2024-11-18
To Ask Israel
- Grant requirements
- Problems with frameworks (OOM etc?)
- Constraints enforced?
- Datasets tested?
- How did you set it up
- Snellius or something else?
Info
OpenMLSuites:
Sparse datasets = turn off inference time evaluation (inference_time_measurements
)
tasks_ids = [360932, 360933]AMLB:
Increase time overhead_time_multiplier
or overhead_time_multiplier
inference_time_measurements
= 5? (do it to 10)
Fix versions of the frameworksFrameworks
- flaml:
If there is a problem with “pandas”
Only install it inside the virtual env of flaml
If there is a problem with “callback”:
Only install the next inside the virtual env of flaml
pip install xgboost==2.0.3
- GAMA:
a package from sklearn should be updated (I don’t remember)
- H2OAutoML:
Define the memory manually (exec.py
)jvm_memory = str(round(config.max_mem_size_mb * 2/3))+"M"
- NaiveAutoML:
- FEDOT:
- AutoGluon
“NoResultError: Trainer has no fit models that can infer.” (Figure)
Slurm:
AutoGluon does not respect the resources defined.
Run each fold per each framework:
python runbenchmark.py AutoGluon_benchmark:2024Q1 openml/t/360975 5min8c_gp3 -f 0
Clean cache
find -O3 /gpfs/scratch1/nodespecific/ -maxdepth 2 -type d -user icampero -exec rm -rf {} \;
Touch:
find israelAutoML2 -exec touch -a -m {} \;
(edited)
Slurm
- myquota
- rome/thin or genoa
- run it with 28GB memory
- problem with h20 and autogloun
- global save : false, retrieve save from backup
- clean cache
- inode
- 2 weeks without touching folder
- use scratch to run your stuff