This project tackles a core challenge in automated machine learning: how do you efficiently explore vast hyperparameter spaces across diverse neural architectures when compute is limited?
Text classification models—LSTMs, Transformers, CNNs—each have unique hyperparameter landscapes. Traditional HPO methods waste significant compute evaluating poor configurations. Our approach introduces a proxy search space: a lightweight benchmark that mirrors the structure of the full optimization problem but evaluates in a fraction of the time.
By running optimizer selection on this proxy first, we identified DEHB as significantly outperforming PriorBand and RandomSearch for our task—before committing any compute to the actual optimization.