Poster Schedule

Tuesday

All papers from Sessions I, II & III will present a poster in this slot. Additionally the following non-archival content will be presented:

  • MetaARIMA: Automatic Configuration of ARIMA using Classifier Chains
  • Towards Dynamic Priors in Bayesian Optimization for Hyperparameter Optimization
  • Quickly Tuning Foundation Models for Image Segmentation
  • A Preliminary Evaluation of Large Language Models for Data Science Code Generation
  • Cost-aware Stopping for Bayesian Optimization
  • Stitching Disparate Neural Network Layers with Complex Adapters and Spatial Rescaling
  • ReLU is all you need for NASWOT
  • Successive Halving with Learning Curve Prediction via Latent Kronecker Gaussian Processes
  • Improved Gaussian Process Hyperparameter Fitting for Bayesian Optimization
  • LISTEN to Your Preferences: An LLM Framework for Multi-Objective Selection


Wednesday

All papers from Sessions IV & V will present a poster in this slot. Additionally the following non-archival content will be presented:

  • Towards Exploiting Early Termination for Multi-Fidelity Hyperparameter Optimization
  • Quantifying Module Interactions in the PSO-X Framework
  • Bayesian Optimisation Against Climate Change: Applications and Benchmarks
  • Zero-Cost Benchmarks: Towards Lower Reliance on Spearman Rank Correlation
  • Cost-aware Bayesian Optimization via the Pandora’s Box Gittins Index
  • Stress Testing Classifiers around the Decision Boundary with Latent Diffusion
  • The Gittins Index: A Design Principle for Decision-Making Under Uncertainty
  • Tune My Adam, Please!
  • Multi-objective Hyperparameter Optimization in the Age of Deep Learning
  • How Usable is Automated Feature Engineering for Tabular Data?
  • Automated Data Preparation for Machine Learning
  • Surrogate Benchmarks for Model Merging Optimization
  • ParticleML: AutoML Through Electromagnetic Physics Simulation
  • Prometheus: A Recursively Self-Improving NAS System
  • Data-Efficient Ranking of Recommendation Models
  • Object-Flow Machine Learning: Active learning framework utilizing protocols information
  • Algorithm Configuration for Structured Pfaffian Settings

Accepted Non-Archival Content

  • MetaARIMA: Automatic Configuration of ARIMA using Classifier Chains
  • Towards Dynamic Priors in Bayesian Optimization for Hyperparameter Optimization
  • Algorithm Configuration for Structured Pfaffian Settings
  • Quickly Tuning Foundation Models for Image Segmentation
  • A Preliminary Evaluation of Large Language Models for Data Science Code Generation
  • Cost-aware Stopping for Bayesian Optimization
  • Stitching Disparate Neural Network Layers with Complex Adapters and Spatial Rescaling
  • ReLU is all you need for NASWOT
  • Successive Halving with Learning Curve Prediction via Latent Kronecker Gaussian Processes
  • Improved Gaussian Process Hyperparameter Fitting for Bayesian Optimization
  • LISTEN to Your Preferences: An LLM Framework for Multi-Objective Selection
  • Towards Exploiting Early Termination for Multi-Fidelity Hyperparameter Optimization
  • Quantifying Module Interactions in the PSO-X Framework
  • Bayesian Optimisation Against Climate Change: Applications and Benchmarks
  • Zero-Cost Benchmarks: Towards Lower Reliance on Spearman Rank Correlation
  • Cost-aware Bayesian Optimization via the Pandora’s Box Gittins Index
  • Stress Testing Classifiers around the Decision Boundary with Latent Diffusion
  • The Gittins Index: A Design Principle for Decision-Making Under Uncertainty
  • Tune My Adam, Please!
  • Multi-objective Hyperparameter Optimization in the Age of Deep Learning
  • How Usable is Automated Feature Engineering for Tabular Data?
  • Automated Data Preparation for Machine Learning
  • Surrogate Benchmarks for Model Merging Optimization
  • ParticleML: AutoML Through Electromagnetic Physics Simulation
  • Prometheus: A Recursively Self-Improving NAS System
  • Data-Efficient Ranking of Recommendation Models
  • Object-Flow Machine Learning: Active learning framework utilizing protocols information