Suggester | Paper |
George De Ath |
Random tessellation forests |
George De Ath |
Approximate Cross-Validation in High Dimensions with Guarantees |
George De Ath |
Differentiation of Blackbox Combinatorial Solvers |
George De Ath |
The Implicit Regularization of Ordinary Least Squares Ensembles |
George De Ath |
Near-linear Time Gaussian Process Optimization with Adaptive Batching and Resparsification |
George De Ath |
Predicting the outputs of finite networks trained with noisy gradients |
George De Ath |
Differentiable Expected Hypervolume Improvement for Parallel Multi-Objective Bayesian Optimization |
George De Ath |
Semi-supervised Embedding Learning for High-dimensional Bayesian Optimization |
George De Ath |
Gradient Estimation with Stochastic Softmax Tricks |
George De Ath |
Optimal Lottery Tickets via SubsetSum: Logarithmic Over-Parameterization is Sufficient |
George De Ath |
NGBoost: Natural Gradient Boosting for Probabilistic Prediction |
George De Ath |
Explaining The Behavior Of Black-Box Prediction Algorithms With Causal Learning |
George De Ath |
Uncertainty quantification using martingales for misspecified Gaussian processes |
George De Ath |
Latent variable modeling with random features |
George De Ath |
Discovering Symbolic Models from Deep Learning with Inductive Biases |
George De Ath |
Survey of machine-learning experimental methods atNeurIPS2019 and ICLR2020 |
Tinkle Chugh |
Learning Search Space Partition for Black-box Optimization using Monte Carlo Tree Search |
George De Ath |
A Unifying Perspective on Neighbor Embeddings along the Attraction-Repulsion Spectrum |
George De Ath |
Meta-Surrogate Benchmarking for Hyperparameter Optimization |
George De Ath |
Optimizer Benchmarking Needs to Account for Hyperparameter Tuning |
George De Ath |
Liberty or Depth: Deep Bayesian Neural Nets Do Not Need Complex Weight Posterior Approximations |
George De Ath |
Reverse engineering learned optimizers reveals known and novel mechanisms |
George De Ath |
Learning Search Space Partition for Black-box Optimization using Monte Carlo Tree Search |
George De Ath |
Mastering Atari, Go, chess and shogi by planning with a learned model |
George De Ath |
A Tutorial on Sparse Gaussian Processes and Variational Inference |
George De Ath |
Good practices for Bayesian Optimization of high dimensional structured spaces |
George De Ath |
Expressive yet Tractable Bayesian Deep Learning via Subnetwork Inference |
George De Ath |
GIBBON: General-purpose Information-Based BayesianOptimisatioN |
George De Ath |
Accounting for Variance in Machine Learning Benchmarks |
Melike Karatas |
Loss Surface Simplexes for Mode Connecting Volumes and Fast Ensembling |
George De Ath |
Deep Evidential Regression |
George De Ath |
Sparse Uncertainty Representation in Deep Learning with Inducing Weights |
George De Ath |
Convergence Guarantees for Gaussian Process Means With Misspecified Likelihoods and Smoothness |
George De Ath |
Reinforced Few-Shot Acquisition Function Learning for Bayesian Optimization |
George De Ath |
LocoProp: Enhancing BackProp via Local Loss Optimization |
Richard Everson |
Sharpness-aware Minimization for Efficiently Improving Generalization |
George De Ath |
Kernel Identification Through Transformers |
George De Ath |
What can linear interpolation of neural network loss landscapes tell us? |
George De Ath |
Bayesian Optimization with High-Dimensional Outputs |
George De Ath |
Training BatchNorm and Only BatchNorm: On the Expressive Power of Random Features in CNNs |
George De Ath |
Pretrained Transformers as Universal Computation Engines |
George De Ath |
Gradient-based Hyperparameter Optimization Over Long Horizons |
George De Ath |
Stochastic Contrastive Learning |
George De Ath |
Stochastic Training is Not Necessary for Generalization |
George De Ath |
Deep Neural Networks and Tabular Data: A Survey |
George De Ath |
What's a good imputation to predict with missing values? |
George De Ath |
Revisiting Design Choices in Model-Based Offline Reinforcement Learning |
George De Ath |
SAINT: Improved Neural Networks for Tabular Data via Row Attention and Contrastive Pre-Training |
George De Ath |
Random Hypervolume Scalarizations for Provable Multi-Objective Black Box Optimization |
George De Ath |
High-Dimensional Bayesian Optimisation with Variational Autoencoders and Deep Metric Learning |
George De Ath |
Grokking: Generalization Beyond Overfitting on Small Algorithmic Datasets |
George De Ath |
Correspondence between neuroevolution and gradient descent |
George De Ath |
Gradients are Not All You Need |
George De Ath |
Bayesian Modeling and Uncertainty Quantification for Learning to Optimize: What, Why, and How |
George De Ath |
Fourier Features Let Networks Learn High Frequency Functions in Low Dimensional Domains |
George De Ath |
Do We Really Need Deep Learning Models for Time Series Forecasting? |
George De Ath |
Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer |
George De Ath |
Model soups: averaging weights of multiple fine-tuned models improves accuracy without increasing inference time |
George De Ath |
A Modern Self-Referential Weight Matrix That Learns to Modify Itself |
George De Ath |
Beyond Separability: Analyzing the Linear Transferability of Contrastive Representations to Related Subpopulations |
George De Ath |
The Distributed Information Bottleneck reveals the explanatory structure of complex systems |
George De Ath |
Random Features for Kernel Approximation: A Survey on Algorithms, Theory, and Beyond |
George De Ath |
Learning Strides in Convolutional Neural Networks |
George De Ath |
Comparing Distributions by Measuring Differences that Affect Decision Making |
George De Ath |
Machine Learning State-of-the-Art with Uncertainties |
George De Ath |
It's DONE: Direct ONE-shot learning without training optimization |
George De Ath |
A Learning-based Innovized Progress Operator for Faster Convergence in Evolutionary Multi-objective Optimization |
George De Ath |
Sharpness-Aware Training for Free |
George De Ath |
Transfer Learning with Deep Tabular Models |
George De Ath |
(Certified!!) Adversarial Robustness for Free! |
George De Ath |
Neural Diffusion Processes |
George De Ath |
Sample-Efficient Optimisation with Probabilistic Transformer Surrogates |
George De Ath |
Learning Iterative Reasoning through Energy Minimization |
George De Ath |
Understanding Dataset Difficulty with V-Usable Information |
George De Ath |
Formal Algorithms for Transformers |
Tinkle Chugh |
Bayesian Optimization over Discrete and Mixed Spaces via Probabilistic Reparameterization |
Tinkle Chugh |
A Parallel Technique for Multi-objective Bayesian Global Optimization: Using a Batch Selection of Probability of Improvement |
George De Ath |
AutoML Loss Landscapes |
George De Ath |
Visualizing high-dimensional loss landscapes with Hessian directions |
George De Ath |
Transformers are Sample Efficient World Models |
George De Ath |
Well-tuned Simple Nets Excel on Tabular Datasets |
George De Ath |
Git Re-Basin: Merging Models modulo Permutation Symmetries |
George De Ath |
Diffusion Models: A Comprehensive Survey of Methods and Applications |
George De Ath |
Pre-Train Your Loss: Easy Bayesian Transfer Learning with Informative Prior |
George De Ath |
What Can Transformers Learn In-Context? A Case Study of Simple Function Classes |
George De Ath |
Ensemble-based gradient inference for particle methods in optimization and sampling |
George De Ath |
Protein structure generation via folding diffusion |
George De Ath |
Metadata Archaeology: Unearthing Data Subsets by Leveraging Training Dynamics |
George De Ath |
Learning to Learn with Generative Models of Neural Network Checkpoints |
George De Ath |
Joint Embedding Self-Supervised Learning in the Kernel Regime |
George De Ath |
Batch Normalization Explained |
George De Ath |
An Embarrassingly Simple Approach to Semi-Supervised Few-Shot Learning |
George De Ath |
Efficient Non-Parametric Optimizer Search for Diverse Tasks |
George De Ath |
TabPFN: A Transformer That Solves Small Tabular Classification Problems in a Second |
George De Ath |
Efficient Transformers: A Survey (v2) |
George De Ath |
Foundation Transformers |
George De Ath |
How Much Data Are Augmentations Worth? An Investigation into Scaling Laws, Invariance, and Implicit Regularization |
George De Ath |
Invariance Learning in Deep Neural Networks with Differentiable Laplace Approximations |
George De Ath |
Bayesian Optimization with Conformal Coverage Guarantees |
George De Ath |
Targeted active learning for probabilistic models |
George De Ath |
Flow Network based Generative Models for Non-Iterative Diverse Candidate Generation |
George De Ath |
Multi-Objective GFlowNets |
George De Ath |
Train Short, Test Long: Attention with Linear Biases Enables Input Length Extrapolation |
George De Ath |
HesScale: Scalable Computation of Hessian Diagonals |
George De Ath |
Optimisation & Generalisation in Networks of Neurons |
George De Ath |
Convexifying Transformers: Improving optimization and understanding of transformer networks |
George De Ath |
Compute-Efficient Deep Learning: Algorithmic Trends and Opportunities |
Tinkle Chugh |
GAUCHE: A Library for Gaussian Processes in Chemistry |
George De Ath |
Information-Theoretic Safe Exploration with Gaussian Processes |
George De Ath |
On the role of Model Uncertainties in Bayesian Optimization |
George De Ath |
A Succinct Summary of Reinforcement Learning |
George De Ath |
The Forward-Forward Algorithm: Some Preliminary Investigations |
George De Ath |
A Comprehensive Survey to Dataset Distillation |
George De Ath |
ExcelFormer: A Neural Network Surpassing GBDTs on Tabular Data |
Tinkle Chugh |
SnAKe: Bayesian Optimization with Pathwise Exploration |
George De Ath |
A Comprehensive Survey on Pretrained Foundation Models: A History from BERT to ChatGPT |
Tinkle Chugh |
Multi-Objective Learning to Predict Pareto Fronts Using Hypervolume Maximization |
George De Ath |
Transformer models: an introduction and catalog |
George De Ath |
Efficiently Modeling Long Sequences with Structured State Spaces |
George De Ath |
Adaptive Experimentation at Scale: Bayesian Algorithms for Flexible Batches |
George De Ath |
Effectively Modeling Time Series with Simple Discrete State Spaces |
George De Ath |
Self-Distillation for Gaussian Process Regression and Classification |
Tinkle Chugh |
Double Descent Demystified: Identifying, Interpreting & Ablating the Sources of a Deep Learning Puzzle |
George De Ath |
Prediction-oriented Bayesian active learning |
George De Ath |
A Study of Bayesian Neural Network Surrogates for Bayesian Optimization |
Tinkle Chugh |
Finding Robust Solutions for Many-Objective Optimization Using NSGA-III |
Tinkle Chugh |
Visualization-aided multi-criteria decision-making using interpretable self-organizing maps |