Skip to main content
Back to News
NeurIPs Paper Reviews 2023 #2

NeurIPs Paper Reviews 2023 #2

23 January 2024
  • Quantitative Research

Our team of quantitative researchers have shared the most interesting research presented during workshops and seminars at NeurIPs 2023.

Discover the perspectives of quantitative researcher Paul, as he discusses his most compelling findings from the conference.

NeurIPs Booth 2022

Sharpness-Aware Minimization Leads to Low-Rank Features

Maksym Andriushchenko, Dara Bahri, Hossein Mobahi, Nicolas Flammarion

In overparametrised neural networks, sharpness of minima has been observed to correlate negatively with the generalisation error of the model. Sharpness-aware minimisation (SAM) is a recent algorithm that introduces an explicit sharpness penalty to the optimisation objective which has been shown to improve model performance.

In this paper, the authors investigate the effect that SAM has on the features of the model. They demonstrate that SAM reduces the feature rank at different layers, as measured by the number of principal components that are needed to capture 99% of the variance, compared to networks that are trained using standard minimisation algorithms. This can for instance be used to reduce the dimensionality of the feature space, improving the performance of downstream tasks. In contrast, the authors found that directly imposing a lower feature rank on the model itself did not lead to improved generalisation. This suggests that the low rank is a useful side effect but not a full explanation of the benefits of SAM.

To further understand the mechanism behind this effect, the authors study a two-layer ReLU network. They show, both experimentally and theoretically, that SAM decreases pre-activation values within the network. This, in turn, reduces the number of non-zero activations and results in the observed low rank of the features.

Sharpness-Aware Minimization Leads to Low-Rank Features
NeurIPS 2022 Paper Reviews

Read paper reviews from NeurIPS 2022 from a number of our quantitative researchers and machine learning practitioners.

Read now

When Do Neural Nets Outperform Boosted Trees on Tabular Data?

Duncan C. McElfresh, Sujay Khandagale, Jonathan Valverde, Vishak Prasad C, Ganesh Ramakrishnan, Micah Goldblum, Colin White

This paper presents a comprehensive study comparing the performance of neural network (NN), gradient boosted decision trees (GBDT), and baseline algorithms like linear or k-nearest neighbour models on a large number of tabular datasets. It also introduces a benchmark suite of challenging tabular datasets to accelerate research in this area.

The study shows that no single algorithm dominates on all datasets, nearly all algorithms examined ranked first on at least one dataset. When aggregating the algorithms by their respective family, GBDTs are high-performing on slightly more datasets than NNs and baseline methods while also being faster than NNs. However, in many cases the difference in performance between NNs and GBDTs is either negligible or, at fixed budget, tuning the hyperparameters of GBDT is more useful than trying out different methods.

Additionally, the authors present a metafeature analysis to identify dataset properties that correlate with superior performance of certain techniques, which is helpful for practitioners selecting suitable methods for their respective datasets. For example, the authors demonstrate that GBDTs tend to outperform NNs on datasets with heavy-tailed, skewed, or high-variance features.

Quantitative Research and Machine Learning

Want to learn more about life as a researcher at G-Research?

Learn more

Read more of our quantitative researchers thoughts

NeurIPs Paper Reviews 2023 #1

Discover the perspectives of Danny, one of our machine learning engineers, on the following papers:

  • A U-turn on Double Descent: Rethinking Parameter Counting in Statistical Learning
  • Normalization Layers Are All That Sharpness-Aware Minimization Needs
Paper Review #1
NeurIPS Paper Review 2023 #3

Discover the perspectives of Szymon, one of our quantitative researchers, on the following papers:

  • Convolutional State Space Models for Long-Range Spatiotemporal Modeling
  • How to Scale Your EMA
Paper Review #3
NeurIPS Paper Review 2023 #4

Discover the perspectives of Dustin, our scientific director, on the following papers:

  • Abide by the law and follow the flow: conservation laws for gradient flows
  • The Tunnel Effect: Building Data Representations in Deep Neural Networks
Paper Review #4
NeurIPS Paper Review 2023 #5

Discover the perspectives of Laurynas, one of our machine learning engineers, on the following papers:

  • Monarch Mixer: A Simple Sub-Quadratic GEMM-Based Architecture
  • QLoRA: Efficient Finetuning of Quantized LLMs
Paper Review #5
NeurIPS Paper Review 2023 #6

Discover the perspectives of Rui, one of our quantitative analyst, on the following papers:

  • Predict, Refine, Synthesize: Self-Guiding Diffusion Models for Probabilistic Time Series Forecasting
  • Conformal Prediction for Time Series with Modern Hopfield Networks
Paper Review #6

Latest News

Distinguished Speaker Symposium on Prime Numbers and Cryptography
  • 04 Dec 2024

The G-Research Distinguished Speaker Symposium returns on Thursday 21st November 2024, with an exciting line-up of speakers, as well as an opportunity to learn more about G-Research and network with our team.

Read article
G-Research November 2024 Grant Winners
  • 02 Dec 2024

Each month, we provide up to £2,000 in grant money to early career researchers in quantitative disciplines. Hear from our November grant winners.

Read article
G-Research October 2024 Grant Winners
  • 08 Nov 2024

Each month, we provide up to £2,000 in grant money to early career researchers in quantitative disciplines. Hear from our October grant winners.

Read article

Latest Events

  • Technology Innovation and Open Source

Tech Careers in Finance – University of Cambridge

23 Jan 2025 New Museum Site, Bene't St, Cambridge CB2 3PT
  • Technology Innovation and Open Source

Spring into Quant Finance 2025

06 Apr 2025 - 11 Apr 2025 Saint-Jean-Cap-Ferrat, France
  • Machine Learning
  • Quantitative Research

NeurIPS 2024

10 Dec 2024 - 15 Dec 2024 Vancouver Convention Centre, 1055 Canada Pl, Vancouver, BC V6C 0C3, Canada

Stay up to date with
G-Research