Skip to main content

NeurIPS 2022: Paper review #2

25 January 2023
  • Quantitative Research

G-Research were headline sponsors at NeurIPS 2022, in New Orleans.

ML is a fast-evolving discipline; attending conferences like NeurIPS and keeping up-to-date with the latest developments is key to the success of our quantitative researchers and machine learning engineers.

Our NeurIPS 2022 paper review series gives you the opportunity to hear about the research and papers that our quants and ML engineers found most interesting from the conference.

Here, Simon L, Senior Quantitative Researcher at G-Research, discusses two papers from NeurIPS:

  • Agreement-on-the-Line: Predicting the Performance of Neural Networks under Distribution Shift
  • Deep Ensembles Work, But Are They Necessary?

Agreement-on-the-Line: Predicting the Performance of Neural Networks under Distribution Shift

Christina Baek, Yiding Jiang, Aditi Raghunathan, Zico Kolter

This paper presents a method for predicting out-of-distribution (OOD) accuracy of a neural network from its in-distribution (ID) performance, without the need for a labelled test set.

The method is based on the observation that ID/OOD agreement of predictions from two neural network classifiers are strongly linearly related (post a probit transform), in a very similar way to how ID/OOD accuracy was shown to be in a more general setting in recent work (Accuracy on the line by Miller et al., 2021). Specifically, the values of the intercept and slope coefficients of these two regressions are very close.

Another interesting observation is that, for neural networks, ID/OOD agreement tends to hold whenever ID/OOD accuracy on the line holds. Hence an agreement regression (which doesn’t require any labels), plus a calculation of in-distribution accuracy can be used to estimate out of distribution accuracy.

The authors even demonstrate that the agreement regression could be performed on checkpoints of a single model fit. These are fascinating empirical findings and hopefully further understanding of the specificity of these observations to neural networks will follow in the future.


Deep Ensembles Work, But Are They Necessary?

Taiga Abe, E. Kelly Buchanan, Geoff Pleiss, Richard Zemel, John P. Cunningham

There were a number of interesting papers at NeurIPS making progress in the understanding of different properties of deep ensembles.

This paper, for instance, questions the claim that predictions from deep ensembles have robustness and uncertainty quantification benefits over and above those of single neural networks.

Taking robustness, the authors show empirically that ensembles sit on the same ID-OOD accuracy line as their constituent models, and hence, the diversity of ensemble predictions does not yield additional protection to dataset distribution shift, relative to a single model (perhaps larger or with a different architecture) with the same predictive performance.

Further, they show empirically that some common metrics of ensemble diversity take similar values on both ID and OOD datasets. By decomposing common ensemble uncertainty metrics into average single model uncertainty, plus a diversity term, they conclude that deep ensemble uncertainty is driven by the uncertainty of a single model.

Deep ensembles have lots of interesting and useful properties, and I found this work helpful in clarifying some of these more carefully.

Read all of our NeurIPS 2022 paper reviews
NeurIPS 2022: Paper review #3
  • 25 Jan 2023

Our NeurIPS 2022 paper review series gives you the opportunity to hear about the research and papers that our quants and ML engineers found most interesting from the conference. Here, a Machine Learning Engineer at G-Research, discusses two papers from NeurIPS.

Read more

G-Research at NeurIPS 2022

We work on a very mature problem at GE Research predicting financial markets, which means we need to stay at the cutting edge of what we do. That's why events like noritz, where we've ATO tier sponsors are crucial for our business as they bring together the best machine learning practitioners to present and discuss the latest research and innovation in ml. We encourage our quant researchers and machine learning engineers to attend leading conferences in person to further develop their skills and stay abreast of the latest technological developments from some of the brightest minds in the industry. Additionally, a number of our talent acquisition team were on hand throughout the week to talk to attendees about what we do, including the various research and engineering roles we are currently hiring for, and we kept everyone fueled with the help of our head barista as busy as we were inside Europe as a headline sponsor. We also ran a number of events outside the conference hall during the week as well. Not least the first ever G Research boat party held on a classic paddle steamer on the Mississippi. What better way to bring together like-minded people in New Orleans? And we were delighted. So many people wanted to come along, As well as providing a unique networking opportunity. This event also gave us the chance to give our guests a flavor of what life is like. At Achieve research, we pride ourselves on cultivating an environment where smart people come together to challenge themselves, enjoy their work, and achieve things as a team. And there's also plenty of opportunity for fun along the way. You know what they say about all work and no play. Want to learn more about GE Research or meet us at a future event? Visit our website to find out more.
Open video transcript

Stay up to date with
G-Research