MITB Banner

6 Outstanding Papers Presented at NeurIPS 2023 

Out of the 13,321 papers submitted at NeurIPS 2023, the top of the lot have won the outstanding awards this year.

Share

Listen to this story

During the ongoing Neural Information Processing Systems (NeurIPS) annual conference, reviewers and chairpersons are currently evaluating tens of thousands of submissions.

Out of the 13,321 papers submitted by authors and researchers worldwide, the top of the lot have won the outstanding awards this year. Here are the 6 outstanding papers announced by NeurIPS in 2023: 

Outstanding Main Track Papers

Privacy Auditing with One (1) Training Run

Steinke, Nasr, and Jagielski propose an efficient auditing scheme for assessing the privacy of differentially private machine learning (ML) systems in a single training run. They leverage the parallelism of adding or removing multiple training examples independently. They avoid the computational cost of group privacy by analysing the connection between differential privacy and statistical generalisation.

 Their approach works in both black-box and white-box settings, requiring minimal assumptions about the algorithm. They demonstrate the effectiveness of their framework on DP-SGD, achieving meaningful privacy bounds with just one model, while standard methods would need hundreds of models.

Are Emergent Abilities of Large Language Models a Mirage?

Schaeffer, Miranda, and Koyejo challenge the idea that large language models (LLMs) exhibit true emergent abilities. They propose that perceived emergent abilities are often a result of the researcher’s metric choices rather than fundamental changes in model behaviour with scale. They support this with a mathematical model and three analyses:

  1. Confirming predictions on metric effects using InstructGPT/GPT-3
  2. Validating predictions in a meta-analysis on BIG-Bench
  3. Demonstrating how metric choices can create apparent emergent abilities in vision tasks across different networks.

Their findings suggest that alleged emergent abilities may vanish with different metrics, questioning the notion that they are intrinsic to scaled AI models.

Runner-Ups

Scaling Data-Constrained Language Models

In the paper, researchers explored scaling language models in data-limited scenarios, given the potential constraint on internet text data. They conducted extensive experiments, varying data repetition and computed budgets of up to 900 billion tokens and 9 billion parameters. Results showed that with limited data and a fixed computing budget, up to 4 epochs of repeated data had minimal impact on loss. However, further repetition diminished the value of additional compute. 

They proposed a scaling law for compute optimality, considering the declining value of repeated tokens and excess parameters. Additionally, they tested methods to alleviate data scarcity, such as augmenting with code data or removing common filters.

Models and datasets from 400 training runs are freely available on GitHub.

Direct Preference Optimization: Your Language Model is Secretly a Reward Model

Here, researchers introduced Direct Preference Optimization (DPO) as a streamlined alternative to Reinforcement Learning from Human Feedback (RLHF) for controlling large unsupervised language models. Unlike RLHF, DPO avoids the complexity and instability of fitting reward models and fine-tuning. Leveraging a mapping between reward functions and optimal policies, DPO directly optimises a single-stage policy training process, solving a classification problem on human preference data. 

The experiments demonstrate that DPO can effectively align language models with human preferences, outperforming RLHF in sentiment control and improving response quality in summarization and dialogue. Notably, DPO is more straightforward to implement and train.

Datasets and Benchmarks Papers:

ClimSim: A large multi-scale dataset for hybrid physics-ML climate emulation

Machine learning experts have introduced ClimSim, the largest hybrid ML-physics dataset, co-created by climate scientists and ML researchers. With 5.7 billion pairs of input-output vectors, it isolates the impact of high-resolution physics on macro-scale climate states. Global and spanning multiple years, the dataset facilitates emulators compatible with operational climate simulators.

The data and code are released openly to support the development of hybrid ML-physics and high-fidelity climate simulations. 

DecodingTrust: A Comprehensive Assessment of Trustworthiness in GPT Models

With the rise of GPT models, practitioners are considering using them for sensitive applications like healthcare and finance, but research reveals undisclosed vulnerabilities. GPT models, including GPT-4, can produce biassed, toxic outputs and unintentionally leak private information. 

Despite GPT-4’s generally improved trustworthiness, it exhibits vulnerability to jailbreaking systems or misleading user prompts. This study highlights previously unrecognised trustworthiness gaps in GPT models.

The benchmark is publicly available on GitHub.

Share
Picture of Tasmia Ansari

Tasmia Ansari

Tasmia is a tech journalist at AIM, looking to bring a fresh perspective to emerging technologies and trends in data science, analytics, and artificial intelligence.
Related Posts

CORPORATE TRAINING PROGRAMS ON GENERATIVE AI

Generative AI Skilling for Enterprises

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

Upcoming Large format Conference

May 30 and 31, 2024 | 📍 Bangalore, India

Download the easiest way to
stay informed

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

AI Forum for India

Our Discord Community for AI Ecosystem, In collaboration with NVIDIA. 

Flagship Events

Rising 2024 | DE&I in Tech Summit

April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore

MachineCon GCC Summit 2024

June 28 2024 | 📍Bangalore, India

MachineCon USA 2024

26 July 2024 | 583 Park Avenue, New York

Cypher India 2024

September 25-27, 2024 | 📍Bangalore, India

Cypher USA 2024

Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA

Data Engineering Summit 2024

May 30 and 31, 2024 | 📍 Bangalore, India

Subscribe to Our Newsletter

The Belamy, our weekly Newsletter is a rage. Just enter your email below.