MITB Banner

5 Amazing Papers Presented by Meta at ICML 2023

The papers cover important topics like machine vision, computational biology, speech recognition, and robotics

Share

Listen to this story

The 40th International Conference on Machine Learning (ICML) is happening this year in Honolulu, Hawaii. AI/ML experts from across the globe have gathered at the venue to showcase and release state-of-the-art research focussing on all facets of machine learning deployed in closely interconnected domains such as AI, statistics, and data science. This year, the spotlight is on vital application areas like machine vision, computational biology, speech recognition, and robotics. The first IMCL took place in 1980 in Pittsburgh.

Let’s take a look at the papers that big tech Meta presented at the conference this year. 

ELSA: Efficient Label Shift Adaptation through the Lens of Semiparametric Models

This study focuses on the problem of domain adaptation with label shift, where the distribution of labels differs between training and testing datasets, while the distribution of features remains the same. The existing methods for label shift adaptation have estimation errors or complex post-prediction calibrations. To overcome these issues, the researchers propose a moment-matching framework called Efficient Label Shift Adaptation (ELSA).

ELSA estimates adaptation weights by solving linear systems, ensuring accurate and efficient performance without post-prediction calibrations. Theoretical analysis proves its consistency and normalcy, while empirical results show state-of-the-art performance.

Reward-Mixing MDPs with a Few Latent Contexts are Learnable

This research focuses on episodic reinforcement learning in a type of decision-making process called “reward-mixing Markov decision processes” (RMMDPs). In these processes, at the start of each episode, nature randomly selects a hidden reward model from M choices, and the agent interacts with the system for H time steps. The goal is to learn a policy that maximises cumulative rewards over H steps for this hidden reward model. The researchers present a new algorithm called EM2, which efficiently finds a nearly optimal policy for any M ≥ 2. They also establish a lower bound on the sample complexity of RMMDPs, showing that high sample complexity in M is unavoidable.

Read more: Meta-Qualcomm Partnership Will Bring Llama 2 to the Masses

Masked Trajectory Models for Prediction, Representation, and Control

Along with  UC Berkeley, Georgia Tech, Google Research, and Meta AI contributed to this project. Masked Trajectory Models (MTM) are a new way of making decisions step by step. The team takes a sequence of states and actions and tries to figure out the sequence by using random parts of it. They learn to be flexible and can do different tasks just by using different parts of the sequence. For example, they can be used as models for predicting future actions, figuring out past actions, or even as a learning agent. In tests, the same MTM network can perform as well as or even better than specialised networks designed for specific tasks. MTM also helps speed up learning in traditional RL algorithms and competes well with specialised offline RL methods in benchmarks.

Hyperbolic Image-Text Representations

Meta introduced MERU which helps organise visual and written ideas in a hierarchy. For example, when we say “dog”, it includes all dog images. Existing models like CLIP don’t explicitly capture this hierarchy. MERU uses hyperbolic spaces, which are good for representing tree-like data, allowing it to better capture the relationships between images and text. Results show that MERU creates a clear and understandable representation while performing as well as CLIP on tasks like image classification and image-text matching.

Hiera: A Hierarchical Vision Transformer without the Bells-and-Whistles

Researchers have enhanced vision transformers for supervised classification, but the extra vision-specific elements have made them slower than the original ViT version. This paper introduces Hiera, a simple hierarchical vision transformer. By pre-training it with a strong visual task (MAE), unnecessary complexity is removed while maintaining accuracy. Hiera outperforms previous models, proving to be faster in both training and inference. Its performance is evaluated on various image and video recognition tasks.

Read more: Top 6 Papers Presented by Meta at CVPR 2023

Share
Picture of Shritama Saha

Shritama Saha

Shritama (she/her) is a technology journalist at AIM who is passionate to explore the influence of AI on different domains including fashion, healthcare and banks.
Related Posts

CORPORATE TRAINING PROGRAMS ON GENERATIVE AI

Generative AI Skilling for Enterprises

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

Upcoming Large format Conference

May 30 and 31, 2024 | 📍 Bangalore, India

Download the easiest way to
stay informed

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

AI Courses & Careers

Become a Certified Generative AI Engineer

AI Forum for India

Our Discord Community for AI Ecosystem, In collaboration with NVIDIA. 

Flagship Events

Rising 2024 | DE&I in Tech Summit

April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore

MachineCon GCC Summit 2024

June 28 2024 | 📍Bangalore, India

MachineCon USA 2024

26 July 2024 | 583 Park Avenue, New York

Cypher India 2024

September 25-27, 2024 | 📍Bangalore, India

Cypher USA 2024

Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA

Data Engineering Summit 2024

May 30 and 31, 2024 | 📍 Bangalore, India

Subscribe to Our Newsletter

The Belamy, our weekly Newsletter is a rage. Just enter your email below.