Maha Elbayad: “Large Concept Models: Language Modeling in a Sentence Representation Space”

When:
February 7, 2025 @ 12:00 pm – 1:15 pm
2025-02-07T12:00:00-05:00
2025-02-07T13:15:00-05:00
Where:
Hackerman Hall B17
3400 N CHARLES ST
Cost:
Free

Abstract

While large language models (LLMs) have revolutionized AI, their token-level processing contrasts sharply with human cognition’s multi-level abstraction. This talk explores moving beyond token-based manipulation to reason in a latent space. We introduce the Large Concept Model (LCM), an architecture that operates on language- and modality-agnostic “concepts,” represented as sentence embeddings within the SONAR space. Trained for autoregressive sentence prediction, the LCM learns to reason and generate at a higher semantic level. Evaluated on generative tasks like summarization and summary expansion, the LCM demonstrates impressive zero-shot generalization across multiple languages. Crucially, this concept-based representation within the SONAR space naturally facilitates multimodal extension, enabling the model to reason about and generate content grounded in diverse sensory inputs. This paves the way for more robust and human-like AI systems.

Bio

Maha Elbayad is a senior research scientist at Meta AI (Menlo Park, CA), working on massively multilingual and multimodal machine translation, contributing to projects like No-Language-Left-Behind and Seamless. A key focus of her research is latent space reasoning, exploring how to build intelligent systems that can integrate and leverage information from diverse sources like images, text, and audio.

Center for Language and Speech Processing