Cindy Wang (Google DeepMind) “Building Data-Efficient and Reliable Applications with Large Language Models”

When:
March 8, 2024 @ 12:00 pm – 1:15 pm
2024-03-08T12:00:00-05:00
2024-03-08T13:15:00-05:00
Where:
Hackerman Hall B17
3400 N. Charles Street
Baltimore
MD 21218
Cost:
Free

Abstract

Large Language Models (LLMs) have demonstrated remarkable capabilities across various domains. However, it is still very challenging to build highly-reliable applications with LLMs that support specialized use cases. LLMs trained on web data often excel at capturing general language patterns, but they could struggle to support specialized domains and personalized user needs. Moreover, LLMs can produce errors that are deceptively plausible, making them potentially dangerous for high-trust scenarios. In this talk, I will discuss some of our recent efforts in addressing these challenges with data-efficient tuning methods and a novel factuality evaluation framework. Specifically, my talk will focus on building multilingual applications, one crucial use case often characterized by limited tuning and evaluation data.

Bio

Xinyi(Cindy) Wang is a research scientist at Google DeepMind working on Large Language Models(LLM) and its application to generative question-answering. She has worked on multilingual instruction-tuning for Gemini and multilingual generative models used in Google search. Before Google DeepMind, Cindy Wang obtained her PhD degree in Language Technologies at Carnegie Mellon University. During her PhD, she mainly worked on developing data-efficient natural language processing~(NLP) systems. She has made several contributions in data selection, data representation, and model adaptation for multilingual NLP.

Center for Language and Speech Processing