Kenton Murray (JHU) – “Improving Preference Optimization and Pre-Training Algorithms: Insights from focusing on Multilingual NLP and Machine Translation”

When:
February 10, 2025 @ 12:00 pm – 1:15 pm
2025-02-10T12:00:00-05:00
2025-02-10T13:15:00-05:00
Where:
Hackerman Hall B17
3400 N CHARLES ST
Cost:
Free

Abstract

Current Natural Language Processing systems rely heavily on Large Language Models (LLMs) which are trained rather naively on large amounts of text using autoregressive next token prediction. Often, they are fine-tuned to mimic human preferences in order to accomplish more complex tasks. While these methods work surprisingly well, they have inherent biases stemming from design and engineering decisions arising from focusing on monolingual English. In this talk, I’ll discuss improved algorithms that we have developed to overcome deficiencies unearthed by working on multilingual and translation tasks. These improvements are general enough to be applied back to standard monolingual English training and fine-tuning. For instance, our CPO algorithm uses half the GPU memory of other popular Preference Optimization methods, or our Cooldown Temperature Sampling algorithm for Pre-Training converges faster and prevents overfitting. By focusing on languages beyond English, this talk will cover how we can gain insights into developing more efficient and robust general methods.

Bio

Kenton Murray is a Research Scientist at Johns Hopkins University where he directs the Maieutic Lab. His work focuses on Multilingual Natural Language Processing and Machine Translation with a particular emphasis on low-resource and morphologically rich languages across text, speech, and vision modalities. He received his PhD from the University of Notre Dame, a Master’s from Carnegie Mellon University, and a Bachelor’s from Princeton University.

Center for Language and Speech Processing