Events

Jieyu Zhao (USC) – “Trustworthy LLMs — our efforts on mitigating issues regarding social bias, safety and reliability”

November 5, 2024
When: November 8, 2024 @ 12:00 pm – 1:15 pm
Where: Hackerman Hall B17, 3400 N CHARLES ST, Baltimore, MD 21218

Abstract The rapid advancement of large language models (LLMs) has unlocked a myriad of possibilities for positive societal impact, ranging from enhancing accessibility and communication to supporting disaster response and public health initiatives. However, the[…]

Read More

Adam Byerly (JHU) “How Effective Is Self-Consistency for Long-Context Problems?”

October 31, 2024
When: November 4, 2024 @ 12:00 pm – 1:15 pm
Where: Hackerman Hall B17, 3400 N CHARLES ST, Baltimore, MD 21218

Abstract Self-consistency (SC) has been demonstrated to enhance the performance of large language models (LLMs) across various tasks and domains involving short content. However, does this evidence support its effectiveness for long-context problems? In this talk, we examine the[…]

Read More

Yen-ju Lu (JHU) “CA-SSLR: Condition-Aware Self-Supervised Learning Representation for Generalized Speech Processing”

October 31, 2024
When: November 1, 2024 @ 12:00 pm – 1:15 pm
Where: Hackerman Hall B17, 3400 N CHARLES ST, Baltimore, MD 21218

Abstract We introduce Condition-Aware Self-Supervised Learning Representation (CA-SSLR), a generalist conditioning model broadly applicable to various speech-processing tasks. Compared to standard fine-tuning methods that optimize for downstream models, CA-SSLR integrates language and speaker embeddings from[…]

Read More

Leo Du (JHU) “Discrete Gradient-based Sampling with applications to Language Models”

October 21, 2024
When: October 21, 2024 @ 12:00 pm – 1:15 pm
Where: Hackerman Hall B17, 3400 N CHARLES ST, Baltimore, MD 21218

Abstract Gradient-based sampling algorithms are a cornerstone of modern Bayesian computation, widely used in applications ranging from probabilistic programming to diffusion models.  While these methods perform exceptionally well in continuous domains, extending them to discrete[…]

Read More

Roger Grosse (University of Toronto) “Probabilistic Inference in Language Models via Twisted Sequential Monte Carlo”

October 4, 2024
When: October 11, 2024 @ 12:00 pm – 1:15 pm
Where: Hackerman Hall B17, 3400 N CHARLES ST, Baltimore, MD 21218

Abstract Numerous capability and safety techniques of Large Language Models (LLMs), including RLHF, automated red-teaming, prompt engineering, and infilling, can be cast as sampling from an unnormalized target distribution defined by a given reward or[…]

Read More

Center for Language and Speech Processing