[Archimedes NLP Group Invited Talk]Understanding the Trade-Offs Between Hallucinations and Mode Collapse in Language Generation

Dates
2025-01-15 14:30 - 16:00
Venue
Artemidos 1 - Amphitheater
TITLE: Understanding the Trade-Offs Between Hallucinations and Mode Collapse in Language Generation

SPEAKER: Grigoris Velegkas(Yale University, USA)

ABSTRACT: Specifying all desirable properties of a language model is challenging, but certain requirements seem essential. Given samples from an unknown language, the trained model should produce valid strings not seen in the training set, and be expressive enough to capture the language's full breadth. Otherwise, outputting invalid strings constitutes "hallucination," and failing to capture the full breadth leads to "mode collapse." Recent work by Kleinberg and Mullainathan [KM24], building on classical work on the closely related problem of language identification by Gold [Gol67] and Angluin [Ang79, 80], provides a concrete mathematical framework to study the problem of language generation. Kleinberg and Mullainathan showed that for all countable collections of languages, it is possible to create a language model that does not hallucinate but suffers from mode collapse. They asked whether this tension between validity and breadth is inherent for language generation.

In this talk, we define various notions of breadth for language generation, and completely characterize when generation with validity and breadth is possible under each of these notions. Our results answer the question of [KM24] and show that this tension between validity and breadth is indeed inherent for language generation. Moreover, we formalize the notion of stable generation, a natural requirement derived from Gold’s work [Gold67], and discuss when this type of generation is achievable. Finally, we discuss the implications of our results in the universal rates setting of Bousquet, Hanneke, Moran, van Handel, and Yehudayoff [BGMvY21]. The talk is based on joint works with Alkis Kalavasis and Anay Mehrotra.


SHORT BIO: Grigoris Velegkas is a final-year PhD student in Computer Science at Yale University, working with Prof. Amin Karbasi. Before that, he studied Electrical and Computer Engineering at the National Technical University of Athens, where he worked with Prof. Dimitris Fotakis. His research lies at the intersection of machine learning and theoretical computer science, and focuses on three main directions: i) understanding generalization properties of ML algorithms, ii) exploring responsible use of ML systems and designing algorithms with provable replicability guarantees, and iii) understanding the interaction between ML algorithms and mechanisms. He was a research intern at Google Research in summer 2023 and summer 2024, and a student researcher from October 2023 to May 2024.


________________________________________________________________________________
Microsoft Teams Need help?
Meeting ID: 364 928 182 762
Passcode: w24fHy
________________________________________________________________________________

 
 
Mon Tue Wed Thu Fri Sat Sun
1
2
3
4
5
6
7
9
11
12
14
16
18
19
20
21
22
23
24
25
26
27
28
29
30
31
 
 

The project “ARCHIMEDES Unit: Research in Artificial Intelligence, Data Science and Algorithms” with code OPS 5154714 is implemented by the National Recovery and Resilience Plan “Greece 2.0” and is funded by the European Union – NextGenerationEU.

greece2.0 eu_arch_logo_en

 

Stay connected! Subscribe to our mailing list by emailing sympa@lists.athenarc.gr
with the subject "subscribe archimedes-news Firstname LastName"
(replace with your details)