Allegro Lab
Thomas Lord Department of Computer Science, University of Southern California.
This page is under development.
The AI, Language, Learning, Generalization, and Robustness (ALLeGRo) Lab studies natural language processing and machine learning with a focus building reliable NLP systems for a wide range of scenarios. We aim for a deeper understanding of how NLP systems work, when they fail, and how they can be improved.
Here are the research questions we have been working on recently:
-
How can we scientifically understand large language models? Our scientific understanding of LLMs lags far behind our ability to engineer them. To bridge this gap, our recent work has studied in-context learning from both a data-centric and mechanistic perspective; we have also investigated the predictability of different LLM capabilities.
-
How should we benchmark modern NLP systems? I have long advocated for benchmarking robustness and uncertainty of NLP systems. Our recent work has benchmarked generalization to long-tail examples and calibration of LLMs. We have also shown that benchmarking under distribution shift can reveal advantages of neurosymbolic approaches.
-
How can smaller open-source models compete with closed-source LLMs? Continued scientific progress relies on access to strong open-source models. Our recent work has improved smaller models by training them to generate reasoning chains.
-
How can advances in NLP inform other disciplines? Developments in NLP promise to have broad impacts across disparate areas of study. We have collaborated with legal experts to operationalize underspecified requirements in the EU’s Digital Services Act in a manner that is both legally justified and technically feasible. I am also interested in collaborating with experts in other disciplines who want to use NLP for their own research; for example, I have built assisted curation tools for biomedical researchers.
news
Sep 03, 2024 | Welcome to the new Allegro Lab website. |
---|
selected publications
- AIESOperationalizing content moderation "accuracy" in the Digital Services Act2024
- ACL FindingsProving membership in LLM pretraining data via data watermarks2024
- NAACL
- EMNLPChain-of-Questions Training with Latent Answers for Robust Multistep Question AnsweringGithub , 2023
- EMNLP Findings
- EACL Findings
- EMNLP FindingsGeneralization Differences between End-to-End and Neuro-Symbolic Vision-Language Reasoning SystemsGithub , 2022
- ACL
- NAACL
- EMNLPWhen Parts are Greater Than Sums: Individual LLM Components Can Outperform Full Models2024
- NeurIPSPre-trained Large Language Models Use Fourier Features to Compute Addition2024
- arxivLanguage Models can Infer Action Semantics for Classical Planners from Environment Feedback2024
- NeurIPSTransformers Learn Higher-Order Optimization Methods for In-Context Learning: A Study with Linear Models2024