[Colloquium] TODAY, 3PM: Allyson Ettinger - "Understanding" and prediction: Controlled examinations of meaning sensitivity in pre-trained models

Rob Mitchum rmitchum at uchicago.edu
Mon May 2 09:04:07 CDT 2022


*Department of Computer Science Seminar*


*Allyson EttingerAssistant Professor, Department of Linguistics*
*University of Chicago*

*Monday, May 2*
*3:00 - 4:00 PM*
*In Person*: JCL 390
*Remote*: Zoom
<https://uchicagogroup.zoom.us/j/98525152251?pwd=bjEydW9xZkcwTkpPNjNtSktWQkZpdz09>


*"Understanding" and prediction: Controlled examinations of meaning
sensitivity in pre-trained models*
*Abstract*: In recent years, NLP has made what appears to be incredible
progress, with performance even surpassing human performance on some
benchmarks. How should we interpret these advances? Have these models
achieved language “understanding”? Operating on the premise that
“understanding” will necessarily involve the capacity to extract and deploy
the information conveyed in language inputs — the “meaning” — in this talk
I will discuss a series of projects leveraging targeted tests to examine
NLP models’ ability to capture input meaning in a systematic fashion. I
will first discuss work probing model representations for compositional
phrase and sentence meaning, with a particular focus on disentangling
compositional information from encoding of word-level properties. I’ll then
explore models’ ability to extract and use meaning information when
executing the basic pre-training task of word prediction in context. In all
cases, these investigations apply tests that prioritize control of unwanted
cues, so as to target the desired model capabilities with greater
precision. The results of these studies suggest that although models show a
good deal of sensitivity to word-level information, and to certain semantic
and syntactic distinctions, they show little sign of representing
higher-level compositional meaning, or of being able to retain and deploy
such information robustly during word prediction. I will discuss potential
implications of these findings with respect to the goals of achieving
“understanding” with currently dominant pre-training paradigms.

*Bio*: Dr. Allyson Ettinger’s research is focused on language processing in
humans and in artificial intelligence systems, motivated by a combination
of scientific and engineering goals. For studying humans, her research uses
computational methods to model and test hypotheses about mechanisms
underlying the brain’s processing of language in real time. In the
engineering domain, her research uses insights and methods from cognitive
science, linguistics, and neuroscience in order to analyze, evaluate, and
improve natural language understanding capacities in artificial
intelligence systems. In both of these threads of research, the primary
focus is on the processing and representation of linguistic meaning.

*Zoom Link*:
https://uchicagogroup.zoom.us/j/98525152251?pwd=bjEydW9xZkcwTkpPNjNtSktWQkZpdz09
*Meeting ID*: 985 2515 2251
*Password*:  315057


-- 
*Rob Mitchum*

*Associate Director of Communications for Data Science and Computing*
*University of Chicago*
*rmitchum at uchicago.edu <rmitchum at ci.uchicago.edu>*
*773-484-9890*
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mailman.cs.uchicago.edu/pipermail/colloquium/attachments/20220502/93fabb0e/attachment.html>


More information about the Colloquium mailing list