<div dir="ltr"><div dir="ltr"><div><div class="gmail_default" style="font-family:georgia,serif;font-size:small"><font style="color:rgb(0,0,0);vertical-align:inherit"><font style="vertical-align:inherit"><b>When:</b> </font></font><font style="color:rgb(0,0,0);vertical-align:inherit"><font style="vertical-align:inherit"> Monday, January 13th at <b style="background-color:rgb(255,255,0)">11:25am CT</b><b> </b></font></font></div><div><p style="font-variant-numeric:normal;font-variant-east-asian:normal;font-stretch:normal;line-height:normal;margin:0px"><font color="#000000"><font style="vertical-align:inherit"><font face="georgia, serif" style="vertical-align:inherit"><b><span style="background-color:rgb(255,255,0)"><br></span></b></font></font></font></p><div class="gmail_default"><font face="georgia, serif" color="#000000"><b>Where: </b>Talk will be given <font style="font-weight:bold"><u>live, in-person</u></font><font style="font-weight:bold"> </font>at</font></div><p class="MsoNormal" style="margin:0in;line-height:normal;background-image:initial;background-position:initial;background-size:initial;background-repeat:initial;background-origin:initial;background-clip:initial"><font face="georgia, serif" color="#000000"> TTIC, 6045 S. Kenwood Avenue</font></p><p class="MsoNormal" style="margin:0in;line-height:normal;background-image:initial;background-position:initial;background-size:initial;background-repeat:initial;background-origin:initial;background-clip:initial"><font color="#000000" face="georgia, serif"> 5th Floor, Room 530<b> </b></font></p><p class="MsoNormal" style="margin:0in;line-height:normal;background-image:initial;background-position:initial;background-size:initial;background-repeat:initial;background-origin:initial;background-clip:initial"><b><font face="georgia, serif" color="#000000"><br></font></b></p><p class="MsoNormal" style="margin:0in 0in 0.0001pt;line-height:normal;background-image:initial;background-position:initial;background-size:initial;background-repeat:initial;background-origin:initial;background-clip:initial"><font face="georgia, serif" color="#000000"><b style="letter-spacing:0.2px">Virtually:</b><span style="letter-spacing:0.2px"> via Panopto </span>(<a href="https://uchicago.hosted.panopto.com/Panopto/Pages/Viewer.aspx?id=22d6c613-0a51-454a-8e08-b2200108e3b6" target="_blank">livestream</a><span style="letter-spacing:0.2px">)</span><br></font></p><p class="MsoNormal" style="margin:0in 0in 0.0001pt;line-height:normal;background-image:initial;background-position:initial;background-size:initial;background-repeat:initial;background-origin:initial;background-clip:initial"><font face="georgia, serif" color="#000000"><br></font></p><p class="MsoNormal" style="margin:0in 0in 0.0001pt;line-height:normal;background-image:initial;background-position:initial;background-size:initial;background-repeat:initial;background-origin:initial;background-clip:initial"><font face="georgia, serif" color="#000000"><font style="vertical-align:inherit"><font style="vertical-align:inherit"><b>Who: </b> </font></font><span class="gmail_default"></span><span class="gmail_default"></span><span class="gmail_default"></span>Jim Rehg, UIUC</font></p><p class="MsoNormal" style="margin:0in 0in 0.0001pt;line-height:normal;background-image:initial;background-position:initial;background-size:initial;background-repeat:initial;background-origin:initial;background-clip:initial"></p><div><p style="letter-spacing:0.2px"><font face="georgia, serif" color="#000000"><span style="letter-spacing:normal"><b>Title:</b> </span><span style="letter-spacing:normal"><span class="gmail_default"></span></span><span style="letter-spacing:normal"><span class="gmail_default"></span></span><span style="letter-spacing:normal">An Egocentric Approach to Social AI</span></font></p><div><font face="georgia, serif" color="#000000"><b>Abstract: </b>While computer vision and NLP have made tremendous progress in extracting semantics from image, video, and textual data, our computational understanding of human social behavior is still in its infancy. Face-to-face communication, using a rich repertoire of visual, acoustic, and linguistic channels, is the foundation for all social interactions and relationships, as well as all other means of communication. Moreover, the acquisition of social skill in infancy is a critical developmental milestone, and its disruption in conditions such as autism has life-long consequences. The current state-of-the-art in AI consists largely of surface-level analyses, e.g., action detection, recognition, and prediction from video, or inferring the sentiment of an utterance via NLP. A major challenge is to leverage this recent progress and mount an attack on the core constructs of social interaction, such as joint attention, theory of mind, and social appraisals. A key hypothesis is that computational social understanding is enabled by an egocentric perspective, i.e. the capture of social signals from the perspective of each social partner via head- and body-worn sensors. This perspective is well-aligned with existing commercial efforts in Augmented and Virtual Reality, and with the literature on child development.</font></div><p class="MsoNormal"><font face="georgia, serif" color="#000000"><u></u> <u></u></font></p><p class="MsoNormal"><font face="georgia, serif" color="#000000">In this talk, I will provide background on egocentric perception and summarize our current progress towards egocentric social understanding. A key technical challenge is the inference of social attention from multimodal sensor data. Inferential attention is based on the analysis of video recordings of naturalistic interactions using machine learning models, without the use of eye tracking. I will review recent progress on estimating visual and auditory attention from egocentric data. I will also describe our efforts to develop a benchmark dataset for multimodal social understanding, based on multi-person social deduction games such as One Night Werewolf. A key motivation for our work is the modeling of social attention as a means to improve the diagnosis and treatment of autism, and I will review our progress towards this goal. This is joint work with collaborators at UIUC, Georgia Tech, Weill-Cornell, and Meta Reality Labs Research.</font></p><div><font face="georgia, serif" color="#000000"><br></font></div><div><font face="georgia, serif"><font color="#000000"><b>Short Bio</b>: <span class="gmail_default" style="font-family:georgia,serif;font-size:small"> </span></font>James M. Rehg is the Director of the Health Care Engineering Systems Center in CSL. Dr. Rehg also holds appointments as a Founder Professor in the Siebel School of Computing and Data Science and in the Department of Industrial and Enterprise Systems Engineering at UIUC. He received his Ph.D. from CMU in 1995 and worked at the Cambridge Research Lab of DEC (and then Compaq) from 1995-2001, where he managed the computer vision research group. He was a professor in the College of Computing at Georgia Tech from 2001-2022. He received an NSF CAREER award in 2001 and a Raytheon Faculty Fellowship from Georgia Tech in 2005. He and his students have received best student paper awards at ICML 2005, BMVC 2010 and 2022, Mobihealth 2014, and Face and Gesture 2015, and a Method of the Year Award from the journal Nature Methods. Dr. Rehg served as the Program co-Chair for ACCV 2012 and CVPR 2017 and General co-Chair for CVPR 2009. He has authored more than 200 peer-reviewed scientific papers and holds 30 issued US patents. His research interests include computer vision, machine learning, and mobile and computational health (<a href="https://rehg.org">https://rehg.org</a>). Dr. Rehg was the lead PI on an NSF Expedition to develop the science and technology of Behavioral Imaging, the measurement and analysis of social and communicative behavior using multi-modal sensing, with applications to developmental conditions such as autism. He is currently the Deputy Director and TR&D1 Lead for the mHealth Center for Discovery, Optimization, and Translation of Temporally-Precise Interventions (mDOT), which is developing novel on-body sensing and predictive analytics for improving health outcomes (<a href="https://mdot.md2k.org/">https://mdot.md2k.org/</a>)<span class="gmail_default" style="font-size:small">. </span></font></div><font face="georgia, serif" color="#000000"><br><b>Host: </b><a href="mailto:greg@ttic.edu" target="_blank"><b>Greg Shakhnarovich</b></a></font></div></div><br clear="all"></div><div><br></div><span class="gmail_signature_prefix">-- </span><br><div dir="ltr" class="gmail_signature" data-smartmail="gmail_signature"><div dir="ltr"><b style="background-color:rgb(255,255,255)"><font color="#3d85c6">Brandie Jones </font></b><div><div><div><font color="#3d85c6"><b><i>Executive </i></b></font><b style="color:rgb(61,133,198)"><i>Administrative Assistant</i></b></div></div><div><span style="background-color:rgb(255,255,255)"><font color="#3d85c6">Toyota Technological Institute</font></span></div><div><span style="background-color:rgb(255,255,255)"><font color="#3d85c6">6045 S. Kenwood Avenue</font></span></div><div><span style="background-color:rgb(255,255,255)"><font color="#3d85c6">Chicago, IL 60637</font></span></div></div><div><span style="background-color:rgb(255,255,255)"><font color="#3d85c6"><a href="http://www.ttic.edu" target="_blank">www.ttic.edu</a> </font></span></div><div><span style="background-color:rgb(255,255,255)"><font color="#3d85c6"><br></font></span></div></div></div></div>
</div>