[Colloquium] CS Reminder (TODAY): AI+Systems Seminar- Bo Li

Abigail Santana via Colloquium colloquium at mailman.cs.uchicago.edu
Tue May 14 08:19:23 CDT 2024


Department of Computer Science Seminar

Bo Li

  *   Neubauer Associate Professor of Computer Science, Data Science

University of Chicago


Tuesday, May 14th
12:30pm - 2:00pm
In person: John Crerar Library 298

Zoom information: https://uchi-jcl.github.io/seminar

Title: Risk Assessment, Safety Alignment, and Guardrails for Generative Models

Abstract:
Large language models (LLMs) have garnered widespread attention due to their impressive performance across a range of applications. However, our understanding of the trustworthiness and risks of these models remains limited. The temptation to deploy proficient foundation models in sensitive domains like healthcare and finance, where errors carry significant consequences, underscores the need for rigorous safety evaluations, enhancement, and guarantees. Recognizing the urgent need for developing safe and beneficial AI, our recent research seeks to design a unified platform to evaluate the safety of LLMs from diverse perspectives such as toxicity, stereotype bias, adversarial robustness, OOD robustness, ethics, privacy, and fairness; enhance LLM safety through knowledge integration; and provide safety guardrail and certifications. In this talk, I will first outline our foundational principles for safety evaluation, detail our red teaming tactics, and share insights gleaned from applying our Decoding Trust platform to different models, such as proprietary and open-source models, as well as compressed models. Further, I will delve into our methods for enhancing model safety, such as hallucination mitigation. I will also explain how knowledge integration helps align models and prove that the RAG framework achieves provably lower conformal generation risks compared to vanilla LLMs. Finally, I will briefly discuss our robust guardrail framework for risk mitigation in practice.

Bio:
Dr. Bo Li is the Neubauer Associate Professor in the Department of Computer Science at the University of Chicago. She is the recipient of the IJCAI Computers and Thought Award, Alfred P. Sloan Research Fellowship, IEEE AI’s 10 to Watch, NSF CAREER Award, MIT Technology Review TR-35 Award, Dean’s Award for Excellence in Research, C.W. Gear Outstanding Faculty Award, Intel Rising Star Award, Symantec Research Labs Fellowship, Rising Star Award, Research Awards from Tech companies such as Amazon, Meta, Google, Intel, IBM, and eBay, and best paper awards at several top machine learning and security conferences. Her research focuses on both theoretical and practical aspects of trustworthy machine learning, which is at the intersection of machine learning, security, privacy, and game theory. She has designed several scalable frameworks for certifiably robust learning and privacy-preserving data publishing. Her work has been featured by several major publications and media outlets, including Nature, Wired, Fortune, and New York Times.

Her website is: https://aisecure.github.io







Abigail Santana
Business Assistant - Computer Science
The University of Chicago
5730 S Ellis Ave, JCL 212
Chicago, IL 60637
asantana01 at uchicago.edu
Office: 773-834-2750

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mailman.cs.uchicago.edu/pipermail/colloquium/attachments/20240514/d167bfae/attachment.html>


More information about the Colloquium mailing list