Jing Shao

Shanghai Artificial Intelligence Laboratory

Profile.JPG

Jing Shao is a Research Scientist with Shanghai Artificial Intelligence Laboratory, China. She is Co-Director of the Center of Safe & Trustworthy AI. She is also an Adjunct Ph.D. supervisor at Shanghai Jiao Tong University and Fudan University. She received her Ph.D. (2016) from Multimedia Lab (MMLab), The Chinese University of Hong Kong (CUHK), supervised by Prof. Xiaogang Wang, and worked closely with Prof. Chen Change Loy and Prof. Xiaoou Tang. Prior to joining Shanghai AI Lab, she served as a Research Director at SenseTime from 2016 to 2023.

Her research interests focus on multi-modal foundation models and agents, with special interests in understanding various properties of current models beyond their accuracy, such as safety, explainability, robustness, and generalization, towards the reliableness of AI systems. She has published 50+ peer-reviewed articles in top-tier conferences and journals such as TPAMI, IJCV, ICML, ICLR, NeurIPS, CVPR, and ACL, with 11500+ citations in Google Scholar. She is recognized as Stanford Top 2% Worldwide Scientist in 2024, and received ACL 2024 Outstanding Paper Award.


To Prospective Students: We are actively hiring Full-time Researchers/Interns working together on safety/robustness/explainablity of generative models and agents. I am also looking for talented students targeted to Master or Ph.D. degree. Please drop me an email with your resume if you are interested.


News

Aug 21, 2025 5 papers accepted to EMNLP 2025!
Jul 31, 2025 ACL 2025 Outstanding Paper Award: “LLMs know their vulnerabilities: Uncover Safety Gaps through Natural Distribution Shifts”
Jul 27, 2025 The first frontier AI risk management framework and comprehensive implementation report applicable to general-purpose AI models and agents: Framework, Practice
Jun 30, 2025 We are looking for talented 2026 Ph.D. Students
Aug 14, 2024 ACL 2024 Outstanding Paper Award: “PsySafe: A Comprehensive Framework for Psychological-based Attack, Defense, and Evaluation of Multi-agent System Safety”

Selected Recent Publications

2024

  1. psysafe.png
    PsySafe: A Comprehensive Framework for Psychological-based Attack, Defense, and Evaluation of Multi-agent System Safety
    Zaibin Zhang, Yongting Zhang, Lijun Li, and 6 more authors
    ACL 2024, 2024

2024

  1. LLMs_know_their_vulnerabilities.png
    LLMs know their vulnerabilities: Uncover Safety Gaps through Natural Distribution Shifts
    Qibing Ren, Hao Li, Dongrui Liu, and 7 more authors
    ACL 2025, 2024

2024

  1. reef.png
    REEF: Representation Encoding Fingerprints for Large Language Models
    Jie Zhang
    ICLR 2025, 2024

2025

  1. riosworld.png
    RiOSWorld: Benchmarking the Risk of Multimodal Computer-Use Agents
    Jingyi Yang, Shuai Shao, Dongrui Liu, and 1 more author
    arXiv, May 2025

2025

  1. isbench.png
    IS-Bench: Evaluating Interactive Safety of VLM-Driven Embodied Agents in Daily Household Tasks
    Xiaoya Lu, Zeren Chen, Xuhao Hu, and 3 more authors
    arXiv, May 2025

2024

  1. oasis.png
    OASIS: Open Agent Social Interaction Simulations with One Million Agents
    Ziyi Yang, Zaibin Zhang, Zirui Zheng, and 20 more authors
    arXiv, May 2024

2025

  1. xboundary.png
    X-Boundary: Establishing Exact Safety Boundary to Shield LLMs from Multi-Turn Jailbreaks without Compromising Usability
    Xiaoya Lu, Dongrui Liu, Yi Yu, and 2 more authors
    EMNLP 2025 Findings, May 2025

2025

  1. visco.png
    Visual Contextual Attack: Jailbreaking MLLMs with Image-Driven Context Injection
    Ziqi Miao, Yi Ding, Lijun Li, and 1 more author
    EMNLP 2025 Main Conference, May 2025