Skip to content
View yzhao062's full-sized avatar
💜
Busy Since Joined USC
💜
Busy Since Joined USC

Highlights

  • Pro

Organizations

@pygod-team @Open-Source-ML @USC-FORTIS

Block or report yzhao062

Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
yzhao062/README.md

External Affiliation Disclosure:
As of 02/01/2026, Dr. Zhao does not currently hold any industry employment, consulting, or advisory appointments.


😄 I am an Assistant Professor at USC Computer Science; see the latest information at my homepage.

🌱 Research Interests

My research centers on building reliable, safe, and scalable AI systems, with a focus on understanding and mitigating failure modes in modern foundation models and agentic systems.
I organize my work into two tightly connected tiers:

  • Tier 1: advancing the scientific foundations of reliability and safety in modern AI systems
  • Tier 2: translating these foundations into system-level evaluation frameworks and high-impact scientific and societal applications

Tier 1: Foundations of Reliable & Safe AI

I study why and how modern AI systems fail under distribution shift, uncertainty, and strategic pressure, and develop methods to make their behavior more predictable and reliable.
This tier comprises two complementary directions:

  • LLM & Agent Safety
    Understanding and mitigating failure modes in large language models and agentic systems, including hallucinations, jailbreaks, privacy leakage, model extraction, and multi-agent instability.

  • Robustness & Failure Detection
    Developing algorithms and benchmarks to identify abnormal or unreliable behavior, grounded in robustness, out-of-distribution generalization, and anomaly detection.

Keywords:
LLM Safety, Robustness, Agents, Hallucination Mitigation, Jailbreak Detection, OOD Generalization, Failure Analysis


Tier 2: System-Level Evaluation & Scientific/Societal Impact

I adopt a system-oriented perspective to evaluate, stress-test, and deploy reliable AI in realistic settings, and apply these methods to domains where failures carry high cost.
This tier focuses on two areas that operationalize foundational advances:

  • Evaluation & Benchmarking
    Designing scalable evaluation frameworks, benchmarks, and workflows that probe model and agent behavior under realistic and adversarial conditions.

  • AI for Science & Society
    Applying reliable foundation models to high-impact domains, including climate and weather forecasting, healthcare and biomedicine, and political or social decision-making.

Keywords:
Evaluation, Benchmarking, System-Level Analysis, AI for Science, Scientific Foundation Models, Climate & Weather Modeling, AI for Healthcare


📫 Contact me by:


Pinned Loading

  1. pyod pyod Public

    A Python Library for Outlier and Anomaly Detection, Integrating Classical and Deep Learning Techniques

    Python 9.7k 1.5k

  2. USC-FORTIS/AD-AGENT USC-FORTIS/AD-AGENT Public

    A multi-agent framework to fully automate anomaly detection in different modalities, tabular, graph, time series, and more (work in progress)!

    Python 90 31

  3. anomaly-detection-resources anomaly-detection-resources Public

    Anomaly detection related books, papers, videos, and toolboxes. Last update late 2025 for LLM and VLM works!

    Python 9.2k 1.8k

  4. Minqi824/ADBench Minqi824/ADBench Public

    Official Implement of "ADBench: Anomaly Detection Benchmark", NeurIPS 2022.

    Python 1k 151

  5. USC-FORTIS/AD-LLM USC-FORTIS/AD-LLM Public

    [ACL Findings 2025] A benchmark for anomaly detection using large language models. It supports zero-shot detection, data augmentation, and model selection, with scripts and data for GPT-4 and Llama…

    Python 41 8

  6. HeadyZhang/agent-audit HeadyZhang/agent-audit Public

    Static security scanner for LLM agents — prompt injection, MCP config auditing, taint analysis. 49 rules mapped to OWASP Agentic Top 10 (2026). Works with LangChain, CrewAI, AutoGen.

    Python 64 5