🛠 Open Source / Software

I value practical and reproducible research. This page highlights open-source libraries, benchmarks, and system tools built by my group and collaborators. Many of these projects support AI auditing and assurance, AI safety and security, or demanding applications in science and society, from evaluation and tracing to policy enforcement and runtime guardrails. Several of these projects have been accepted into the Anthropic Claude for Open Source Program. For all repositories, see my GitHub profile.

Adoption & Recognition

PyOD — Python Library for Outlier Detection

9,800+ stars · 38M+ downloads · GitHub
TypeEvidenceSource
Space Agency Selected by ESA for OPS-SAT spacecraft telemetry benchmark (all 30 algorithms) Nature Sci. Data
Enterprise Walmart real-time pricing anomaly detection (1M+ daily updates) KDD 2019
Enterprise Databricks Kakapo framework for unsupervised outlier detection Databricks Blog
Enterprise IQVIA healthcare fraud detection (123K+ pharmacy claims) SUOD Paper
Education Featured in 5 books (Manning, O'Reilly, Apress, Routledge, IntechOpen) Manning
Education DataCamp course with dedicated chapter (19M+ platform learners) DataCamp
Enterprise Ericsson Anomaly Detection Framework (E-ADF) built on PyOD Ericsson Blog
Patents 6 patents cite PyOD/COPOD/ECOD (Ericsson, EU, US, China) Example

TrustLLM — Trustworthiness Benchmark for LLMs

620+ stars · ICML 2024 · GitHub
TypeEvidenceSource
U.S. Senate Cited in HSGAC "Hedge Fund Use of Artificial Intelligence" report (footnote 119) Senate PDF
U.S. DoD Listed in CDAO Generative AI Responsible AI Toolkit ai.mil PDF
NIST Named in NIST AI 100-2e2025 Section 3.6 "Benchmarks for AML Vulnerabilities" NIST PDF
Policy Official benchmark in all 3 editions of the FLI AI Safety Index (2024, 2025 x2) FLI Report · Indicator Sheet
National Lab Lawrence Livermore National Laboratory feature article LLNL
International Cited in International AI Safety Report 2026 (citation #881; led by Yoshua Bengio, 100+ experts, 30+ countries) Report
Media Featured by 机器之心 (Jiqizhixin) and 澎湃新闻 (The Paper) 机器之心 · 澎湃

TDC — Therapeutics Data Commons

1,200+ stars · NeurIPS 2021 · with Harvard & Stanford · GitHub
TypeEvidenceSource
Journal Published in Nature Chemical Biology (2022) Nature Chem. Bio.
University Harvard Medical School feature: "Can AI transform drug discovery?" HMS News
Science Press Phys.org syndication of Harvard article Phys.org
Industry Amazon Science feature article Amazon Science
Pharma Cited by researchers at AstraZeneca, Pfizer, Roche, Novartis, Merck, Sanofi, Eli Lilly Audit details

DoxBench — Geolocation Privacy Leakage Benchmark

ICLR 2026 · GitHub
TypeEvidenceSource
Policy Cited by Privacy International in "Nowhere to Hide? Privacy Risks and Policy Implications of AI Geolocation" (p.28, footnote 56) Report
Featured project: anywhere-agents - one config to rule all your AI agents: portable across every project and session, effective through curated writing, routing, and skills, and safer via a PreToolUse guard that blocks destructive Git and GitHub commands. Supports Claude Code and Codex today, with plans to grow. Links: GitHub | Star | Docs | PyPI | npm.
Featured project: agent-audit - an AI agent auditing and policy-checking tool with checks for OWASP Agentic Top 10 risks, taint-style flow analysis, and MCP configuration auditing. It sits mainly in AI Auditing & Assurance, with direct overlap with AI Safety & Security. In March 2026, it scanned 18,899 ClawHub skills and detected 13,947 vulnerabilities (including 1,996 BLOCK-level critical findings). Links: GitHub | Star | PyPI | ClawHub.
Featured project: Aegis - the open-source firewall for AI agents, with pre-execution blocking, human approvals, and tamper-evident audit trails. It sits mainly in AI Safety & Security, while its audit trails also support AI Auditing & Assurance in real deployments. Links: GitHub | Star | Preprint | Demo.
Sort by: