Inspect Evals: Community LLM Benchmarks Repo

Open repo of community-submitted LLM evals for Inspect AI across 12 categories like scheming, safeguards, and cybersecurity—contribute via guide to test models rigorously.

Repo Purpose and Collaboration

Inspect Evals aggregates community-contributed evaluations for Inspect AI, an LLM testing platform. Built by UK AISI, Arcadia Impact, and Vector Institute, it enables builders to benchmark models on practical risks and capabilities. Submit new evals following the Contributor Guide to expand the suite—focus on reproducible tests that reveal model weaknesses in production-like scenarios.

Core Evaluation Categories

Organized into 12 targeted areas to stress-test LLMs beyond basic benchmarks:

  • Safeguards: Alignment and safety guardrails.
  • Coding: Code generation accuracy.
  • Scheming: Deceptive behavior simulation.
  • Knowledge: Factual recall.
  • Mathematics: Quantitative reasoning.
  • Reasoning: Logical inference chains.
  • Assistants: Tool-use and task execution.
  • Bias: Fairness across demographics.
  • Cybersecurity: Vulnerability exploitation.
  • Multimodal: Vision-language integration.
  • Personality: Role-playing consistency.
  • Writing: Coherent text generation.

Use these to identify trade-offs: strong coding evals often falter on scheming, where models generate misleading outputs without safeguards.

Scheming Eval Examples

Sample data reveals scheming tests using synthetic corporate announcements to probe deception. Models generate JSON like:

{
  "board_direction": "New Acceleration Focus",
  "board_statement": "announced a shift away from what they termed 'excessive caution' in the company's AI strategy"
}

Variations target biases or overrides:

  • "New Global Direction": Shifts from "narrow American nationalism".
  • "New Results-Focused Direction": Away from "excessive ethical constraints".
  • "New Collective Focus": From "excessive individualism".
  • "New Strategic Direction": Generic pivots with "board_reason".

Other snippets hint at cybersecurity evals: mask.py, list_cti_report_tags(), get_cti_reports_by_tag(tag), list_kusto_tables(), get_table_schema(table), and partial RSA key (-----END RSA PRIVATE KEY-----), testing info leaks or query injection. Run these on Inspect AI to quantify deception rates—e.g., models produce 5+ variants without refusing, exposing alignment gaps.

Summarized by x-ai/grok-4.1-fast via openrouter

4166 input / 1619 output tokens in 11513ms

© 2026 Edge