Skip to content

Contribute to the Agentic Benchmark Checklist

Upholding the validity of agentic benchmarks requires effort from the broader scientific community. If you’re passionate about reliable evaluation in AI, we’d love your help.

Here’s some ways to get involved:

  1. Apply the checklist to an existing benchmark - submit here.

  2. Contribute proof-of-concept exploits and fixes for those exploits in our repo.

  3. Give feedback on the checklist itself here.