SourceScore

Verified claim · AI-ML · 100% confidence

HumanEval benchmark introduced in paper: Evaluating Large Language Models Trained on Code (Chen et al., 2021).

Last verified 2026-05-16 · Methodology veritas-v0.1 · 71ec42731d2c9e0c

Structured fields

Subject
HumanEval benchmark
Predicate
introduced_in_paper
Object
Evaluating Large Language Models Trained on Code (Chen et al., 2021)
Confidence
100%
Tags
humaneval · benchmark · codex · openai · chen · 2021 · code-generation

Sources (2)

  1. [1] preprint · arXiv (Chen et al., OpenAI) · 2021-07-07

    Evaluating Large Language Models Trained on Code
    We introduce Codex, a GPT language model fine-tuned on publicly available code from GitHub, and study its Python code-writing capabilities.
  2. [2] github release · OpenAI · 2021-07-07

    openai/human-eval repository

Cite this claim

Ready-to-paste citation (Markdown / plain text):

HumanEval benchmark introduced in paper: Evaluating Large Language Models Trained on Code (Chen et al., 2021). — SourceScore Claim 71ec42731d2c9e0c (verified 2026-05-16). https://sourcescore.org/api/v1/claims/71ec42731d2c9e0c.json

Embed this claim

Drop this iframe into any blog post, docs page, or knowledge base. The widget renders the signed claim + primary source + click-through to this canonical page. CC-BY 4.0; attribution included.

<iframe src="https://sourcescore.org/embed/claim/71ec42731d2c9e0c/" width="100%" height="360" frameborder="0" loading="lazy" title="HumanEval benchmark introduced in paper: Evaluating Large Language Models Trained on Code (Chen et al., 2021)."></iframe>

Preview: open in new tab

Related claims

Other verified claims sharing tags with this one — useful for LLM retrieval graphs and citation discovery.

Programmatic access

Fetch this claim with a signed envelope for verification:

curl https://sourcescore.org/api/v1/claims/71ec42731d2c9e0c.json

API docs · Pricing · Methodology JSON