{
  "apiVersion": "v1",
  "methodology": "https://sourcescore.org/methodology/",
  "canonical": "https://sourcescore.org/claims/79be9b25cd64f250/",
  "claim": {
    "vertical": "ai-ml",
    "subject": "Codex",
    "predicate": "introduced_in_paper",
    "object": "Evaluating Large Language Models Trained on Code (Chen et al., 2021)",
    "confidence": 1,
    "sources": [
      {
        "url": "https://arxiv.org/abs/2107.03374",
        "title": "Evaluating Large Language Models Trained on Code",
        "publisher": "arXiv (Chen, Tworek, Jun, Yuan, Pinto, Kaplan, et al.)",
        "publishedDate": "2021-07-07",
        "accessedDate": "2026-05-16",
        "type": "preprint",
        "excerpt": "We introduce Codex, a GPT language model fine-tuned on publicly available code from GitHub, and study its Python code-writing capabilities."
      },
      {
        "url": "https://openai.com/blog/openai-codex",
        "title": "OpenAI Codex",
        "publisher": "OpenAI",
        "publishedDate": "2021-08-10",
        "accessedDate": "2026-05-16",
        "type": "official-blog"
      }
    ],
    "publishedAt": "2026-05-16T00:00:00Z",
    "lastVerified": "2026-05-16",
    "methodologyVersion": "veritas-v0.1",
    "tags": [
      "codex",
      "code-generation",
      "openai",
      "foundational",
      "2021"
    ],
    "id": "79be9b25cd64f250",
    "statement": "Codex introduced in paper: Evaluating Large Language Models Trained on Code (Chen et al., 2021)."
  },
  "signature": {
    "algorithm": "HMAC-SHA256",
    "signedBy": "did:web:sourcescore.org",
    "signedAt": "2026-05-16T00:00:00.000Z",
    "signature": "4f867014c87eb4d79ceaea9219eb616d671f986917e5917d7684fbddd6182122"
  },
  "citedAs": "Codex introduced in paper: Evaluating Large Language Models Trained on Code (Chen et al., 2021). — SourceScore Claim 79be9b25cd64f250 (verified 2026-05-16, signed 4f867014…). https://sourcescore.org/claims/79be9b25cd64f250/"
}