{
  "apiVersion": "v1",
  "methodology": "https://sourcescore.org/methodology/",
  "canonical": "https://sourcescore.org/claims/1ca2cc2864dfb376/",
  "claim": {
    "vertical": "ai-ml",
    "subject": "GPT-3",
    "predicate": "parameter_count",
    "object": "175000000000",
    "confidence": 1,
    "sources": [
      {
        "url": "https://arxiv.org/abs/2005.14165",
        "title": "Language Models are Few-Shot Learners",
        "publisher": "OpenAI / arXiv (Brown et al.)",
        "publishedDate": "2020-05-28",
        "accessedDate": "2026-05-16",
        "type": "preprint",
        "excerpt": "We train GPT-3, an autoregressive language model with 175 billion parameters, 10x more than any previous non-sparse language model."
      },
      {
        "url": "https://openai.com/research/gpt-3-apps",
        "title": "GPT-3 Powers the Next Generation of Apps",
        "publisher": "OpenAI",
        "publishedDate": "2021-03-25",
        "accessedDate": "2026-05-16",
        "type": "official-blog"
      }
    ],
    "publishedAt": "2026-05-16T00:00:00Z",
    "lastVerified": "2026-05-16",
    "methodologyVersion": "veritas-v0.1",
    "tags": [
      "gpt-3",
      "openai",
      "parameters",
      "175b",
      "brown",
      "2020"
    ],
    "id": "1ca2cc2864dfb376",
    "statement": "GPT-3 parameter count: 175000000000."
  },
  "signature": {
    "algorithm": "HMAC-SHA256",
    "signedBy": "did:web:sourcescore.org",
    "signedAt": "2026-05-16T00:00:00.000Z",
    "signature": "4e965acc621b4fef46723003cecf9d2b12a604a026d22bfcca75bb17669f5521"
  },
  "citedAs": "GPT-3 parameter count: 175000000000. — SourceScore Claim 1ca2cc2864dfb376 (verified 2026-05-16, signed 4e965acc…). https://sourcescore.org/claims/1ca2cc2864dfb376/"
}