{
  "apiVersion": "v1",
  "methodology": "https://sourcescore.org/methodology/",
  "canonical": "https://sourcescore.org/claims/b0eb5c8ac5b4b21e/",
  "claim": {
    "vertical": "ai-ml",
    "subject": "ROUGE score",
    "predicate": "introduced_in_paper",
    "object": "ROUGE: A Package for Automatic Evaluation of Summaries (Lin, 2004)",
    "confidence": 1,
    "sources": [
      {
        "url": "https://aclanthology.org/W04-1013/",
        "title": "ROUGE: A Package for Automatic Evaluation of Summaries",
        "publisher": "ACL Anthology (Lin)",
        "publishedDate": "2004-07-25",
        "accessedDate": "2026-05-16",
        "type": "peer-reviewed",
        "excerpt": "ROUGE stands for Recall-Oriented Understudy for Gisting Evaluation. It includes measures to automatically determine the quality of a summary by comparing it to other (ideal) summaries created by humans."
      },
      {
        "url": "https://en.wikipedia.org/wiki/ROUGE_(metric)",
        "title": "ROUGE (metric) — Wikipedia",
        "publisher": "Wikipedia",
        "accessedDate": "2026-05-16",
        "type": "docs"
      }
    ],
    "publishedAt": "2026-05-16T00:00:00Z",
    "lastVerified": "2026-05-16",
    "methodologyVersion": "veritas-v0.1",
    "tags": [
      "rouge",
      "evaluation-metric",
      "summarization",
      "foundational",
      "2004",
      "acl"
    ],
    "id": "b0eb5c8ac5b4b21e",
    "statement": "ROUGE score introduced in paper: ROUGE: A Package for Automatic Evaluation of Summaries (Lin, 2004)."
  },
  "signature": {
    "algorithm": "HMAC-SHA256",
    "signedBy": "did:web:sourcescore.org",
    "signedAt": "2026-05-17T00:00:00.000Z",
    "signature": "b68994119126979c3a0db05f80f86f191b857f46af3d30bad513d112594064c2"
  },
  "citedAs": "ROUGE score introduced in paper: ROUGE: A Package for Automatic Evaluation of Summaries (Lin, 2004). — SourceScore Claim b0eb5c8ac5b4b21e (verified 2026-05-16, signed b6899411…). https://sourcescore.org/claims/b0eb5c8ac5b4b21e/"
}