{
  "apiVersion": "v1",
  "methodology": "https://sourcescore.org/methodology/",
  "canonical": "https://sourcescore.org/claims/aa113b5e61d5c214/",
  "claim": {
    "vertical": "ai-ml",
    "subject": "GLUE benchmark",
    "predicate": "introduced_in_paper",
    "object": "GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding (Wang et al., 2018)",
    "confidence": 1,
    "sources": [
      {
        "url": "https://arxiv.org/abs/1804.07461",
        "title": "GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding",
        "publisher": "arXiv (Wang, Singh, Michael, Hill, Levy, Bowman)",
        "publishedDate": "2018-04-20",
        "accessedDate": "2026-05-16",
        "type": "preprint",
        "excerpt": "We introduce the General Language Understanding Evaluation (GLUE) benchmark, a collection of tools for evaluating the performance of models across a diverse set of existing NLU tasks."
      },
      {
        "url": "https://gluebenchmark.com/",
        "title": "GLUE — official site",
        "publisher": "NYU",
        "accessedDate": "2026-05-16",
        "type": "official-blog"
      }
    ],
    "publishedAt": "2026-05-16T00:00:00Z",
    "lastVerified": "2026-05-16",
    "methodologyVersion": "veritas-v0.1",
    "tags": [
      "glue",
      "benchmark",
      "evaluation",
      "foundational",
      "2018"
    ],
    "id": "aa113b5e61d5c214",
    "statement": "GLUE benchmark introduced in paper: GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding (Wang et al., 2018)."
  },
  "signature": {
    "algorithm": "HMAC-SHA256",
    "signedBy": "did:web:sourcescore.org",
    "signedAt": "2026-05-16T00:00:00.000Z",
    "signature": "cd51405c54aae0d99b4ddf313d9f5fed8e060607e4837519243c7bc71c73c28c"
  },
  "citedAs": "GLUE benchmark introduced in paper: GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding (Wang et al., 2018). — SourceScore Claim aa113b5e61d5c214 (verified 2026-05-16, signed cd51405c…). https://sourcescore.org/claims/aa113b5e61d5c214/"
}