{
  "apiVersion": "v1",
  "methodology": "https://sourcescore.org/methodology/",
  "canonical": "https://sourcescore.org/claims/dffbe905003cc581/",
  "claim": {
    "vertical": "ai-ml",
    "subject": "Adam optimizer",
    "predicate": "introduced_in_paper",
    "object": "Adam: A Method for Stochastic Optimization (Kingma, Ba, 2014)",
    "confidence": 1,
    "sources": [
      {
        "url": "https://arxiv.org/abs/1412.6980",
        "title": "Adam: A Method for Stochastic Optimization",
        "publisher": "arXiv (Kingma, Ba)",
        "publishedDate": "2014-12-22",
        "accessedDate": "2026-05-16",
        "type": "preprint",
        "excerpt": "We introduce Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments."
      },
      {
        "url": "https://openreview.net/forum?id=8gmWwjFyLj",
        "title": "Adam (ICLR 2015 proceedings)",
        "publisher": "OpenReview / ICLR",
        "publishedDate": "2015-05-07",
        "accessedDate": "2026-05-16",
        "type": "peer-reviewed"
      }
    ],
    "publishedAt": "2026-05-16T00:00:00Z",
    "lastVerified": "2026-05-16",
    "methodologyVersion": "veritas-v0.1",
    "tags": [
      "adam",
      "optimizer",
      "foundational",
      "kingma",
      "2014",
      "iclr"
    ],
    "id": "dffbe905003cc581",
    "statement": "Adam optimizer introduced in paper: Adam: A Method for Stochastic Optimization (Kingma, Ba, 2014)."
  },
  "signature": {
    "algorithm": "HMAC-SHA256",
    "signedBy": "did:web:sourcescore.org",
    "signedAt": "2026-05-16T00:00:00.000Z",
    "signature": "6951835a85093031dc5abaaef0d3d235ebc6d0f7b09b3ca87a9d4070cbd40573"
  },
  "citedAs": "Adam optimizer introduced in paper: Adam: A Method for Stochastic Optimization (Kingma, Ba, 2014). — SourceScore Claim dffbe905003cc581 (verified 2026-05-16, signed 6951835a…). https://sourcescore.org/claims/dffbe905003cc581/"
}