{
  "apiVersion": "v1",
  "methodology": "https://sourcescore.org/methodology/",
  "canonical": "https://sourcescore.org/claims/8cdc42759d98b531/",
  "claim": {
    "vertical": "ai-ml",
    "subject": "Hugging Face Transformers library",
    "predicate": "publicly_released_on",
    "object": "2018-10-04 — originally pytorch-pretrained-bert",
    "confidence": 1,
    "sources": [
      {
        "url": "https://github.com/huggingface/transformers/releases/tag/0.1.2",
        "title": "transformers v0.1.2 — earliest tagged release",
        "publisher": "Hugging Face",
        "publishedDate": "2018-11-15",
        "accessedDate": "2026-05-16",
        "type": "github-release",
        "excerpt": "PyTorch implementation of Google AI's BERT (Bidirectional Encoder Representations from Transformers) with Python script to load any pre-trained model."
      },
      {
        "url": "https://arxiv.org/abs/1910.03771",
        "title": "HuggingFace's Transformers: State-of-the-art Natural Language Processing",
        "publisher": "arXiv (Wolf et al.)",
        "publishedDate": "2019-10-09",
        "accessedDate": "2026-05-16",
        "type": "preprint"
      }
    ],
    "publishedAt": "2026-05-16T00:00:00Z",
    "lastVerified": "2026-05-16",
    "methodologyVersion": "veritas-v0.1",
    "tags": [
      "transformers",
      "hugging-face",
      "library",
      "open-source",
      "released_on",
      "2018"
    ],
    "id": "8cdc42759d98b531",
    "statement": "Hugging Face Transformers library publicly released on: 2018-10-04 — originally pytorch-pretrained-bert."
  },
  "signature": {
    "algorithm": "HMAC-SHA256",
    "signedBy": "did:web:sourcescore.org",
    "signedAt": "2026-05-17T00:00:00.000Z",
    "signature": "693fc61b80fe083d6db4e1fffaae28b783e2a5791e93e1b3168235869de4ec73"
  },
  "citedAs": "Hugging Face Transformers library publicly released on: 2018-10-04 — originally pytorch-pretrained-bert. — SourceScore Claim 8cdc42759d98b531 (verified 2026-05-16, signed 693fc61b…). https://sourcescore.org/claims/8cdc42759d98b531/"
}