{
  "apiVersion": "v1",
  "methodology": "https://sourcescore.org/methodology/",
  "canonical": "https://sourcescore.org/claims/2c6ddc094019890c/",
  "claim": {
    "vertical": "ai-ml",
    "subject": "llama.cpp",
    "predicate": "publicly_released_on",
    "object": "2023-03-10 by Georgi Gerganov",
    "confidence": 1,
    "sources": [
      {
        "url": "https://github.com/ggerganov/llama.cpp",
        "title": "llama.cpp — Inference of Meta's LLaMA model in pure C/C++",
        "publisher": "Georgi Gerganov / open-source community",
        "publishedDate": "2023-03-10",
        "accessedDate": "2026-05-16",
        "type": "github-release",
        "excerpt": "Inference of Meta's LLaMA model (and others) in pure C/C++. The main goal of llama.cpp is to enable LLM inference with minimal setup and state-of-the-art performance on a wide range of hardware - locally and in the cloud."
      },
      {
        "url": "https://github.com/ggerganov/llama.cpp/releases/tag/master-26c0846",
        "title": "llama.cpp first tagged release",
        "publisher": "Georgi Gerganov",
        "publishedDate": "2023-03-10",
        "accessedDate": "2026-05-16",
        "type": "github-release"
      }
    ],
    "publishedAt": "2026-05-16T00:00:00Z",
    "lastVerified": "2026-05-16",
    "methodologyVersion": "veritas-v0.1",
    "tags": [
      "llama-cpp",
      "ggerganov",
      "inference",
      "open-source",
      "cpp",
      "released_on",
      "2023"
    ],
    "id": "2c6ddc094019890c",
    "statement": "llama.cpp publicly released on: 2023-03-10 by Georgi Gerganov."
  },
  "signature": {
    "algorithm": "HMAC-SHA256",
    "signedBy": "did:web:sourcescore.org",
    "signedAt": "2026-05-17T00:00:00.000Z",
    "signature": "babac754573e97847eb2bb4c3b0f952c4aa3ebd21acc6f1f52cbdc02db345260"
  },
  "citedAs": "llama.cpp publicly released on: 2023-03-10 by Georgi Gerganov. — SourceScore Claim 2c6ddc094019890c (verified 2026-05-16, signed babac754…). https://sourcescore.org/claims/2c6ddc094019890c/"
}