{
  "apiVersion": "v1",
  "methodology": "https://sourcescore.org/methodology/",
  "canonical": "https://sourcescore.org/claims/a9ab1ec12062f7ae/",
  "claim": {
    "vertical": "ai-ml",
    "subject": "GPTQ",
    "predicate": "introduced_in",
    "object": "Frantar et al. 2022 — accurate post-training quantization for GPT models",
    "confidence": 1,
    "sources": [
      {
        "url": "https://arxiv.org/abs/2210.17323",
        "title": "GPTQ: Accurate Post-Training Quantization for Generative Pre-trained Transformers",
        "publisher": "arXiv (Frantar, Ashkboos, Hoefler, Alistarh / IST Austria)",
        "publishedDate": "2022-10-31",
        "accessedDate": "2026-05-16",
        "type": "preprint",
        "excerpt": "In this paper, we present a new one-shot weight quantization method based on approximate second-order information, that is both highly-accurate and highly-efficient. Specifically, GPTQ can quantize GPT models with 175 billion parameters in approximately four GPU hours, reducing the bitwidth down to 3 or 4 bits per weight, with negligible accuracy degradation relative to the uncompressed baseline."
      },
      {
        "url": "https://github.com/IST-DASLab/gptq",
        "title": "GPTQ — official IST-DASLab GitHub repository",
        "publisher": "IST Austria DAS Lab",
        "publishedDate": "2022-10-31",
        "accessedDate": "2026-05-16",
        "type": "github-release"
      }
    ],
    "publishedAt": "2026-05-16T00:00:00Z",
    "lastVerified": "2026-05-16",
    "methodologyVersion": "veritas-v0.1",
    "tags": [
      "gptq",
      "quantization",
      "ist-austria",
      "inference",
      "post-training",
      "2022",
      "introduced_in"
    ],
    "id": "a9ab1ec12062f7ae",
    "statement": "GPTQ introduced in: Frantar et al. 2022 — accurate post-training quantization for GPT models."
  },
  "signature": {
    "algorithm": "HMAC-SHA256",
    "signedBy": "did:web:sourcescore.org",
    "signedAt": "2026-05-17T00:00:00.000Z",
    "signature": "d38c8b26a7268e643c48f4022de510624e40257040345c7e9d639c8595a18bcd"
  },
  "citedAs": "GPTQ introduced in: Frantar et al. 2022 — accurate post-training quantization for GPT models. — SourceScore Claim a9ab1ec12062f7ae (verified 2026-05-16, signed d38c8b26…). https://sourcescore.org/claims/a9ab1ec12062f7ae/"
}