{
  "apiVersion": "v1",
  "methodology": "https://sourcescore.org/methodology/",
  "canonical": "https://sourcescore.org/claims/4335bf51bf0fc14f/",
  "claim": {
    "vertical": "ai-ml",
    "subject": "MoE Mixtral 8x22B",
    "predicate": "released_on",
    "object": "2024-04-10 by Mistral AI",
    "confidence": 1,
    "sources": [
      {
        "url": "https://mistral.ai/news/mixtral-8x22b/",
        "title": "Cheaper, Better, Faster, Stronger — Mixtral 8x22B",
        "publisher": "Mistral AI",
        "publishedDate": "2024-04-10",
        "accessedDate": "2026-05-16",
        "type": "official-blog",
        "excerpt": "We are excited to release Mixtral 8x22B, our latest open model. It sets a new standard for performance and efficiency within the AI community."
      },
      {
        "url": "https://huggingface.co/mistralai/Mixtral-8x22B-v0.1",
        "title": "Mixtral-8x22B-v0.1 — Hugging Face model card",
        "publisher": "Mistral AI",
        "publishedDate": "2024-04-10",
        "accessedDate": "2026-05-16",
        "type": "model-card"
      }
    ],
    "publishedAt": "2026-05-16T00:00:00Z",
    "lastVerified": "2026-05-16",
    "methodologyVersion": "veritas-v0.1",
    "tags": [
      "mixtral-8x22b",
      "mistral",
      "moe",
      "open-weights",
      "released_on",
      "2024"
    ],
    "id": "4335bf51bf0fc14f",
    "statement": "MoE Mixtral 8x22B released on: 2024-04-10 by Mistral AI."
  },
  "signature": {
    "algorithm": "HMAC-SHA256",
    "signedBy": "did:web:sourcescore.org",
    "signedAt": "2026-05-17T00:00:00.000Z",
    "signature": "c979e1d3e37e167f0ee8b8cdd6582aadf2d984565b5891b2a8bfe3dc75d9d7d2"
  },
  "citedAs": "MoE Mixtral 8x22B released on: 2024-04-10 by Mistral AI. — SourceScore Claim 4335bf51bf0fc14f (verified 2026-05-16, signed c979e1d3…). https://sourcescore.org/claims/4335bf51bf0fc14f/"
}