{
  "apiVersion": "v1",
  "methodology": "https://sourcescore.org/methodology/",
  "canonical": "https://sourcescore.org/claims/85a3ca745eaf4ee0/",
  "claim": {
    "vertical": "ai-ml",
    "subject": "CLIP (Contrastive Language-Image Pretraining)",
    "predicate": "introduced_in_paper",
    "object": "Learning Transferable Visual Models From Natural Language Supervision (Radford et al., 2021)",
    "confidence": 1,
    "sources": [
      {
        "url": "https://arxiv.org/abs/2103.00020",
        "title": "Learning Transferable Visual Models From Natural Language Supervision",
        "publisher": "arXiv (Radford et al., OpenAI)",
        "publishedDate": "2021-02-26",
        "accessedDate": "2026-05-16",
        "type": "preprint",
        "excerpt": "We demonstrate that the simple pre-training task of predicting which caption goes with which image is an efficient and scalable way to learn SOTA image representations from scratch on a dataset of 400 million (image, text) pairs collected from the internet."
      },
      {
        "url": "https://openai.com/research/clip",
        "title": "CLIP: Connecting text and images",
        "publisher": "OpenAI",
        "publishedDate": "2021-01-05",
        "accessedDate": "2026-05-16",
        "type": "official-blog"
      }
    ],
    "publishedAt": "2026-05-16T00:00:00Z",
    "lastVerified": "2026-05-16",
    "methodologyVersion": "veritas-v0.1",
    "tags": [
      "clip",
      "multimodal",
      "vision",
      "radford",
      "2021",
      "openai"
    ],
    "id": "85a3ca745eaf4ee0",
    "statement": "CLIP (Contrastive Language-Image Pretraining) introduced in paper: Learning Transferable Visual Models From Natural Language Supervision (Radford et al., 2021)."
  },
  "signature": {
    "algorithm": "HMAC-SHA256",
    "signedBy": "did:web:sourcescore.org",
    "signedAt": "2026-05-16T00:00:00.000Z",
    "signature": "5b43ae5f036bb684ea2221054725631789590ad598d25b3ac0592b92ba5be6fb"
  },
  "citedAs": "CLIP (Contrastive Language-Image Pretraining) introduced in paper: Learning Transferable Visual Models From Natural Language Supervision (Radford et al., 2021). — SourceScore Claim 85a3ca745eaf4ee0 (verified 2026-05-16, signed 5b43ae5f…). https://sourcescore.org/claims/85a3ca745eaf4ee0/"
}