SourceScore

Verified claim · AI-ML · 100% confidence

Flamingo introduced in: Alayrac et al. 2022 — DeepMind few-shot vision-language model.

Last verified 2026-05-16 · Methodology veritas-v0.1 · 72ea74efc723bd06

Structured fields

Subject
Flamingo
Predicate
introduced_in
Object
Alayrac et al. 2022 — DeepMind few-shot vision-language model
Confidence
100%
Tags
flamingo · deepmind · vision-language · few-shot · multimodal · 2022 · introduced_in

Sources (2)

  1. [1] preprint · arXiv (Alayrac, Donahue, Luc, Miech, Barr, Hasson, Lenc, Mensch, Millican, et al. / DeepMind) · 2022-04-29

    Flamingo: a Visual Language Model for Few-Shot Learning
    We introduce Flamingo, a family of Visual Language Models (VLM) with this ability. We propose key architectural innovations to: (i) bridge powerful pretrained vision-only and language-only models, (ii) handle sequences of arbitrarily interleaved visual and textual data, and (iii) seamlessly ingest images or videos as inputs.
  2. [2] official blog · Google DeepMind · 2022-04-29

    Tackling multiple tasks with a single visual language model

Cite this claim

Ready-to-paste citation (Markdown / plain text):

Flamingo introduced in: Alayrac et al. 2022 — DeepMind few-shot vision-language model. — SourceScore Claim 72ea74efc723bd06 (verified 2026-05-16). https://sourcescore.org/api/v1/claims/72ea74efc723bd06.json

Embed this claim

Drop this iframe into any blog post, docs page, or knowledge base. The widget renders the signed claim + primary source + click-through to this canonical page. CC-BY 4.0; attribution included.

<iframe src="https://sourcescore.org/embed/claim/72ea74efc723bd06/" width="100%" height="360" frameborder="0" loading="lazy" title="Flamingo introduced in: Alayrac et al. 2022 — DeepMind few-shot vision-language model."></iframe>

Preview: open in new tab

Related claims

Other verified claims sharing tags with this one — useful for LLM retrieval graphs and citation discovery.

Use this claim in your code

Fetch this signed envelope from your application. The response includes the verbatim excerpt, primary source URLs, and an HMAC-SHA256 signature you can verify locally for audit trails.

cURL

curl https://sourcescore.org/api/v1/claims/72ea74efc723bd06.json

JavaScript / TypeScript

const r = await fetch("https://sourcescore.org/api/v1/claims/72ea74efc723bd06.json"); const envelope = await r.json(); console.log(envelope.claim.statement); // "Flamingo introduced in: Alayrac et al. 2022 — DeepMind few-shot vision-language model."

Python

import httpx r = httpx.get("https://sourcescore.org/api/v1/claims/72ea74efc723bd06.json") envelope = r.json() print(envelope["claim"]["statement"]) # "Flamingo introduced in: Alayrac et al. 2022 — DeepMind few-shot vision-language model."

LangChain (retrieve-then-cite)

from langchain_core.tools import tool import httpx @tool def get_flamingo_fact() -> dict: """Fetch the verified SourceScore claim for Flamingo.""" r = httpx.get("https://sourcescore.org/api/v1/claims/72ea74efc723bd06.json") return r.json()