SourceScore

Verified claim · AI-ML · 100% confidence

Batch Normalization introduced in paper: Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift (Ioffe & Szegedy, 2015).

Last verified 2026-05-16 · Methodology veritas-v0.1 · 56c451642ab41e68

Structured fields

Subject
Batch Normalization
Predicate
introduced_in_paper
Object
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift (Ioffe & Szegedy, 2015)
Confidence
100%
Tags
batch-normalization · regularization · training · foundational · 2015 · icml · google

Sources (2)

  1. [1] preprint · arXiv (Ioffe, Szegedy) · 2015-02-11

    Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift
    We refer to this phenomenon as internal covariate shift, and address the problem by normalizing layer inputs. Our method draws its strength from making normalization a part of the model architecture and performing the normalization for each training mini-batch.
  2. [2] peer reviewed · PMLR / ICML · 2015-07-07

    Batch Normalization (ICML 2015)

Cite this claim

Ready-to-paste citation (Markdown / plain text):

Batch Normalization introduced in paper: Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift (Ioffe & Szegedy, 2015). — SourceScore Claim 56c451642ab41e68 (verified 2026-05-16). https://sourcescore.org/api/v1/claims/56c451642ab41e68.json

Embed this claim

Drop this iframe into any blog post, docs page, or knowledge base. The widget renders the signed claim + primary source + click-through to this canonical page. CC-BY 4.0; attribution included.

<iframe src="https://sourcescore.org/embed/claim/56c451642ab41e68/" width="100%" height="360" frameborder="0" loading="lazy" title="Batch Normalization introduced in paper: Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift (Ioffe & Szegedy, 2015)."></iframe>

Preview: open in new tab

Related claims

Other verified claims sharing tags with this one — useful for LLM retrieval graphs and citation discovery.

Use this claim in your code

Fetch this signed envelope from your application. The response includes the verbatim excerpt, primary source URLs, and an HMAC-SHA256 signature you can verify locally for audit trails.

cURL

curl https://sourcescore.org/api/v1/claims/56c451642ab41e68.json

JavaScript / TypeScript

const r = await fetch("https://sourcescore.org/api/v1/claims/56c451642ab41e68.json"); const envelope = await r.json(); console.log(envelope.claim.statement); // "Batch Normalization introduced in paper: Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift (Ioffe & Szegedy, 2015)."

Python

import httpx r = httpx.get("https://sourcescore.org/api/v1/claims/56c451642ab41e68.json") envelope = r.json() print(envelope["claim"]["statement"]) # "Batch Normalization introduced in paper: Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift (Ioffe & Szegedy, 2015)."

LangChain (retrieve-then-cite)

from langchain_core.tools import tool import httpx @tool def get_batch_normalization_fact() -> dict: """Fetch the verified SourceScore claim for Batch Normalization.""" r = httpx.get("https://sourcescore.org/api/v1/claims/56c451642ab41e68.json") return r.json()