Data Availability Sampling (DAS) is the cryptographic mechanism that allows Ethereum light nodes to verify that a block’s full data is available on the network — without downloading the entire block. It is the key enabling technology for full danksharding: Ethereum’s plan to scale blob transaction throughput to 64–128 blobs per block (from 6 in Pectra) without requiring validators or light nodes to download and store all that data themselves. DAS works by combining erasure coding (encoding data with redundancy so that any subset of the whole can reconstruct the full data) and random sampling (light nodes request a random handful of coded chunks; if they all respond, the full data is available with high probability). Without DAS, scaling blob throughput would force validators to download more data per block — defeating decentralization goals. With DAS, even light nodes can participate in availability verification.
The Data Availability Problem
Before understanding DAS, the underlying problem:
Block producers could withhold data. A malicious block proposer could:
- Propose a block header with a valid cryptographic commitment (e.g., a KZG commitment to “some data”)
- Never reveal the actual data
- Validators who can only verify the commitment header (not the underlying data) cannot detect the withholding
This matters for rollups: if an L2 sequences transactions and posts a blob, but the blob data is never actually made available (withheld by the blob poster), then L2 state cannot be reconstructed from L1 data. Users who need to exit the L2 would be stuck — fraud proofs can’t be computed without the data.
Solution requirement: A mechanism that makes it computationally hard to hide data once you’ve committed to it, while keeping data download requirements for validators minimal.
How DAS Works
The following sections cover this in detail.
Step 1: Erasure Coding
Before publishing blob data, it is erasure coded — encoded with 2x redundancy:
- Original data: N chunks
- Encoded data: 2N chunks
- Property: Any N of the 2N chunks are sufficient to reconstruct the original data
Erasure coding means that if an attacker withholds >50% of chunks, the data is unrecoverable — but any combination of ≥50% of chunks lets you decode everything.
Step 2: KZG Commitments over chunks
The encoded chunks are committed using KZG polynomial commitments:
- Each chunk has a verifiable commitment
- Light nodes can verify that a chunk is authentic (matches the blob’s root commitment) without downloading the full blob
Step 3: Random Sampling by Light Nodes
Light nodes perform verification by random sampling:
- Request a small number of random chunks (e.g., 30–75 chunks) from the 2N total
- Verify each chunk against the published KZG commitment
- If all requested chunks are returned and valid → the light node considers the data “probably available”
Mathematical guarantee:
If a malicious producer withholds ≥50% of chunks, the probability that a light node’s random 30-chunk sample all hit available chunks is:
“`
P = (0.5)^30 ≈ 1 in 1 billion
“`
So sampling 30 chunks gives overwhelming statistical confidence the data is available (or flags it as likely unavailable). With 10,000 light nodes each sampling 30 chunks, the sampling covers essentially the entire encoded block.
DAS and Danksharding
Full danksharding is the Ethereum upgrade that brings data throughput from ~6 blobs/block (Pectra) to 64–128 blobs/block. At 128 blobs × 128KB = 16MB per block of blob data, if every validator had to download 16MB per block (at 1 block every 12 seconds), the bandwidth requirement would be ~10 Mbps just for blob data — significantly above consumer bandwidth levels, threatening validator decentralization.
With DAS:
- Full nodes download a random 50%+ sample — they don’t need the full 16MB/block
- Light nodes download their small random sample (~30 chunks)
- The network collectively verifies the full data without any individual node bearing the full load
- Validators can participate with consumer hardware and bandwidth
DAS enables data throughput to scale without proportionally increasing per-node requirements.
Current Status vs. Full DAS
EIP-4844 (Proto-Danksharding) — What we have now (post-Dencun):
- Blobs are introduced; KZG commitments per blob
- Full blob data is downloaded by beacon nodes (not DAS)
- Data is pruned after ~18 days
- No DAS yet — this is proto-danksharding, the stepping stone
Full Danksharding — The future:
- 64–128 blobs/block
- Full DAS implemented — light nodes sample instead of full-download
- Full Danksharding enables massive blob throughput without forcing full download on validators
The KZG commitments in EIP-4844 are designed specifically to be compatible with DAS — making the Dencun/EIP-4844 work a prerequisite for future full Danksharding.
Celestia and Alternative DA Layers
Celestia is a modular blockchain dedicated to data availability that already implements DAS on its live network — making it an alternative data availability layer for rollups instead of posting blobs to Ethereum. Celestia’s DAS implementation uses Reed-Solomon erasure coding + random sampling, similar to Ethereum’s planned design.
This created the DA wars debate: should rollups post data to Ethereum L1 (EIP-4844 blobs, eventually full danksharding with DAS) or to Celestia/other DA layers for cheaper/faster DA? Ethereum advocates argue that L1 DA provides stronger security guarantees; modular DA advocates argue separate DA layers enable much faster iteration.
History
- 2019 — Data availability problem formally analyzed in Ethereum research context by Mustafa Al-Bassam, Alberto Sonnino, and Vitalik Buterin (paper: “Fraud Proofs: Maximising Light Client Security and Scaling Blockchains”).
- 2020 — Danksharding concept evolves; Dankrad Feist formalizes the combined erasure coding + KZG commitment + DAS design.
- 2022 — EIP-4844 published as proto-danksharding — laying KZG commitment groundwork for future DAS.
- 2023 — Celestia mainnet — First live network with DAS in production.
- 2024, March — Dencun/EIP-4844 activates, enabling blobs without yet full DAS.
- 2025+ — DAS research and development continues; targeted for post-Pectra Ethereum upgrade cycles.
Common Misconceptions
“DAS is already live on Ethereum.”
EIP-4844 uses KZG commitments for blobs but does not yet implement DAS — beacon nodes download full blob data. DAS is planned for full danksharding, a future upgrade.
“DAS makes Ethereum fully trustless for data availability.”
DAS provides a strong probabilistic guarantee — not a deterministic proof. It relies on: (1) enough light nodes participating in sampling to cover the data, (2) correct erasure coding by the block producer. The KZG commitment verifies the erasure encoding is correct, removing the second concern; the first depends on sufficient network participation.
“Erasure coding without DAS is enough.”
Erasure coding alone doesn’t solve data availability if nodes don’t actually sample the encoded data. Simply saying “data is erasure-coded” doesn’t help if nobody checks that the chunks are available. DAS is the sampling mechanism that makes erasure coding useful for DA verification.
Criticisms
- Network participation dependency — DAS requires enough light clients actively sampling to achieve meaningful coverage. If light client participation is low (as it typically is today), the statistical guarantee weakens.
- Erasure coding proof complexity — Verifying that the erasure coding was done correctly requires KZG proofs — the KZG trusted setup is required (same KZG ceremony as EIP-4844). This inherits the KZG trust assumption.
- Implementation complexity — Full DAS requires changes to both consensus and networking layers (p2p protocols must support chunk-level requests). This is significantly more complex than EIP-4844’s blob handling and contributes to danksharding’s long development timeline.
- Not yet battle-tested on Ethereum — Celestia is the only live DAS implementation at scale; Ethereum’s DAS hasn’t been production-tested. Whether the theoretical guarantees hold under adversarial network conditions at Ethereum’s scale remains to be demonstrated.
Social Media Sentiment
DAS is primarily a Ethereum researcher Twitter/X topic — discussed in threads about danksharding, Celestia vs. Ethereum DA, and rollup data posting strategy. The EIP-4844 aactivation brought DAS into semi-mainstream discussion as the “next step after blobs.” Celestia’s launch in 2023 made the DA debate practical — forcing rollup teams to explicitly decide on their DA layer strategy. r/ethereum and Ethereum research forums treat DAS as foundational to long-term scaling; token-focused communities rarely engage with the technical details. The “DA wars” narrative (Ethereum DA vs. Celestia/EigenDA) was a major 2023–2024 multi-chain interoperability debate.
Last updated: 2026-04
Related Terms
Sources
- Al-Bassam, M. et al. (2019). “Fraud and Data Availability Proofs: Maximising Light Client Security and Scaling Blockchains with Dishonest Majorities.” arXiv:1809.09044.
- Feist, D. (2022). Proto-Danksharding FAQ and Full Danksharding Design. notes.ethereum.org.
- Buterin, V. (2021). A data availability FAQ. vitalik.eth.limo.
- Lazaridis, K. et al. (2023). “Analysis of Data Availability Sampling on Ethereum.” arXiv:2307.02842.