Currently, there are 3 sources of BSC snapshot could be used, if you are unclear about their differences, you may refer Clarification of the snapshots first, to decide which snapshot suit you the most.
Usage: usage/legacyfullnode_usage.md
Currently, snapshots are provided as block history data divided by year. Each list includes world states, historical block files, URL, MD5, file size. Users can use the provided tools to quickly download.
The Pruned Snapshot is better for internal/private nodes, it only keeps the latest 9w historical blocks and requires BSC client >=v1.5.5.
| Snapshot Type | Snapshot File | Total Size | Remark |
|---|---|---|---|
| Full Snapshot | mainnet-geth-pbss-20260202 | ~5.0TB | BSC >= v1.6.4 |
| Pruned Snapshot | mainnet-geth-pbss-20260202-pruneancient | ~1.4TB | BSC >= v1.6.4 |
| Snapshot Type | Snapshot File | Total Size | Remark |
|---|---|---|---|
| Full Snapshot | testnet-geth-pbss-20251013 | ~300GB | BSC >= v1.6.1-beta |
| Pruned Snapshot | testnet-geth-pbss-20251013-pruneancient | ~120GB | BSC >= v1.6.1-beta |
Attention: The 20251013 snapshot has no filtermap index with flag
--history.logs.disable.
You can download the mainnet or testnet files separately in the list and unzip them in a same directory. Or you can use the following script:
# install aria2 on your os
yum install aria2
wget https://raw.githubusercontent.com/bnb-chain/bsc-snapshots/main/dist/fetch-snapshot.shParameters:
-ddownload,-eextract,-cverify MD5 checksum,-puse pruned snapshot-D <dir>directory to store downloaded archives-E <dir>extraction target, should be the node's--datadir(e.g./data/bsc), files will be extracted to<dir>/geth/chaindata/...--auto-deletedelete each archive immediately after extraction to save disk space-pauto-appends-pruneancientto the snapshot name, do NOT use with a name that already has the suffix
Quick start (download, verify, extract, auto-delete in one step):
# full snapshot (~5TB, needs at least 6TB free, or 4TB with --auto-delete)
bash fetch-snapshot.sh -d -e -c --auto-delete -D /data/snapshot -E /data/bsc mainnet-geth-pbss-20260202
# pruned snapshot (~1.4TB, needs at least 2TB free with --auto-delete)
bash fetch-snapshot.sh -d -e -c -p --auto-delete -D /data/snapshot -E /data/bsc mainnet-geth-pbss-20260202After extraction, files will be at /data/bsc/geth/chaindata/..., start geth with --datadir /data/bsc.
Step by step (download first, extract later):
# step 1: download & checksum
bash fetch-snapshot.sh -d -c -D /data/snapshot mainnet-geth-pbss-20260202
# step 2: extract to datadir
bash fetch-snapshot.sh -e -D /data/snapshot -E /data/bsc mainnet-geth-pbss-20260202You can remove the -c option to skip MD5 checking. Run bash fetch-snapshot.sh --help for all options.
Please keep
fetch-snapshot.shthe latest version.
- mainnet:
- mainnet-geth-pbss-20260104, mainnet-geth-pbss-20260104-pruneancient
- mainnet-geth-pbss-20251205, mainnet-geth-pbss-20251205-pruneancient
- mainnet-geth-pbss-20251114, mainnet-geth-pbss-20251114-pruneancient
- mainnet-geth-pbss-20251013, mainnet-geth-pbss-20251013-pruneancient
- mainnet-geth-pbss-20250906, mainnet-geth-pbss-20250906-pruneancient
- mainnet-geth-pbss-20250806, mainnet-geth-pbss-20250806-pruneancient
- mainnet-geth-pbss-20250715, mainnet-geth-pbss-20250715-pruneancient
- mainnet-geth-pbss-20250605, mainnet-geth-pbss-20250605-pruneancient
- mainnet-geth-pbss-20250520, mainnet-geth-pbss-20250520-pruneancient
- mainnet-geth-pbss-20250501, mainnet-geth-pbss-20250501-pruneancient
- mainnet-geth-pbss-20250404, mainnet-geth-pbss-20250404-pruneancient
- mainnet-geth-pbss-20250310, mainnet-geth-pbss-20250310-pruneancient
- testnet:
- testnet-geth-pbss-20250610, testnet-geth-pbss-20250610-pruneancient
- testnet-geth-pbss-20250407, testnet-geth-pbss-20250407-pruneancient
- testnet-geth-pbss-20240711.tar.lz4(md5: 64626987189d739bd1a3ee743387f8a6)
Usage: https://github.com/BNB48Club/bsc-snapshots
Special thanks to BNB48Club on contributing another dump of snapshot.
Usage: usage/erigon3_archivenode_usage.md, Erigon 3 release: v1.3.10
Endpoints: erigon_51600000
Usage: reth-bsc snapshot usage
| Type | Network | File | Size | Special Parameters |
|---|---|---|---|---|
| Archive MDBX | BSC-MainNet | 20260202_mainnet_reth_mdbx_static_files_archive_node.tar.zst 20260202_mainnet_reth_mdbx_db_archive_node.tar.zst |
2.65 TiB 3.28 TiB |
--db.max-size=12TB --db.page-size=8KB |
| Archive MDBX | BSC-TestNet | 20260202_testnet_reth_mdbx_archive_node.tar.zst | 423.58 GiB | - |
| Full MDBX | BSC-TestNet | 20260202_testnet_reth_mdbx_full_node.tar.zst | 270.69 GiB | - |
| Full MDBX | BSC-MainNet | 20260202_mainnet_reth_mdbx_full_node.tar.zst | 3.0 TiB | - |
| Full TrieDB | BSC-TestNet | 20260202_testnet_reth_triedb_full_node.tar.zst | 356.63 GiB | - |
| Full TrieDB | BSC-MainNet | 20260202_mainnet_reth_triedb_full_node.tar.zst | 3.64 TiB | - |
| Archive TrieDB | BSC-TestNet | 20260202_testnet_reth_triedb_archive_node.tar.zst | 463.92 GiB | - |
| Archive TrieDB | BSC-MainNet | 20260202_mainnet_reth_triedb_static_files_archive_node.tar.zst 20260202_mainnet_reth_triedb_db_archive_node.tar.zst 20260202_mainnet_reth_triedb_triedb_archive_node.tar.zst |
2.65 TiB 3.0 TiB 798.13 GiB |
--db.max-size=12TB --db.page-size=8KB |
Incremental snapshots is a feature that allows you to download only the changes made since the last snapshot, significantly reducing the amount of data you need to transfer.
If this is your first time using snapshots, start by downloading the full snapshot; thereafter, use incremental snapshots to keep your node up to date.
To enable incremental snapshots, use the following flags when starting your BSC node:
--incr.use-remote: Enable incremental snapshots--incr.remote-url="https://download.snapshots.bnbchain.world/incr-snapshot": Set the remote URL--incr.datadir: Specify a directory for storing incremental data
geth --incr.use-remote \
--incr.remote-url="https://download.snapshots.bnbchain.world/incr-snapshot" \
--incr.datadir=/path/to/incremental/data \
<other-flags>Note: This feature requires BSC client >= v1.6.0-alpha-feature-incr-snapshot
As the node snapshot of bsc becomes larger and larger, backup, upload and download will become increasingly unmaintainable, and it will occupy more disk space and take up more upload and download time.
At the same time, in order to support the history expiry and state expiry of bsc later, it is planned to split the node snapshot according to historical data and active data, and the following advantages can be obtained:
- When updating the snapshot, only the changed part can be updated to reduce the difficulty of operation and maintenance;
- Support annual backup of historical data, which also helps with the subsequent historical data pruning;
- Support archiving multiple snapshot versions, avoiding wasting disk space;
- Support downloading and decompressing a single part immediately, and snapshot download and decompression can be completed on a smaller disk;