Blockchain Data in Cloud
Bitquery provides ready-to-use blockchain data dumps in Parquet format via popular cloud providers such as AWS S3, Google Cloud Storage, Snowflake, and BigQuery.
You can plug these datasets directly into your existing analytics stack and build custom data pipelines without running your own blockchain infrastructure or maintaining complex indexing systems.
Overview
Our cloud data export service delivers production-ready blockchain datasets optimized for large-scale analytics, historical backfills, and data lake integrations. All data is provided in Apache Parquet format, ensuring optimal compression, columnar storage, and compatibility with modern analytics engines. We can also provide other file formats if required.
Key Benefits
- No Infrastructure Management – Skip running blockchain nodes, indexers, or data processing infrastructure
- Production-Ready Format – Parquet files optimized for analytics workloads
- Cloud-Native – Direct integration with AWS S3, Google Cloud Storage, Snowflake, and BigQuery
- Historical Coverage – Complete blockchain history from genesis blocks
- Multi-Chain Support – Access data from major blockchain networks
- Cost-Effective – Pay only for the data you need, when you need it
- Scalable – Handle petabytes of blockchain data with ease
Available Blockchain Data Dumps
Bitquery provides comprehensive cloud data dumps for the following blockchains:
EVM Chains Data Export
Export blockchain data for Ethereum, BSC, Base, Polygon/Matic, Optimism, Arbitrum, and other EVM-compatible chains. Includes:
- Blocks – Block-level metadata and timestamps
- Transactions – Full transaction-level data with gas information
- Transfers – Native token and ERC-20 token transfers
- Balance Updates – Account balance changes per block
- DEX Trades – Decentralized exchange trading data
- DEX Pools – Liquidity pool metadata and state
- Smart Contract Calls – Function calls and interactions
- Events – Ethereum event logs and emissions
- Miner Rewards – Block rewards and transaction fees
- Uncle Blocks – Ethereum uncle block data
Use Cases: DeFi analytics, NFT tracking, smart contract analysis, token holder analysis, DEX volume analysis, cross-chain analytics.
Solana Blockchain Data Export
Export Solana blockchain data including slot-level blocks, transactions, transfers, and DEX activity:
- Blocks – Slot-level block metadata
- Transactions – Full transaction-level data with signatures
- Transfers – Native SOL and SPL token transfers
- Balance Updates – Account balance changes per slot
- DEX Pools – Decentralized exchange pool metadata
- DEX Orders – Order-level DEX activity and fills
- DEX Trades – Executed trades on Solana DEXs
- Rewards – Validator and staking rewards
Use Cases: Solana DeFi analytics, NFT marketplace analysis, token transfer tracking, DEX volume analysis, validator performance monitoring.
Tron Blockchain Data Export
Export Tron blockchain data for comprehensive network analysis:
- Blocks – Block-level metadata
- Transactions – Full transaction-level data
- Transfers – Native TRX and TRC-20 token transfers
- Balance Updates – Account balance changes per block
- DEX Trades – Executed trades on Tron DEXs
Use Cases: Tron DeFi analytics, TRC-20 token tracking, DEX volume analysis, account balance monitoring, transaction flow analysis.
Bitcoin Blockchain Data Export
Export Bitcoin blockchain data including transaction inputs, outputs, and OMNI Layer protocol data:
- Blocks – Block-level metadata
- Transactions – Full transaction-level data
- Inputs – Transaction input data and UTXO references
- Outputs – Transaction output data and addresses
- OMNI Transactions – OMNI Layer protocol transactions
- OMNI Transfers – OMNI Layer token transfers
Use Cases: Bitcoin transaction analysis, UTXO tracking, address clustering, OMNI token analysis, blockchain forensics, historical price analysis.
Data Format and Structure
Blockchain data is provided by default in Apache Parquet format, a columnar storage file format optimized for analytics workloads. We can also provide data in other file formats (CSV, JSON, Avro, etc.) based on your requirements. Parquet offers:
- High Compression – Reduces storage costs by up to 90%
- Columnar Storage – Enables efficient column pruning and predicate pushdown
- Schema Evolution – Supports schema changes over time
- Universal Compatibility – Works with all major analytics engines
File Organization
Data is organized by blockchain and topic, with files named using block/slot ranges:
bitquery-blockchain-dataset/
├── ethereum/
│ ├── blocks/
│ ├── transactions/
│ ├── transfers/
│ ├── balance_updates/
│ ├── dex_trades/
│ └── ...
├── solana/
│ ├── blocks/
│ ├── transactions/
│ ├── transfers/
│ ├── dex_trades/
│ └── ...
├── bitcoin/
│ ├── blocks/
│ ├── transactions/