Our data approach
We process and store over a petabyte of transaction-level onchain data every day directly from the source, to deliver accurate, auditable metrics you can trust.
Token Terminal uses an ELT (Extract, Load, Transform) paradigm to process raw blockchain data into standardized metrics. This approach ensures data is transparent, reproducible, and auditable at every stage.
In a typical ETL workflow, data is transformed before it is stored. This can work well if the structure of your data is stable.
But in our industry, things change constantly in ways that are different from most traditional data pipelines. For example:
-
Protocol upgrades Blockchains regularly introduce upgrades that can include breaking changes to data schemas. A protocol might add new transaction types, restructure contract storage, or modify how events are emitted.
-
Smart contract deployments Decentralized applications are continuously launching new products or business lines, often across multiple chains. Each new contract introduces additional data that needs to be standardized and integrated.
-
Methodology changes Even the meaning of common metrics evolves over time. How you define metrics like fees, active users, or token incentives can change as protocols mature and industry practices develop.
These factors all make blockchain data feel like ongoing R&D that happens in production.
Because we use ELT, we don’t have to re-ingest everything from scratch when something changes. Instead, we can re-process the same raw data and update our standardized metrics quickly, without interrupting downstream workflows.
By keeping raw data in its original form, we gain a few important advantages over the traditional ETL approach:
- Traceability Every metric can be linked directly back to the source.
- Reproducibility When new smart contracts are deployed or methodologies change, we can re-process historical data within minutes without waiting days to re-ingest it.
- Scale This approach allows us to maintain standardized metrics across hundreds of blockchains and thousands of applications efficiently.
Whether you’re exploring our dashboards, doing analysis in spreadsheets, or querying the API, all data comes from this same high-fidelity data infrastructure.