System Design Space
Knowledge graphSettings

Updated: February 21, 2026 at 11:59 PM

Brief overview of the T-Bank data platform

hard

Evolution of T-Bank's data platform: from DWH approaches to Lakehouse, key contours of the platform, scale and practical architectural conclusions.

Source

T-Bank Data Platform

Analysis of the evolution of the platform, the current landscape of systems and architectural contours.

Read the article

Brief overview of the T-Bank data platform shows how the data platform evolves as the company grows: from classic DWH approaches (Inmon/Kimball) to Data Lake and Lakehouse. At the scale of 17k+ users and 144M+ requests per month, the architecture turns into a set of specialized layers where ingestion, storage, processing and governance are designed as a single platform.

Platform scale

Evolution period

18+ years

Consistent transition from classic DWH to Data Lake and further to Lakehouse.

Platform users

17 000+

Engineers, analysts and product teams use a single data contour.

Requests per month

144 million+

High load requires constant work on scalability and latency.

Key systems

19

End-to-end lifecycle: from ingestion and processing to governance, observability and security.

Architectural blocks of the platform

Data collection and transportation

The ingest/replication circuit, which connects OLTP, event streams and downstream data products.

Data Replication: BODS + Chrono

Historical batch circuit and modern streaming replication for stable delivery of changes.

Event Sourcing: SDP

Streaming Data Transfer Platform built using Data Mesh principles for domain streams.

Reverse ETL: Spheradian

Return enriched data back to operating systems with up to 100ms latency.

Architectural Conclusions

At the scale of a data platform, there cannot be one universal DBMS: specialized storage/compute circuits are needed for different latency and throughput profiles.

Data Contracts and the observability loop become critical: without them, an increase in the number of teams turns the platform into an unmanageable set of integrations.

Reverse ETL closes the gap between analytics and operational products: data should be returned to business flow in almost real time.

The stability of the platform is ensured not only by technology, but also by organizational ownership: clear owners, incident processes and uniform standards for data delivery.

Practical checklist

  • Separate ingestion, storage, processing and governance as separate architectural layers with explicit SLAs/SLOs.
  • For each data product, record contracts, owner, and backward compatibility when schemas change.
  • Plan multi-engine execution in advance: the optimal engine for batch, stream and ad-hoc analytics is usually different.
  • Implement data incident management as a mandatory process, rather than a post-factum practice, after the first major failure.

References

Related chapters

Enable tracking in Settings

System Design Space

© 2026 Alexander Polomodov