TransLink: Transit Insights (ETL + Dashboards)

Dec 15, 2025 · 1 min read
projects

What I built

  • Architected scalable ETL pipelines using PySpark + Spark SQL to ingest 23GB of logs into a Data Lake.
  • Implemented transformations across bronze/silver/gold layers to ensure consistency and data quality.
  • Built interactive Tableau and Power BI dashboards for KPI tracking and performance trends.

Tools & methods

PySpark, Spark SQL, Data Lake, Medallion Architecture, Tableau, Power BI

Outcome

A clean pipeline-to-dashboard workflow that enables business-facing KPI visibility and faster decision-making.


Dhruv Saikia
Authors
Data | Game Dev | Cybersecurity
Master’s student at SFU specializing in Big Data.
Background in Data Science, Cybersecurity, and Game Development.
I like building big data pipelines that are secure and are user friendly.