I'm a Data Engineer building automated pipelines that turn raw data into analysis-ready datasets. I work across the full data stack — from scraping and ingestion to transformation, quality validation, and cloud warehousing.
Currently focused on batch pipelines, BigQuery, and Airflow orchestration. Previously worked as a Data Analyst, so I understand both sides — how data is built and how it's used.
- End-to-end pipelines — scrape → transform → validate → load
- Automated workflows with Apache Airflow on Docker
- Cloud data warehouses on Google BigQuery
- Analytics-ready datasets for business intelligence and reporting
- Apache Spark — distributed data processing
- dbt — SQL-based transformations inside the warehouse
- Medallion Architecture — Bronze / Silver / Gold data layers
- Kafka — event streaming and real-time pipelines
Pipelines ship data. Quality checks protect it. Partitions make it fast.


