Specialising in Azure Databricks, Delta Lake, and PySpark. Architecting high-performance lakehouse systems.
Senior Data Engineer with 4 years at Celebal Technologies building production-grade pipelines on Azure Databricks. Delivered end-to-end medallion architecture lakehouse solutions for a large-scale conglomerate, migrated 600+ legacy ETL jobs for a leading private-sector bank, and re-implemented Oracle PL/SQL sales datamarts at 500M+ record scale.
Comfortable owning the full pipeline lifecycle - from Autoloader ingestion and PySpark transformation to Delta Lake optimisation, ADF orchestration, and CI/CD deployment via Databricks Asset Bundles.
Designed and delivered the end-to-end data ingestion layer for a large-scale Enterprise Data Platform (EDP). Owned the full pipeline lifecycle from heterogeneous source systems through the Bronze layer to the Silver layer following a strict medallion architecture.
Migrated 600+ Pentaho ETL jobs from on-prem Cloudera Hadoop to Azure Databricks. Handled 80M unique customer records for the bank's credit risk departments.
Converted a large body of Oracle PL/SQL procedures and Unix shell-scripts into production-grade PySpark, re-implementing 5 regional sales datamarts on Delta Lake handling 500M+ records.
CDAC | Pune, Maharashtra
Grade: BCenturion University of Technology
CGPA: 9.2 / 10Currently open to Senior Data Engineer roles and complex cloud migration projects.