✔️ 3+ years of experience in data engineering and analytics, with a strong focus on designing, building, and
maintaining scalable data pipelines and cloud-based data solutions on Microsoft Azure.
✔️ Strong problem-solving and optimization skills, with a proven track record of identifying data pipeline
bottlenecks, improving data processing performance, and ensuring reliability, scalability, and cost efficiency
of data platforms.
✔️ Proficient in Azure data engineering tools and technologies, including Azure Data Factory, Azure Synapse
Analytics, Azure Data Lake Storage (ADLS Gen2), SQL, Python, and PySpark, with hands-on experience in data
modeling, ETL/ELT processes, and big data processing.
✔️ Experienced in stakeholder collaboration and requirement analysis, translating business needs into robust
data architectures, delivering high-quality datasets for analytics and reporting, and supporting downstream BI
and advanced analytics use cases.
Tech stack
Python
SQL
PySpark
DataBricks
Azure
Work History
🚧 Senior Data Engineer | Infosys - Pune
Domain: Retail
01/2024 – Present
Designed and implemented a multi-zone Data Lakehouse architecture in ADLS Gen2 (Raw, Curated, Gold
layers),
leveraging Delta Lake and Medallion Architecture for efficient ETL, ACID compliance, and big data
transformations
using PySpark in Azure Databricks.
Built and optimized ETL pipelines with Azure Data Factory to ingest data from diverse sources (SQL DB,
APIs, flat files), migrate
on-premises SSIS packages, and reduce processing time by 40%, while enabling real-time data ingestion via
Event Hubs and
Stream Analytics.
Enhanced analytics and reporting by integrating Azure Synapse Analytics with partitioned tables,
materialized views, and
columnstore indexing, improving query performance by 60%, and connecting Power BI for interactive dashboards
on sales,
inventory, and customer insights.
🚧 Junior Data Engineer | Infosys - Bangalore
Domain: Banking
11/2022 – 12/2023
Designed and developed end-to-end data pipelines using Azure Data Factory, Databricks (PySpark & Delta
Lake),
and Azure Synapse Analytics to integrate structured and unstructured data from CRM systems, web analytics,
and
third-party sources into Azure Data Lake Storage (ADLS Gen2) for unified analytics and reporting.
Built real-time streaming solutions leveraging Azure Event Hubs, Kafka, and Spark Streaming for processing
live
inventory, sales, and telemetry data, while implementing Azure Functions and Logic Apps to automate event-
driven workflows and data orchestration.
Developed ETL workflows in Databricks using PySpark and Delta Lake for large-scale data transformation,
deduplication, feature engineering, and enrichment, ensuring high data accuracy and consistency across
analytical
environments.
🚧 Data Engineer Intern | iNeuron - Remote
Domain: Fintech & Financial
07/2021 – 01/2022
Designed and developed end-to-end ETL solutions using Azure Data Factory, Spark-Scala, and SSIS to
extract, transform,
and load large datasets from multiple sources, including Oracle and MySQL, ensuring optimized data flow and
scalability.
Implemented CDC (Change Data Capture) for near real-time updates, automated data partitioning, indexing,
and
performance tuning in Azure SQL Database, improving query execution and report generation efficiency.
Built and optimized data models, stored procedures, and complex SQL queries to support analytics,
reporting, and
business intelligence, including rank-based customer segmentation using advanced SQL window functions.
Certifications & Accomplishments
Microsoft Certified: Azure Data Engineer Associate Certified.
Databricks Certified: Data Engineer Associate Certified.
Databricks Certified: Data Engineer Professional Certified.
Honored with the "Best Team (2023)" award and the "On the Spot (2024)" award by Infosys for exceptional
performance.
Earned multiple client appreciations for delivering high-quality work and ensuring prompt delivery of
solutions.