BS Computer Science
Sir Syed University of Engineering and Technology
Data Engineer focused on data pipelines, modeling, and analytics, designing end-to-end solutions that bridge raw data to meaningful intelligence.
I build data pipelines that seamlessly move information from sources like S3 and APIs into Microsoft Fabric and SQL Warehouses—ensuring every workflow is scalable, efficient, and cloud-native.
Recently, I delivered a lakehouse architecture and automated ETL solution that cut processing time by 75%, empowering teams with faster, real-time analytics and smarter decision-making.
Build Scalable ETL pipelines.
Make BI Dashboards
Dimensional modeling: fact and dimension tables.
Build classification and regression models.
Sir Syed University of Engineering and Technology
Predictive models and time series models.
SQL and Python related ETL pipelines
Certified Fabric Data Engineer. Also work on Azure, AWS and GCP.
A curated collection of data engineering and analytics projects — including end-to-end ETL pipelines, data warehousing models, cloud integrations, and performance-optimized Spark workflows. Click a card to view the repository or explore the project overview.
Core strengths across data engineering, warehousing, and analytics—focused on building scalable, reliable, and cloud-based data solutions.
ETL/ELT pipeline design with PySpark, Databricks, and Azure Data Factory; workflow orchestration, data validation, and monitoring.
Hands-on with Azure, AWS (S3, EC2), and Microsoft Fabric—OneLake, Dataflows Gen2, and Pipelines for unified data management.
Dimensional modeling, SCD Type 2, fact/dimension design, and performance tuning for BI and analytics workloads.
SQL, Python, Power BI, and Delta-Parquet for data analysis, reporting, and seamless integration across structured sources.
Also: SQL Server, PostgreSQL, DataBricks, Spark optimization, version control (Git), and data quality frameworks.
View my resume or download the PDF.
I help teams collect, organize, and optimize their data ecosystems from pipelines to dashboards ensuring reliable, high-quality data for business and AI initiatives. Share a brief on your use case and I’ll get back within 24–48 hours.
For any inquiries, please contact me via email.
Email is best for initial scoping or used LinkedIn too.
Happy to schedule a call after your message.
Remote — Worldwide
Replies within 12–24 hours (business days)