Results-driven data engineering professional with over 10 years of expertise in designing, developing, and optimizing OLTP databases, Data Warehouses, ODS systems, OLAP cubes, and Data Marts. Successfully managed end-to-end data engineering processes and created business intelligence solutions for Finance, HR, Supply Chain, Operations, and E-Commerce industries. Proficient in leveraging Azure, AWS, GCP cloud platforms and Power BI, Tableau, Looker BI tools to deliver insightful analytics. Skilled in Python, SQL, PowerShell scripting, ETL/ELT pipelines, and modern data solutions focused on efficiency and ROI. Solid foundation in designing and maintaining scalable data systems. Demonstrated expertise in developing efficient ETL processes and ensuring data accuracy for impactful business insights. Strong collaborative skills and ability to adapt to dynamic project requirements, consistently delivering reliable and timely solutions.
Overview
11
11
years of professional experience
Work History
Data Engineer
FHLB Bank Cincinnati
02.2022 - Current
Led the migration of an on-premises data warehouse to the cloud, enhancing system performance, scalability, and data accessibility
Designed and implemented data quality frameworks using Azure Data Factory (ADF) and Logic Apps to automate validation checks, ensuring accurate data flow between applications
Developed monitoring and alerting mechanisms for data quality issues, providing timely notifications to relevant stakeholders
Trained and mentored interns on the deployment process, utilizing Azure DevOps for CI/CD pipelines and automation best practices
Collaborated with cross-functional teams to troubleshoot and optimize data pipelines for improved efficiency and reliability
Created and managed ETL pipelines using Azure Data Factory to extract, transform, and load data from various sources into the data warehouse
Maintained data transformation processes (ETL, SQL stored procedures, etc.) to support operational data flows
Collaborated with data engineers and data scientists to ensure data consistency, quality, and availability for analytical and reporting purposes
Participated in the evaluation, selection, and integration of new technologies and tools to enhance the database environment
Troubleshooted and resolved data warehouse related issues in a timely manner
Documented database design, architecture, and processes for future reference
Developed and tuned SQL queries in Snowflake, utilizing SnowSQL for data transformation and optimized view creation for reporting purposes
Built data models with DBT, implementing SQL transformations and creating reusable data pipelines for consistent ETL operations
Worked extensively with Azure Data Factory (ADF), using key activities like Copy Data, Lookup, ForEach, Execute Pipeline, Stored Procedure, and Web activity for diverse data integration needs
Managed cloud resources across Azure services, including Azure SQL Database, Cosmos DB, Event Hubs, Kusto SQL, Logic Apps, Function Apps, ADLS, and Synapse SQL Pools
Provided production support, resolving incident tickets and handling failures across ADF, Databricks, SQL Server, and Power BI environments, ensuring data continuity
Led a team of 3-4 junior developers, providing technical oversight and guidance throughout project lifecycles
Collaborated with Solution Engineers to design data models and ETL workflows aligned with client needs
Engaged directly with clients to gather business requirements, define desired outcomes, and translate them into actionable development tasks
Decomposed development work into user stories and tasks in Azure DevOps, assigning and tracking progress to ensure timely delivery
Monitored development tickets, resolved blockers, and mentored junior developers to address technical challenges effectively
Data Engineer
PricewaterhouseCoopers (PwC)
02.2022 - 02.2023
Collaborated with the Data Control and Automation Management (DCAM) team to automate Azure Data Factory (ADF) logs processing
Utilized Azure Data Factory (ADF), Apache NiFi, HALO, and Github to design and develop robust data pipelines for efficient data ingestion, transformation, and storage
Implemented monitoring and alerting solutions to ensure data pipeline reliability and quality
Worked closely with cross-functional teams to identify and resolve data quality issues and performance bottlenecks
Contributed to the maintenance and enhancement of existing data engineering infrastructure
Assisted in the development of best practices and standards for data engineering processes
Utilized Databricks for data processing with PySpark, employing libraries like Pandas, NumPy, Matplotlib, and Seaborn for data manipulation and visualization
Supported business users in defining KPIs and developing Power BI reports, utilizing data modeling, DAX, and RLS for secure and insightful data visualizations
Created ETL workflows in ADF and Databricks, leveraging Medallion architecture to cleanse data in data lakes, building data warehouses for Power BI reporting
Implemented error handling, validation steps, generic pipelines, control tables, and parallel processing within ADF, Logic Apps, and Function Apps
Employed Databricks Unity Catalog for data governance, performance tuning, and managing access control on data assets
Enhanced query performance in Databricks with PySpark, using libraries like Pandas, Scipy, PyArrow, and Plotly for efficient data handling
Authored Snow SQL scripts for Snowflake, creating structured datasets and data models supporting Power BI report generation
Conducted data validation and exploration on AWS S3 bucket files using AWS Athena, AWS Glue ETL, Lambda Functions, and AWS Crawler
Designed Power BI dashboards with M Query, DAX functions, and custom visuals to build interactive and insightful dashboards tailored for business users
Designed and implemented enterprise data warehousing model, metadata solution and data life cycle management in both RDBMS, Data Lakes, Data Marts and Big Data environments
Involved in Logical modelling using Dimensional Modelling techniques such as Star Schemas and Snowflake Schemas
Designed and developed end to end Azure Data Factory (ADF) pipelines using multiple nested copy and transform activities
Involved in architecture and design of data ingestion, data processing and data warehousing strategy to meet BI requirements using Dimensional Model
Data Lake delivered the initiative to fulfil the future data ingestion needs from various data sources to Azure Data Lake and Data bricks
SENIOR CLOUD DATA ENGINEER
GARTNER
07.2020 - 02.2022
Gathered business requirements, met with business users and clients, performed data profiling source system data, documentation, analysis, and design of proposed systems
Developed mapping documents that demonstrated data lineage
Updated legacy data dictionaries, governance practices, and standards
Designed and implemented enterprise data warehousing model, metadata solution and data life cycle management in both RDBMS, Data Lakes, Data Marts and Big Data environments
Involved in Logical modelling using the Dimensional Modelling techniques such as Star Schemas and Snowflake Schemas
Designed and developed end to end Azure Data Factory (ADF) pipelines using multiple nested copy and transform activities
Involved in architecture and design of data ingestion, data processing and data warehousing strategy to meet BI requirements using Dimensional Model
Delivered Data Lake initiative to fulfil the future data ingestion needs from various data sources to Azure Data Lake and Data bricks
Performed analysis, development and migration of Stored Procedures, Triggers, Views, and other related database objects from on-premises SQL Server to Azure Data Factory
Involved in migrating existing SSIS packages to Azure Data Factory using Lift and Shift strategy
Implemented VSTS/ Azure DevOps for source control and engaged in developing CICD pipeline for continuous ETL releases to multiple environments
Authored PowerShell script to Trigger, deploy and monitor ADF pipeline using command line interface and Azure ARM templates
Integrated Azure Logic app for Data Factory pipeline monitoring and notification services
Integrated Azure Key Vault for data security best practices and to securely pass the connection string to the Data Pipeline
Implemented Microsoft cloud technologies like Azure Data Factory, Azure Data Lake, Azure Synapse (Azure Data warehouse), Azure Kusto, Logic Apps, Key Vault, and Azure SQL
Built data pipelines using Azure Data factory (ADF) loading data to Azure Data Lake, Azure SQL Database, Azure SQL Data warehouse
Performed Cloud Migration using Azure SQL Server, Azure Data Lake, Azure SQL Database and implementing incremental loading on premise data to Azure Data Lake store using Azure Data factory
Developed and architected Power BI Data Model and Reports for business users
Developed and published Power BI multiple reports with time intelligence functionality
Created Power BI tabular reports for time-based reporting
Created complex DAX measures
SQL BI DEVELOPER
TOYOTA MOTOR NORTH AMERICA
10.2013 - 07.2020
Collaborated with BI developers to convert functional requirements into technical requirements
Gathered and documented business requirements and transformed these requirements into BRD
Used Test Cases, Prototypes, Network diagrams, and UML (E-R Diagram) for system integration and software implementation projects
Initiated discussion to obtain in-depth customer feedback for service improvement opportunities
Designed Complex SQL scripts with recursive CTEs, CTE, temp tables, and effective DDL/DML Triggers to facilitate efficient data manipulation and consistency and support the existing applications
Worked on the SSAS cube deployment and star schema deployment to different databases
Did Performance tuning to optimize SQL queries using SQL profiler and query analyzer
Worked on deploying the code and unit testing in different team environments
Involved in performance tuning using indexing (Cluster Index, Non-Cluster index) tables
Created complex Stored Procedures, Triggers, Cursors, Tables and other SQL Joins and Statements for Applications by using SQL
Created Data Marts for Staging and Production databases for Operational Data Stores (ODS)
Created SSIS packages to extract data from OLTP to OLAP systems and scheduled Jobs to call the packages and Stored Procedures
Using Power BI and SSRS, created various visualization like Pie charts, bar graphs, Tree maps etc., that will helpful easily to understand the data
Continuously interacted with End-users and Stakeholders to gather business requirements
Used DDL and DML for writing triggers, stored procedures, and data manipulation
Extensively used joins and sub queries to simplify complex queries involving multiple tables
Developed scripts for maintaining administration tasks
Streamlined the Data Load process using SSIS/Stored Procedures and Azure Data Factory
Created Pipelines and Data Flows using Azure ADF and troubleshoot data issues using Azure Data Factory
Built Complex Data Ingestion Pipelines using Mapping Data Flows in ADF
Designed Data Pipelines for Batch Processing in Azure Data Factory and Azure Databricks to perform data transformation from OneStream
Created Data Flows of existing SSIS Packages in Azure Data Factory
Created Data Solutions for Structured and Unstructured Data from Storage to Processing Using Azure SQL Database and Azure Cosmos Database for Data Store, Azure Blob Storage for Staging of Data, Azure Data Factory and Azure Data Bricks for Data Processing
Conducted logical and physical database design including data modelling, maintenance, problem diagnosis and resolution
Developed and created data dictionary, stored procedures, advanced queries, triggers, views, indices, and functions for databases
Maintained SQL databases, performed software installations and upgrades, monitored database performance, performed capacity planning and SQL Server clustering, managed database quality assurance including database consistency check
Performed Dimensional Modelling using Star and Snowflake Schema, Identifying Facts and Dimensions, Physical and logical data modelling using Erwin and ER-Studio
Implemented Slowly Changing Dimensions (Type 1 and Type 2) and Surrogate keys
Designed ADF and SSIS ETL data pipelines from various sources into Data Warehouses and Data Marts
Designed Power BI Reports from scratch and to create self-service BI capabilities and use tabular models
Performed time series analysis using DAX expressions in Power BI
Was responsible for creating and changing the visualizations in Power BI reports and Dashboards on client requests
Created hierarchies in Power BI reports using visualizations like Bar charts, Line charts, etc
Published and managed Power BI reports on dashboards in Power BI server
Created and developed the stored procedures, triggers to handle complex business rules, history data and audit analysis
Performed DBA tasks including creation/alteration, grant of system/db
Roles and permission on various database objects
Created complex SSIS packages using proper control and data flow elements with error handling
Created several complex reports (sub reports, graphical, multiple groupings, drilldowns, parameter driven, formulas, summarized and condition formatting) using SSRS and Tableau
Education
Master of Business Administration (MBA) -
Texas Woman's University
01.2017
Bachelor's Degree -
Moscow State University
01.2009
Skills
Azure (ADF, Synapse, Logic Apps, Data Lake, Fabric, Azure Databricks)