10+ years of professional experience in IT, with 5+ years of work experience in BigData Technologies (PySpark, Python, AWS Cloud, Airflow, Databricks) and API development using FastAPI and 5+ years of experience in Microsoft SQL server and Business Intelligence (SSRS, SSIS & SSAS and Microsoft Power BI).
Overview
14
14
years of professional experience
1
1
Certification
Work History
Sr. Data Engineer
Nike, Inc.
08.2022 - Current
Migrated data assets—including tables and views—from Snowflake to Databricks, enhancing performance, scalability, and reliability.
Designed and deployed FastAPI applications with GET and PUT endpoints; used SQLAlchemy ORM for PostgreSQL schema modeling and database interaction.
Ingested external API data into Databricks via PostgreSQL, validated using Bruno, supporting operational data products.
Orchestrated scalable ETL pipelines using Apache Airflow for batch and streaming loads into Snowflake and Databricks.
Containerized FastAPI services and ETL workflows using Docker, integrated into CI/CD pipelines for automated deployment across dev, QA, and prod.
Transitioned project environment and dependency management from pip/venv to Hatch, enabling modern, reproducible builds.
Cleaned and optimized AWS S3 storage during migration to DBFS, reducing costs and eliminating redundant files.
Led schema normalization efforts by identifying PK/FK violations and proposing column deprecations while maintaining 3rd Normal Form (3NF), improving data integrity and clarity.
Implemented a standardized Tech Solution ID framework across GitHub repositories to give leadership visibility into active, deprecated, and under-development data products.
Maintained high code quality through enforcement of Black, Mypy, and Flake8 in CI workflows.
Utilized Pydantic for data validation and parsing, especially to validate and serialize incoming API request data in FastAPI applications, ensuring data integrity and reducing runtime errors.
Environment: Spark, Python, AWS S3, Snowflake, Jenkins, GitHub, Airflow Scheduler, PyCharm, Hive, Confluence, SQL Server Integration Services 2019(ETL), SQL Server Management Studio 2019, Microsoft Visual Studio 2019 (Business Intelligence Development Studio).
BigData Engineer
Florida Department of Health (FDOH)
02.2020 - 07.2022
Worked in AWS, PySpark environment, develop the batch scripts in Python.
Handled data from source systems and transform using Spark framework and load to S3, Hive and Snowflake for the consumption layers.
As part of the ETL phases engaged in pre and post data quality checks.
Develop the Pyspark scripts, Unit Testing, deploy through Jenkins pipeline to prod environment.
Provided on call support for the batch scripts developed.
Worked as a Production Support and followed the escalation matrix to escalate the production ticket.
Worked on Airflow DAGs in python. Build the Job flows while setting the dependencies.
Experienced in using AWS cloud services S3, EMR cluster.
Expertise in EMR cluster configurations, On Demand and Autoscaling the EMR capacity.
Monitored real time failed tasks, perform data analysis and improve performance of the spark execution time by tuning the spark application with appropriate memory values for different use cases.
Use Git Repo for code repository.
Power BI, MSBI Developer.
Microsoft – Tata Consultancy Services (India)
12.2015 - 08.2016
Advanced Analytics (Forecasting).
Build new capabilities per business needs.
Driving Business Optimization through data insights
Fuzzy matching on ISV names
Requirement gathering for Business insight and analysis and Functional Specifications
Document high level functional and technical design specifications
Testing – Functional, Regression
Deployment on- Power BI DX site - ISV, Audience, Tools ,Azure
Manage Power BI Reporting Platform (ISV, Audience, Tools) for user access, deployment.
Environment: Power BI, SSIS, SSAS.
Microsoft Business Intelligence (MSBI) Developer.
Blue Cross & Blue Shield of Louisiana - Tech Mahindra
02.2011 - 11.2015
Requirement gathering & analysis in Claims, Provider and Medical Management area while interacting with OSCs/Clients
Responsible for review of the deliverable from offshore Claims, Provider and Medical Management area.
Assist team to understand domain specific requirement in Claims, Provider and Medical Management area.
Involved in Design and Development of SSIS Packages using various Control Flow and Data Flow items to Extract, Transform and Load the Data using SQL Server Integration Services (SSIS).
Experienced in configuring with Error logging and Event handling to redirect error rows and fix the errors in SSIS Testing - unit testing & integration testing.
Worked on Huge Data Transfers from & to SQL Server Databases using utilities / tools like DTS, SSIS, BULK INSERT etc ., and used configuration files and variables for production deployment.
Generated multiple Enterprise reports using SSRS from SQL Server Database (OLTP) and included various reporting features such as group by, drilldowns, drill through, sub-reports, navigation reports (Hyperlink) etc.
Created different Parameterized Reports (SSRS 2005) which consist of report Criteria in various reports to make minimize the report execution time and to limit the no of records required.
Report parameters included single valued parameters, multi-value parameters which also consisted of different parameter types.
Worked on all types of report types like tables, matrix, charts, sub reports etc.
Scheduled reports to refresh and run on daily, weekly and monthly basis in the report manager.
Trained freshers on SSRS and SSIS and also on domain knowledge (includes HIPAA rules and regulations).
Developed and maintained T-SQL stored procedures.
Environment: Microsoft SQL Server Management Studio, SSRS, SSIS, SSAS.
Education
Master of Science - Information Technology Management
Campbellsville University
Campbellsville, KY
05-2020
Bachelor's - Electrical, Electronics And Communications Engineering