Summary
Overview
Work History
Education
Skills
Personal Information
Certification
Timeline
Generic

Jagadish

Minneapolis,MN

Summary

I have extensive experience in data analysis and building ETL pipelines using technologies such as Azure, Databricks, Python, Spark, Snowflake, Synapse, and SQL. I excel in architecting scalable Cloud solutions. I design efficient data pipelines using Azure Data Factory (ADF) and Databricks, leveraging Snowflake and Synapse for advanced analytics. Proficient in CI/CD practices, to ensure seamless integration and deployment.

My expertise spans Cloud, Bigdata and traditional RDBMS databases, and I have successfully migrated on-prem applications to the cloud, ensuring data integrity and performance. I lead cross-functional teams, translating business requirements into actionable technical strategies, and providing mentorship to drive innovation and excellence in data engineering and analytics. I communicate effectively with leadership, identifying cost-saving opportunities and providing detailed cost summary reports. Additionally, I have expertise in change management using ServiceNow, effectively managing and tracking changes to ensure smooth transitions and minimal disruptions.

Overview

18
18
years of professional experience
1
1
Certification

Work History

Data Engineering

Optum
07.2018 - Current
  • Company Overview: United Health Care
  • Scalable Solution design, building ETL pipelines, Data quality, integrating pipelines with CI/CD, delivering the products and collaboration/train with stakeholders, Operational support and Service now
  • Have worked on design solution design for ETL pipelines to process the claims to identify the payment recovery line of businesses
  • Collaborating with analytics teams and product team to build the ETL data pipelines
  • Worked closely with the Analytics Product teams to enhance engineering solutions for analytics, emphasizing data engineering and ensuring a smooth transition of data from on-premises databases to cloud-based platforms
  • Extract Transform and Load data from Sources Systems to Azure Data Storage services using a combination of Azure Data Factory, Spark SQL Azure Data Lake Analytics
  • Migrated data from on-premises databases like SQL Server, Oracle, and DB2 to cloud platforms such as Snowflake, involved in the data analysis and ensuring data integrity before building pipelines or analytical tools
  • Data Ingestion to one or more Azure Services - (Azure Data Lake, Azure Storage, Delta Lake, Azure Synapse, Snowflake) and processing the data in Azure Databricks
  • Worked on modernizing the ETL pipelines using Azure Data Factory (ADF) for scalable solutions to handle the large volume with data quality checks
  • Performed data quality issue analysis using Snow SQL by building analytical warehouses on Snowflake
  • Worked on large datasets, monitoring performance challenges and providing insights for business on challenges and recommendations
  • Worked on integrating ETL pipelines with GIT to ensure seamless integration and deployment
  • Extensively worked on the building automated cost analysis reports to share with stakeholders
  • Managed change initiatives using ServiceNow, effectively tracking and implementing changes to ensure smooth transitions and minimal disruptions

Technologies: python, Spark, Databricks, Azure, snowflake, Synapse, SAS, github, SQL Server, Oracle, DB2

Lead - Data Engineering

Optum Services Inc
09.2016 - 06.2018
  • BigData Modernization – Performed role as a Team Lead for project delivery, independent contractors and resources to identify and define deliverables, including all developing activities to build modernize ETL pipelines to perform Cleansing, Parsing, Analyzing and distributed Bigdata technology using spark, hive, sqoop, hbase, oracle to provide modern solution for business
  • Developed architecture document and design proto type jobs based on the business requirements along with high-level design (HLD)
  • Experience in managing project tasks and milestones with cross functional teams and coordinate follow up on Action items
  • Responsible for designing and delivery end to end ETL process implementations
  • Worked on large volumes of data, Design ETL pipelines and provide the insights for business
  • Successfully implemented the bigdata ETL pipelines using Sqoop, HQL and spark as the main execution engines
  • Developed and designed spark jobs to consume data from multiple data sources like Oracle, hive, HDFS file systems
  • Optimized and fine-tuned spark, HQL jobs to effectively make use of memory and resources which resulted in reduced in run times and running more jobs in parallel
  • Used parquet as the main file format to store data in bigdata and cloud echo systems
  • Used Airflow to schedule and automate spark jobs
  • Handled Change control requests effectively using system like Service now to track the enhancements/changes for upcoming releases
  • Actively participates in all sprint meetings (Planning, Grooming, Daily Stand-ups, Sprint Review and Retrospectives.
  • Works closely with scrum master and technical product manager on day to days business as part of scrum, grooming and planning sessions
  • Collaborated with other teams to share and implement best practices

Environment: Bigdata, Hive, Sqoop, Spark, Mapreduce, Mapr, Git, Flat files, SQL Server, UNIX, TWS, Airflow, DataStage

Sr BigData Engineer

Optum- United Health Group
03.2014 - 08.2016
  • Roles: Business requirements gathering, Data Analysis, Technical Design, ETL development, Unit Testing, Code review and Production Support/Monitoring

Environment: DataStage 8.7/11.3 (Designer, Director, Manager and Administrator), Oracle11g, DB2, Shell Script, Perl, SQL Loader, Toad, TWS, Service-Now, Subversion

ETL Developer/Data Engineer

INGENIX
02.2010 - 02.2014
  • Roles: Business requirements gathering, Data Analysis, Technical Design, ETL development, Unit Testing, Code review and Production Support/Monitoring

Environment: IBM Data Stage 8.7, Oracle 11g, Netezza 7.0, SQL, PL/SQL, UNIX (AIX), Windows NT 4.0, HPSM, ITG self-service, Tivoli work load scheduler

ETL Developer

Paramount Global Solutions
07.2007 - 12.2010

Roles & Responsibilities: Business requirements gathering, Data Analysis, ETL development, Unit Testing, Code review and Production Monitoring

Environment: Data Stage 7.5, Oracle 10g, IBM AIX 5.x, Test Director, TOAD 7.3, ERWIN 4.2, Win 2000/NT and AIX-UNIX

Education

Master's -

Electronics Engineering

Skills

  • Azure
  • ADLS
  • Data Bricks
  • Data Factory
  • Lakehouse
  • ARM template
  • Az copy
  • Bigdata
  • Sqoop
  • Hive
  • Parquet
  • Spark
  • Pig
  • HDFS
  • Orc
  • Sql
  • Shell Script
  • Terraform
  • Spark scala
  • Python
  • ETL
  • SAS
  • DevOps
  • GitHub
  • Jenkins

Databases

  • Oracle
  • Synapse
  • Snowflake
  • Databricks Lakehouse
  • DB2
  • Teradata
  • SQL Server
  • PostgreSQL
  • Storage
  • Unix
  • ADLS-2
  • Blob

Change Management

  • Service now

Data Analysis

  • RDBMS
  • Data engineering
  • Data analyzation
  • Data mining proficiency
  • Data transmission protocols
  • Pipeline engineering
  • Data management strategy
  • Data governance frameworks
  • Engineering design development

Personal Information

Work Permit: Authorized to work in the US for any employer (PR)

Certification

Snow Pro Certification

Timeline

Data Engineering

Optum
07.2018 - Current

Lead - Data Engineering

Optum Services Inc
09.2016 - 06.2018

Sr BigData Engineer

Optum- United Health Group
03.2014 - 08.2016

ETL Developer/Data Engineer

INGENIX
02.2010 - 02.2014

ETL Developer

Paramount Global Solutions
07.2007 - 12.2010

Master's -

Electronics Engineering
Jagadish