Summary
Overview
Work History
Education
Skills
Certification
Additional Information
Timeline
Generic

Govindaraj Sanjeevi

Seattle,WA

Summary

A highly motivated, goal oriented and easy to work with data professional with 13+ years of experience in Data Engineering, Data Analytics, Data Warehousing and Business Intelligence. Passionate about leveraging quantitative techniques, programming skills and analytical skills to provide strategic decision making to stakeholders of all levels. Excellent analytical, problem-solving and documentation skills. Team player with excellent Interpersonal and communication skills.

Overview

14
14
years of professional experience
1
1
Certification

Work History

LEAD DATA ENGINEER

Amazon
09.2024
  • Migrating jobs on the Redshift clusters from one instance to the other
  • Working on several internal tools to build the data pipelines for the Digital organization
  • Working on a Model Registry with all the features that we could add as part of the Digital Organization.

LEAD DATA ENGINEER

Hims & Hers
05.2024 - 09.2024
  • Migrated several Aptile and Composer jobs to Astronomer Airflow instances
  • Built Frameworks around dbt & Cosmos for the migration to Databricks
  • Did POC for data ingestion into Databricks from Bigquery with a more cost optimal way
  • Built Confluent connectors for ingesting postgres data to S3
  • Brianstormed Databricks migration from Google Cloud Platform
  • Built Dataflow and DataSync combination of jobs for ingesting data from Bigquery to S3
  • Have setup a CI/CD using CircleCI to deploy the code from Git to Astronomer instances
  • Created Google Cloud Functions to make an API call to trigger tasks when there is a Git merge.

SENIOR DATA ENGINEER

Wayfair
09.2021 - 03.2024
  • Spearheaded efforts to migrate legacy systems onto cloud-based platforms, resulting in improved scalability and cost efficiency
  • Increased team productivity by implementing streamlined processes and effective communication strategies
  • Reduced operational costs through comprehensive process improvement initiatives and resource management
  • Mentored junior team members for career advancement, fostering a pipeline of future leaders within the organization
  • Accomplished multiple tasks within established timeframes
  • Reengineered existing ETL workflows to improve performance by identifying bottlenecks and optimizing code accordingly
  • Ingested variety of sources into the Foundational layer and Curated them later for consumption
  • Supported the consumption for users by creating GraphQL end points and also posting the messages to the Kakfa topic
  • Built CDC components, automation scripts to move existing scripts in internal scheduler to use with Airflow compatibility.

DATA ANALYTICS CONSULTANT

Nationwide Insurance
10.2019 - 09.2021
  • Worked to build Unified Claims Analytical System by bringing in the data from various sources into one single system(Bronze, Silver, Gold layes) in Delta Lake then loaded into Snowflake for consumption
  • Created Apache Spark pipelines to bring in data from various sources
  • Optimized the code to bring data from 1970's till today since it is massive
  • Built Fact tables over 20 billion rows to support the machine learning capabilities
  • Built Claim Events table with all the different type of stages a claim can go through into one single table
  • Trained team members on best practices in data management and analytics methodologies, promoting a culture of continuous learning and improvement.

DATA ENGINEER - BIG DATA

Lbrands
03.2018 - 10.2021
  • Involved in migrating all data warehouse and business intelligence reports from Teradata warehouse to Mapr Hadoop cluster
  • Automating the ingestion process of moving all the data from the Teradata to the Hadoop Distributed File system
  • Created Hive Query Language scripts to populate the data warehouse tables
  • And all these data is merged into the semantic layer(Hbase - NoSql DB) were other sources also loads the data that enable the reporting team to showcase into business users
  • The business objective of the project is to have a centralized data repository that helps the users in their business development decision making in a better performed way.

CONSULTANT

Nationwide Insurance
06.2016 - 03.2018
  • Involved in creating Informatica mappings to map the XML data coming from JMS(Java Messaging Queue) system to the Teradata tables for the underwriting team to perform their analytics
  • Involved in creating Informatica mappings to create a dataset in the XML format and send to the third party company(Ausum) for analytics in the 5MB file size that was requested
  • Involved in various stages of the project to build the pipeline end to end.

ASSOCIATE

Blue Cross Blue Shield of Minnesota
02.2015 - 06.2016
  • Built datasets to help calculate the Member Out of Pocket expenses for decision making for health plan members
  • Involved in Business analysis and requirements gathering and architecting the project
  • Coordinate with the offshore team to develop the project
  • Built several Informatica mappings and Wherescape Red scripts to automate the process
  • Ensured quality, technical guidance, code review and mentoring the juniors.

ASSOCIATE

Blue Cross Blue Shield of Minnesota
02.2011 - 01.2015
  • Involved in migration of the existing Mainframe DB2 data architecture and the allied components (including the JCL, COBOL scripts) to a better model, from a data warehousing perspective, involving Teradata database
  • Involved in analyzing the existing process and scripts and designed for enhancements
  • Built several Informatica mappings, Wherescape Red scripts and SAS scripts to automate the process
  • Ensured quality, technical guidance, code review and mentoring the juniors.

Education

Bachelor of Engineering - Computer Science And Engineering

Magna College of Engineering
Chennai, India
05.2010

Skills

Cloud & Big Data

  • Google Cloud Platform (BigQuery, Dataproc, Dataflow, Bigtable, Cloud Storage, Cloud SQL, Vertex AI, Composer)
  • AWS (S3, Redshift, EC2, EMR, RDS)
  • Snowflake, Teradata

Programming & Analytics

  • Languages: Python, SQL, SAS
  • Big Data: PySpark, Apache Kafka, Confluent
  • APIs: REST, GraphQL

Data Visualization & BI

  • Tableau, Looker

DevOps & Infrastructure

  • Containerization: Docker, Kubernetes
  • CI/CD: CircleCI, GitHub
  • IaC: Terraform

ETL & Orchestration

  • Apache Airflow
  • Informatica

Certification

  • Google Cloud Certified - Data Engineer
  • SAS Certified Base Programmer for SAS 9
  • Informatica Powercenter Developer
  • Oracle 9i: SQL
  • Cognizant Certified Programmer - AHM250(Heath care)

Additional Information

https://www.linkedin.com/in/govind arajsanjeevi/

Timeline

LEAD DATA ENGINEER

Amazon
09.2024

LEAD DATA ENGINEER

Hims & Hers
05.2024 - 09.2024

SENIOR DATA ENGINEER

Wayfair
09.2021 - 03.2024

DATA ANALYTICS CONSULTANT

Nationwide Insurance
10.2019 - 09.2021

DATA ENGINEER - BIG DATA

Lbrands
03.2018 - 10.2021

CONSULTANT

Nationwide Insurance
06.2016 - 03.2018

ASSOCIATE

Blue Cross Blue Shield of Minnesota
02.2015 - 06.2016

ASSOCIATE

Blue Cross Blue Shield of Minnesota
02.2011 - 01.2015

Bachelor of Engineering - Computer Science And Engineering

Magna College of Engineering
Govindaraj Sanjeevi