Summary
Overview
Work History
Education
Skills
Websites
Timeline
Generic

JANARDHANA BHOGIREDDY

Jacksonville,FL

Summary

Data Engineer with 11+ years of professional experience with an analytical mindset, strong communication skills and proficiency in Data warehousing on large scale distributed data, Data Modelling, ETL design & implementation, schema design, AWS cloud platform, building scalable and robust data pipelines to enable data-driven decisions for the business.

Overview

12
12
years of professional experience

Work History

Sr Data Engineer

Momentive/SurveyMonkey
05.2022 - Current
  • Designed and built out a flagship customer data product on Snowflake Cloud Data Warehouse allowing the Sales, Marketing, Customer Operations, Product teams to have a clear understanding and representation of our customers to drive the business insights using AWS services, Airflow, PySpark, Snowflake and data modeling concepts
  • Developed a centralized data pipeline to handle GDPR compliance and Data Anonymization using dynamic conditional masking across the Company level Enterprise data warehouse with python, DBT and SQL stored procedures, which strengthened the trust of security team by 70%
  • Ingested real-time billing data from STRIPE APIs and marketing data from Google Analytics APIs into Snowflake using Fivetran, configured for recurring 15-minute updates
  • Architected and implemented a multi-dimensional galaxy data model for product usage data in Snowflake using SQLDBM, powered by Fivetran, DBT, and Airflow
  • Optimized complex ETL data pipelines by introducing proper partitioning and clustering using PySpark, Snowflake, DBT and Airflow to meet the desired SLAs
  • Built Report Automation Framework using python to automate 3000+ SQL reports simultaneously and delivered recursive reports to clients on AWS platform with 98.9% efficacy
  • Built a common AWS cost optimization framework using python to terminate all EMR and EC2 instances seamlessly when not in use, which reduced over all AWS spending by 30%
  • Created custom Airflow operators using python to interact with services like EMR, EC2, Athena, S3, DynamoDB and Snowflake which are being used with in Enterprise by ~30 teams
  • Migrated 30+ Legacy Airflow complex DAGs to Amazon Managed Workflows for Apache Airflow (MWAA) for more scalable and robust infrastructure.

Data Engineer

Nike
07.2020 - 05.2022
  • Application design, development, customization and implementation using PySpark transformation in AWS environment
  • Build SalesDataProduct1 & 2 for commercial analytics to plan their marketplace inventory
  • Designed and developed data pipelines (AWS redshift transformations) using bash scripting to implement SCD1 on dimension tables and SCD2 on fact tables to maintain historical data with the current data in enterprise data warehouse
  • Created a master validation script using python to check for missing values and duplicate values in enterprise data warehouse tables by comparing it with source database which yielded 50% increase in data quality
  • Enhanced existed pipelines for stage1 and stage 2 environments for downstream consumers
  • Scheduling ETL workflows with Airflow, working on AWS Databricks clusters and AWS EMR clusters
  • Worked with team in design sessions and translate complex functional and technical requirements into detailed design for successful implementation of projects.

Data Engineer

Florida Blue
02.2017 - 07.2020
  • Perform complex analysis and evaluation of existing and proposed applications on Hadoop
  • Created Spark modules, write complex queries, functions, views to generate reports for business users
  • Analyze business requirements & perform impact analysis for the same on existing application
  • Translate business requirements into logical, component-based technical design documentation
  • Translate complex functional and technical requirements into detailed design
  • Abide by security and data privacy standards for data access
  • Created scalable and high-performance Hadoop platforms.

Software Engineer

Ikea
01.2012 - 07.2015
  • Create data mapping, workflow and deployment documents
  • Created Talend jobs to copy the files from one server to another and utilized Talend FTP components
  • Using Analytix Ds migrated Data Stage Jobs into Talend jobs
  • Created and managed Source to Target mapping documents for all Facts and Dimension tables Migrating existing Datastage Jobs into Talend and design new jobs to pull data from heterogeneous sources (XML, SQL Server, Sybase, MS Excel) using Talend most used components (tMap, tDie, tConvertType, tJoin, tBufferOutput, tFlowMeter, tSetGlobalVar, tHashInput & tHashOutput and many more)
  • Exported datastage jobs as XML file and migrated by using ETL converter then loaded into Talend jobs.

Education

Master of Computer Application -

Ellenki College of Engineering & Technology
Hyderabad, India
09.2008

Skills

  • Python
  • SQL
  • Bash Scripting
  • HighTouch
  • Monte Carlo
  • Pentaho Data Integration
  • Git
  • JIRA
  • Jenkins
  • Snowflake
  • Amazon Redshift
  • PostgreSQL
  • Hive
  • MySQL
  • S3
  • EC2
  • EMR
  • Glue
  • Athena
  • Redshift
  • DynamoDB
  • Kinesis
  • MWAA
  • AWS Local Runner
  • Spark
  • PySpark
  • Hadoop
  • Kafka

Timeline

Sr Data Engineer

Momentive/SurveyMonkey
05.2022 - Current

Data Engineer

Nike
07.2020 - 05.2022

Data Engineer

Florida Blue
02.2017 - 07.2020

Software Engineer

Ikea
01.2012 - 07.2015

Master of Computer Application -

Ellenki College of Engineering & Technology
JANARDHANA BHOGIREDDY