Summary
Overview
Work History
Education
Skills
Accomplishments
Certification
Timeline
Generic

Praveen Rao Gundavarapu

Leander,TX

Summary

As a Senior engineer with deep expertise in data architecture, pipeline development, and big data technologies. Proven track record in optimizing data workflows, enhancing system efficiency, and driving business intelligence initiatives. Strong collaborator, adaptable to evolving project demands, with focus on delivering impactful results through teamwork and innovation. Skilled in SQL, Python, Spark, and cloud platforms, with strategic approach to data management and problem-solving.

Overview

16
16
years of professional experience
1
1
Certification

Work History

SR. DATA ENGINEER

EVERESTRE Insurance
11.2022 - Current
  • Design azure applications and pipelines for global insurance space for analyzing claims data using SQL, ADF, spark, Databricks and snowflake build dashboards to provide insights for analyzing claims policy and forecast of the target budget strategy
  • Design data modelling and schema development for claims approval, Cost of claims data, Continuous Integration and Continuous Delivery (CI/CD), Test Automation, Infrastructure as Code, Secure Coding Practices
  • Use of kafka and other sources systems ( salesforce cloud platform) api ingestion . load data using synapse and warehouse to connect PBI reports
  • Ingest using confluence kafka data of survey and reviews to snowflake DB and flatten the data
  • Load data to stage to curated and build the bi warehouse for loss impact of policies
  • Build fact and key dimensions for financial reporting needs
  • Languages and Technologies: Azure Data Factory, Snowflake, Web API , hive,DBT,GIT, data bricks, Azure Functions, Azure SQL Polybase, BLOB, Data lake, snowflake, kafka , sales force cloud, U-SQL, logic apps, EventHub, data bricks, Azure DevOPS, Jenkins, Azure key vaults and python
  • Collaborated with cross-functional teams to define requirements and develop end-to-end solutions for complex data engineering projects.
  • Evaluated emerging technologies and tools to identify opportunities for enhancing existing systems or creating new ones.

SR. DATA ENGINEER

Microsoft
11.2021 - 11.2022
  • Design azure applications and pipelines for using flow for analyzing payment data using cosmos, spark, Databricks and snowflake build dashboards to provide insights for analyzing payment billing charge back and provider history
  • Design data modelling and schema development for payment approval, Cost of payment data, WorldPay/account update data cleansing and testing for various source Cloud services components like, storage, ADF framework, function app logic app and Synapse
  • As a senior lead developer, I am responsible for design and development of the payment applications for C+ E finance group accessing data engineering and data analytics project related to GPS/GF’s payment services PI and AU data
  • CLAMOR/DARSY services: - Preparing Asimov data for darsy portal submissions for third party apps development in cloud services. Work with Azure Functions Kusto query& EventHub
  • Metrics are stored in Azure and kusto database.
  • Reengineered existing ETL workflows to improve performance by identifying bottlenecks and optimizing code accordingly.

SR. DATA ENGINEER

Halliburton
03.2021 - 11.2021
  • Design azure applications and pipelines for large datasets using API, json format, for analyzing the fleet , fuel and pump using pyspark, Databricks and build dashboards to provide insights for analyzing fleet metrics
  • Build big data applications using Azure Data Factory, spark, python and Power BI which offers services to analyze and perform store performance. Use of Google Cloud Storage and BigQuery for extracts and API
  • Designed data bricks application for updating the history of engine performance data using PySpark. Design data modelling and schema development for pe asset data, data cleansing and testing for various source Cloud services components like, storage, ADF logic app and spark used to ingest, load by apply transformation logics for data. Build PowerBI reports for analytics like for data marts & cubes. Build and connect kafka sources for dwh data load to snowflake database
  • Languages and Technologies: Azure Data Factory, Web API, VSTF/TFS, DBT, GIT, PowerBI data bricks, Databricks, hive, Azure Functions Kusto query& EventHub, Azure SQL Polybase, BLOB, Data lake, U-SQL,COSMOSDB logic apps, EventHub, data bricks, Azure DevOPS, Jenkins, Azure key vaults and python

SR. BIG DATA LEAD ENGINEER

Ahold-Delhaize USA
07.2018 - 03.2021
  • As a senior lead developer, I am responsible for design and development of the applications and code deployments. build big data applications using Azure Data Factory, spark, U-SQL, python and Power shell and Web APIs which offers services to analyze and perform store performance. Use of Google Cloud Storage and BigQuery for extracts and API calls. Designed data bricks application using PySpark. Design data modelling and schema development, data cleansing and testing for various sources FTP, Oracle (EDW). Cloud services components like, storage, ADF and spark used to ingest, load by apply transformation logics for data. Build PowerBI reports for analytics like for data marts & cubes. Develop and integrate customer data MDM and marketing campaign with salesforce.
  • Languages and Technologies: Azure Data Factory, snowflake, Data bricks pipelines, Web API, Databricks, hive, Azure Functions Kusto query& EventHub, Azure SQL Polybase, BLOB, Data lake, U-SQL, logic apps, EventHub, data bricks, Azure DevOPS, Jenkins, Azure key vaults and python

SR. SOFTWARE DEVELOPMENT ENGINEER

Microsoft
06.2013 - 07.2018
  • C & E finance: Building Data warehouse and data lake analytics: - Build MS finance analytics warehouse using Azure services like data lake storage, U-SQL and Power shell for MS product sales insights and KPI’s by feeding different areas of sources using Azure ADF and spark/python
  • Surface Device Telemetry &Analytics: - Preparing Asimov data for surface devices. Metrics are stored in Azure database. Data is processed and loaded using Cosmos scripts and Xflow configuration. Wrote DAX queries to derive datasets and build Power BI dashboards to visualize.
  • Web Analytics: - Office Max Team is assisting Office product website through web analytics. Such as how many page searches, legitimate page visits help to broaden the development of site. Data is tracked via fiddler events in COSMOS cluster. Develop and build ETL pipelines using SSIS and stored procedures to track history of those events. Once loaded to SQL data warehouse build Tabular cube for analyzing its trends
  • Device Configuration: Use Power shell scripts to re-home, includes install, configure windows and database tools for devices to be used for RISK analytics and reporting
  • MSG Campaign Tools: - Design and developed tool named OSCAR database, for tracking machines for vulnerability which uses Global Foundation
  • Technologies: SQL Server 2010, T-SQL, SSIS, Data warehouse, SSAS, SSRS, Power shell
  • Delivered reliable software solutions on time and within budget, meeting project milestones consistently.
  • Designed user-friendly interfaces for optimal usability, enhancing end-user experiences.

SR. DATABASE ENGINEER

Disney- ABC TV Network
11.2009 - 02.2013
  • ETL Architecture, Developing Source to target data mapping (STDM) document defining transformation logic for SAP data using SSIS and SQL stored procedures. Translated business logic to transformation logic to generate pseudo code for the ETL process
  • Disney Consumer Products: - Develop Online Product Approval (OPA) application database using to manage Product Lifecycle Management (PLM) system used to manage the approval lifecycle for Disney, Marvel, and ESPN consumer products

Education

Master of Science in Engineering - Mechatronics

RWTH Aachen

Bachelor of Science and Engineering - Mechanical

JNTU Anantapur

Skills

  • Power shell & shell Programming &T-SQL
  • Spark RDD, data frame, spark different contexts, UDF & notebook
  • Azure data factory, data lake, storage
  • Linux and windows Server Platforms
  • Hive, beeline, Ambari, Azure HDInsight, Data Lake, Blob storage
  • Data Bricks, pyspark, json, Parquet
  • Power BI, ADF, U-SQL
  • Python, NumPy & matplotlib
  • Azure DevOps, Git, YAML, Jenkins
  • Sales Force marketing cloud, data extracts and campaign data
  • Snowflake DWH, variant stream and JSON load
  • Kafka connecter and Topic creation and troubleshoot data load for azure storage
  • Data warehouse modelling, SQL DW (Azure synapse), play base
  • Build releases & integration
  • Compute and scale application processing
  • Azure Functions Kusto query& EventHub
  • Integrate snow pipe with Azure, Data mapping and stage& integration
  • Build AS cubes and deploy using VSTS as well as CICD
  • Python programming
  • ETL development
  • Performance tuning
  • Data warehousing
  • Advanced SQL
  • Machine learning
  • Data analysis
  • Data science research methods
  • Statistical analysis
  • Decision-making
  • Effective communication
  • Self motivation
  • Goal setting
  • Complex Problem-solving
  • Adaptability

Accomplishments

  • Supervised team of 6 staff members.
  • Documented and resolved best practice and run books which led to prioritize and complete projects in time.
  • Achieved smooth deployments through effectively designing and testing and CI/CD tasks.

Certification

  • SQL Server 2008 MCITP Certificate
  • Hadoop 2.x Administration
  • Python for Data science (MIT Institute/online)
  • Azure Databricks(spark) (edX)

Timeline

SR. DATA ENGINEER

EVERESTRE Insurance
11.2022 - Current

SR. DATA ENGINEER

Microsoft
11.2021 - 11.2022

SR. DATA ENGINEER

Halliburton
03.2021 - 11.2021

SR. BIG DATA LEAD ENGINEER

Ahold-Delhaize USA
07.2018 - 03.2021

SR. SOFTWARE DEVELOPMENT ENGINEER

Microsoft
06.2013 - 07.2018

SR. DATABASE ENGINEER

Disney- ABC TV Network
11.2009 - 02.2013

Bachelor of Science and Engineering - Mechanical

JNTU Anantapur

Master of Science in Engineering - Mechatronics

RWTH Aachen
Praveen Rao Gundavarapu