Summary
Overview
Work History
Education
Skills
Certification
Timeline
Generic

CHAITHANYA SAGAR GOTTIMUKKULA

Frisco,TX

Summary

Over 9 years of experience in Information Technology, worked on Data warehousing and providing solutions, develop, maintain and support the client requirements. Microsoft certified: Azure Data Engineer Associate Snowflake’s SnowPro Core Certified Experience as an Azure Data Engineer in various Microsoft Azure Cloud technologies including Azure Data Factory (ADF), Azure Data Lake Storage (ADLS), Azure Data Box, Azure SQL, Azure Data bricks, Azure Cosmos DB, Azure Key vaults and Azure Analytical services Experience in creating pipelines in ADF using activities such as Move & Transform, Copy, filter, for each, Get Metadata, Lookup etc. Experience working with different file formats like , csv, xml etc., In-depth knowledge and Hands on experience implementing cloud data lake like Azure Data Lake Gen1 and Azure Data Lake Gen2. Experience in transforming data from one server to other servers using tools like Bulk Copy Program (BCP), SSIS and Azure Data Factory Worked on latest reporting tools like Power BI to create advanced visualizations to give appropriate insights and support the business. Expert in Coding SQL and PL/SQL like Stored Procedures, Functions and Triggers. Expert in databases like SQL Server, Oracle, Snowflake and BigQuery. Hands-on experience with Oracle databases (e.g., 11g, 12c, 19c) and their integration with PL/SQL. Worked extensively on Google Cloud Platform (GCP)’s BigQuery by creating tables and loading data from different source tables. Experienced in scheduling and process automation using Shell/Python scripting. Optimized and tuned ETL processes & SQL Queries for better performance. Implemented Optimization techniques for better performance on the ETL side and also on the database side. Experienced with SQL Server Migration Assistant (SSMA) to migrate database objects from other databases to SQL Server. Experience in developing SQL scripts to validate the databases tables and reports data for backend database testing. Hands-on experience with Amazon EC2, Amazon S3, Amazon RDS, Redshift, VPC, IAM, Amazon Elastic Load Balancing, Auto Scaling, Cloud Front, CloudWatch, SNS, SES, SQS and other services of the AWS family. Experienced in both Agile as well as Waterfall methodologies. An excellent team member with an ability to perform individually, good interpersonal relations, strong communication skills, hardworking and high level of motivation.

Overview

9
9
years of professional experience
1
1
Certification

Work History

Azure Data Engineer

Albertsons Companies Inc.
09.2021 - Current
  • Contributed to the full Software Development Life Cycle (SDLC), engaging in Requirement gathering, Business Analysis, Design, Development, Testing and implementation of business rules
  • Worked on creating tables in Snowflake and BigQuery for Customer360 and loaded data into them using various sources tables
  • Extract Transform and Load data from Sources Systems to Azure Data Storage services using a combination of Azure Data Factory, T-SQL
  • Data Ingestion to one or more Azure Services - (Azure Data Lake, Azure Storage, Azure SQL)
  • Cultivated a comprehensive understanding of data sources, implemented robust data standards and upheld data quality
  • Implemented high-performance data transformations using Spark’s RDDs, DataFrames, and Datasets in Scala
  • Implemented Scala pipelines for efficient incremental loading of data into Snowflake and BigQuery tables, seamlessly integrated with Databricks jobs
  • Created Extract Transform and Load (ETL) processes, utilizing Azure Data Factory and T-SQL to extract data from Source Systems into Azure Data Storage Services
  • Orchestrated data ingestion into Azure Services like Azure Data Lake, Azure Storage, and Azure SQL for seamless data management
  • Developed and optimized Spark-based ETL pipelines in Azure Databricks for processing data
  • Utilized PySpark/Scala to build and fine-tune distributed data processing workflows
  • Worked on the optimization of long running jobs in BigQuery to reduce the slot contention
  • Utilized the Snowflake platform to create tables and efficiently loaded data from files by employing the 'COPY INTO' command
  • Loaded data into snowflake tables, leveraging various database, schemas, roles, and warehouses, including configurations such as small, big, large to optimize the data loading process
  • Spearheaded the creation of Azure Data Factory (ADF) pipelines with Linked Services/Datasets, facilitating the ETL of data from diverse sources like Oracle into Azure SQL, Azure Data Lake Storage and Azure Data Box
  • Applied SQLPlus for extracting substantial tables from Oracle to flat files and established continuous monitoring protocols
  • Automated the creation of scripts for streamlined data extraction from Oracle
  • Leveraged Jupyter Python notebooks to craft data validations reports based on Column name, Data type, Record Count and sample data validations
  • Orchestrated the migration of database objects, including procedures and tables, from Oracle to Azure SQL using SQL Server Migration Assistant (SSMA)
  • Designed and implemented high-performance PL/SQL scripts for ETL processes, reducing data processing time
  • Created tables and ingested data into them in Snowflake and BigQuery based on Business requirements
  • Gathered Business requirements and business rules from key client stakeholders
  • Environment: BigQuery, Oracle, Azure SQL, Snowflake, Azure Data Factory, Azure Data Lake, Azure Databox, Python, MS Excel, Jupyter, Microsoft Teams, SQL Server Migration Assistant (SSMA)

Azure Data Engineer

CareCentrix Inc.
09.2019 - 08.2021
  • Designed and Developed pipelines, data flows, complex data transformations and manipulations using Azure Data Factory (ADF) and PySpark with Databricks
  • Created, provisioned multiple Databricks clusters needed for batch and continuous streaming data processing and installed the required libraries for the clusters
  • Created Azure Data Factory pipelines to extract the data from Relational sources like Oracle, SQL Server, DB2 and non-relational sources like Flat files, JSON files, XML files, Shared folders etc
  • Developed streaming pipelines using Apache Spark with Python
  • Developed Azure Databricks notebooks to apply the business transformations and perform data cleansing operations
  • Develop Databricks Python notebooks to Join, filter, pre-aggregate, and process the files stored in Azure data lake storage
  • Ingested huge volume and variety of data from disparate source systems into Azure Data Lake Gen2 using Azure Data Factory
  • Created reusable pipelines in Data Factory to extract, transform and load data into Azure SQL DB and SQL Data warehouse
  • Implemented both ETL and ELT architectures in Azure using Data Factory, Databricks, SQL DB and SQL Data warehouse
  • Used Azure Logic Apps to develop workflows which can send alerts/notifications on different jobs in Azure
  • Used Azure Devops to build and release different versions of code in different environments
  • Developed UNIX scripts to automate different tasks involved as part of loading process
  • Automated jobs using Scheduled, Event based, Tumbling window triggers in ADF
  • Worked on real-time data processing and analytics solutions leveraging PL/SQL capabilities
  • Created External tables in Azure SQL Database for data visualization and reporting purpose
  • Create and setup self-hosted integration runtime on virtual machines to access private networks
  • Well-versed with Azure authentication mechanisms such as Service principal, Managed Identity, Key vaults
  • Environment: Azure SQL, Azure Data Factory, Databricks, Python, Apache Spark, MS Excel, Microsoft Teams, Power BI, Power Automate, Visual Studio 2014

Azure Data Engineer

Insurance Auto Auctions Inc. (IAA)
11.2018 - 09.2019
  • Involved in Technical and Business decisions for Business requirement, Interaction with Business Analysts, Client team, and Development team through Agile Kanban process
  • Creating Azure Data factories for loading the data to Azure SQL database from Cosmos platform
  • Acted as build and release engineer, deployed the services by VSTS (Azure DevOps) pipeline
  • Created and Maintained pipelines to manage the IAC for all the applications
  • Understand the latest features like (Azure DevOps, OMS, NSG Rules, etc..,) introduced by Microsoft Azure and utilized it for existing business applications
  • Created complex Power BI dashboards
  • Develop Databricks Python notebooks to Join, filter, pre-aggregate, and process the files stored in Azure data lake storage based on business logic
  • Performed Column Mapping, Data Mapping and Maintained Data Models and Data Dictionaries
  • Built system to perform real-time data processing using Spark streaming and Kafka
  • Involved in retrieving multi-million records for data loads using SSIS and by querying against Heterogeneous Data Sources like SQL Server, Oracle, Text files and some Legacy systems
  • Expertise in using different Transformations like Lookups, Derived Column, Merge Join, Fuzzy Lookup, For Loop, For Each Loop, Conditional Split, Union all, Script component etc
  • Transferred data from various data sources/business systems including MS Excel, MS Access, and Flat Files to SQL Server using SSIS/DTS packages using various features
  • Involved in Performance tuning of ETL transformations, data validations and stored procedures
  • Strong experience in designing and implementing ETL packages using SSIS for integrating data using OLE DB connection from heterogeneous sources
  • Created Complex ETL Packages using SSIS which upsets data from staging table to database tables
  • Developed and designed system to collect data from multiple portal using Kafka and then process it using spark
  • Extensively worked on UNIX Shell Scripting for splitting group of files to various small files
  • Analyzed the SQL scripts and designed it by using PySpark SQL for faster performance
  • Experience in creating reports from scratch using Power BI
  • Created roles using SSAS to restrict cube properties
  • Implemented cell level security in cubes using MDX expressions to restrict users of one region seeing data of another region using SSAS
  • Created calculated measures using MDX implementing business requirement
  • Experienced working on Star and Snowflake Schemas and used the fact and dimension tables to build the cubes, perform processing and deployed them to SSAS database
  • Designed aggregations and pre-calculations in SSAS
  • Involved in designing Partitions in Cubes to improve performance using SSAS
  • Experienced in Developing Power BI Reports and Dashboards from multiple data sources using Data Blending
  • Responsible for creating and changing the visualizations in Power BI reports and Dashboards on client requests
  • Created Calculated Columns and Measures in Power BI and Excel depending on the requirement using DAX queries
  • Optimized SQL queries and PL/SQL blocks to reduce execution time and resource consumption
  • Created hierarchies in Power BI reports using visualizations like Bar chart, Line chart, etc
  • Worked with both live and import data into Power BI for creating reports
  • Managed relationship between tables in Power BI using star schema
  • Used different type of slicers available in Power BI for creating reports
  • Environment: Azure SQL, Azure Data Factory, Kafka, SQL Server, MS Excel, Microsoft Teams, Kafka, Visual Studio 2014, SSIS, Power BI, UNIX
  • Collaborated on ETL (Extract, Transform, Load) tasks, maintaining data integrity and verifying pipeline stability.

Sr. BI Developer

Mr. Cooper
10.2016 - 07.2018
  • Built business intelligence and data visualization dashboards using various technologies such as Power BI and Power Automate
  • Designed SSIS Packages to transfer data between servers, load data into database, and archived data file from different DBMS using SQL enterprise manager/SSMS on SQL server
  • Involved in Normalization and De-Normalization of existing tables for faster query retrieval
  • Involved in designing Parameterized Reports for generating Ad-Hoc reports as per the client requirements
  • Used Python in manipulating structured and unstructured datasets
  • Gathered business requirements, definition and design of the data sources and data flows
  • Written SQL statements for retrieval of data and Involved in performance tuning of TSQL
  • Monitored and provided front-line support of daily processes
  • Helped create process logging and new monitoring tools, integrity reports, and mapping tools
  • Created SSIS packages for File Transfer from one location to the other using FTP task
  • Create and maintain SSIS packages to extract transform and load data into SQL Server
  • Experience in creating scheduling Jobs, Alerts, SQL Mail Agent, and scheduled DTS Packages
  • Involved in ETL architecture enhancements to increase the performance using query optimizer
  • Configured the loading of data into slowly changing dimensions using Slowly Changing Dimension wizard
  • Designed various SSIS modules in order to fetch the data in the data staging environment based on the different types of incoming data
  • Designed and created Report templates, bar graphs and pie charts based on the financial data
  • Delivered enterprise, web-enabled reports using SSRS
  • Implemented Database project solution for deploying database objects
  • Involved in requirement gathering, analysis, design, development & deployment
  • Configured and deployed different reports to the servers using SQL Server 2012 SSRS
  • Created complex stored procedures, triggers, cursors, tables, and views, using SQL
  • Performed data validation on the flat files that were generated in UNIX environment using UNIX commands as necessary
  • Worked with version control systems like Git for managing PL/SQL scripts and database changes
  • Experienced in SQL design, coding, testing and implementation
  • Developed and optimized stored procedures, views, and user-defined functions for the application
  • Environment: Azure SQL, Azure Data Factory, SQL Server 2016, Unix, Power BI, Power Automate, MS Excel, Microsoft Teams, Visual Studio 2014

MS SQL/BI Developer

Acaira Technologies
12.2015 - 10.2016
  • Created and executed SQL Server Integration Service packages to populate data from the various data sources, created packages for different data loading operations for many applications
  • Created SSIS Packages using SSIS Designer for export heterogeneous data from OLE DB Source (Oracle), Excel Spreadsheet to SQL Server 2008
  • Created logging for ETL load at package level and task level to log number of records processed by each Package and each task in a package using SSIS
  • Designed ETL packages dealing with different data sources (SQL Server, Flat Files, and XMLs etc.) and loaded the data into target data sources by performing different kinds of transformations using SQL Server Integration Services (SSIS)
  • Created Stored Procedures to import data from csv and text files tables using .fmt files
  • Used Merge to perform insert, update and delete operations on a target table based on the results of a join with a source table
  • Reviewed the existing Stored Procedures to increase the performance
  • Involved on migrating SQL Server databases to SQL Azure Database using SQL Azure Migration Wizard
  • Experience working on Azure SQL Databases and Azure SQL DataWarehouse
  • Exposure on Azure HDInsight Microsoft's Hadoop based service
  • Coordinated and worked with the business analysts and system analysts and defined technical specification documents for the defined business rules
  • Uploaded the created documents to Confluence so that the complete team can have access to it
  • Worked and fixed production issues assigned in JIRA
  • Attended daily Scrum meetings to provide status updates
  • Environment: MS SQL Server 2012/2008R2/2005, SSIS, SSRS, MS Azure, MS Excel, JIRA, SVN, GIT, Bit-Bucket, Confluence, Visual Studio 2012

Education

Master of Science - Management Information Systems

University of Mary Hardin Baylor
Belton, TX
05-2020

Skills

  • T-SQL
  • PL/SQL
  • HTML
  • XML
  • Oracle
  • Python
  • Java
  • Spark
  • PySpark
  • Scala
  • Unix
  • Shell
  • MS SQL server 2016
  • MS SQL server 2012
  • MS SQL server 2008 R2
  • MS SQL server 2005
  • Oracle 8i
  • Oracle 9i
  • Oracle 10g
  • Snowflake
  • PostgreSQL
  • AWS
  • ADFv2
  • Blob Storage
  • ADLS
  • Azure Data Box
  • Azure SQL DB
  • SQL server
  • Azure Synapse
  • Azure Analysis Services
  • Databricks
  • Azure Cosmos DB
  • Azure Stream Analytics
  • Azure Event Hub
  • Logic Apps
  • Event Grid
  • Azure DevOps
  • ARM Templates
  • Power BI
  • Windows
  • Linux
  • Azure Data Factory
  • ETL processes
  • Data transformation
  • Data storage
  • BigQuery optimization
  • Snowflake integration
  • Cloud data architecture
  • Apache Spark
  • Python programming
  • Data visualization
  • Business analysis
  • Agile methodology
  • ETL development
  • Data warehousing
  • Data modeling
  • Data pipeline design
  • Data migration
  • Big data processing
  • Scripting languages
  • Spark framework
  • Performance tuning
  • Machine learning
  • Data pipeline control
  • Data integration
  • SQL and databases
  • SQL programming
  • Business intelligence
  • Database design
  • Relational databases
  • RDBMS
  • Query optimization
  • Database development
  • Data warehousing expertise
  • RDMS design

Certification

  • Microsoft certified: Azure Data Engineer Associate
  • Snowflake’s SnowPro Core Certified

Timeline

Azure Data Engineer

Albertsons Companies Inc.
09.2021 - Current

Azure Data Engineer

CareCentrix Inc.
09.2019 - 08.2021

Azure Data Engineer

Insurance Auto Auctions Inc. (IAA)
11.2018 - 09.2019

Sr. BI Developer

Mr. Cooper
10.2016 - 07.2018

MS SQL/BI Developer

Acaira Technologies
12.2015 - 10.2016

Master of Science - Management Information Systems

University of Mary Hardin Baylor
CHAITHANYA SAGAR GOTTIMUKKULA