Summary
Overview
Work History
Education
Skills
Projects
Timeline
Generic

MANNAVA VINOD KUMAR

Frisco,CO

Summary

Responsive expert experienced in monitoring database performance, troubleshooting issues and optimizing database environment. Possesses strong analytical skills, excellent problem-solving abilities, and deep understanding of database technologies and systems. Equally confident working independently and collaboratively as needed and utilizing excellent communication skills.

Overview

5
5
years of professional experience

Work History

Azure Data Engineer

Bottomline Technologies
08.2021 - 07.2022
  • Designed and implemented database solutions in Azure SQL Data Warehouse and Azure SQL Database
  • Implemented ETL and data movement solutions using Azure Data Factory and SSIS
  • Extracted, transformed, and loaded data from various source systems to Azure Data Storage services
  • Developed scalable data processing pipelines using PySpark for big data analytics
  • Designed and implemented ETL pipelines using Azure DataFactory for data movement and transformation
  • Developed and managed data analytics workflows using Databricks on Azure
  • Managed and orchestrated data transformations in DBT with SQL-based models
  • Conducted data modeling and schema design to support analytical workloads in Snowflake
  • Planned and executed data migration projects, transferring data from on-premises systems to Azure
  • Implemented data security measures to protect sensitive information in accordance with company policies
  • Optimized database and ETL processes for performance improvement and conducted regular performance monitoring and tuning
  • Collaborated with cross-functional teams, including data analysts, developers, and business stakeholders, to gather requirements and deliver solutions
  • Stayed updated with the latest Azure technologies and best practices, implementing innovative solutions to enhance data management and processing capabilities.

Azure Data Engineer

Capgemini
01.2020 - 07.2021
  • Developed data pipelines using Spark, Hive, and HBase for customer behavioral data analysis
  • Implemented data ingestion from various source systems using Sqoop and PySpark
  • Developed robust data pipelines using PySpark to handle large-scale data processing tasks and ensured efficient data ingestion, transformation, and loading from various source systems
  • Developed and maintained data architecture blueprints, standards, and guidelines for Azure-based data solutions
  • Automated data movement and transformation processes to support continuous data integration and delivery
  • Developed and maintained Databricks notebooks for ETL tasks and data analysis
  • Wrote and optimized complex SQL queries for data extraction, transformation, and reporting
  • Designed and implemented Snowflake data warehousing solutions to support analytical reporting
  • Documented all processes, including PySpark scripts, ADF pipelines, Databricks workflows, DBT transformations, and Snowflake integrations, to ensure knowledge sharing and project transparency
  • Provided support for troubleshooting and resolving issues related to PySpark, ADF, Databricks, SQL, DBT, and Snowflake
  • Implemented enhancements to increase efficiency, reduce costs, and improve data quality
  • Conducted code reviews, mentored junior team members, and provided technical guidance and support to project teams
  • Worked closely with stakeholders to understand business requirements, translate them into technical solutions, and ensure successful delivery of projects within defined timelines and budget constraints.

Data Consultant

Accenture
05.2018 - 12.2019
  • Orchestrated streamlined data workflows using AWS Data Pipeline and Lambda
  • Designed and implemented a Snowflake Schema for comprehensive data analysis
  • Leveraged AWS services like Redshift, Glue, and EMR for data processing and optimized Redshift performance through efficient query design and data distribution strategies
  • Integrated data from various source systems into AWS data storage and processing services
  • Automated data ingestion and transformation tasks using Glue jobs
  • Ensured data consistency, integrity, and accuracy during the integration process
  • Conducted performance tuning for AWS Data Pipeline, Redshift, Glue, and EMR jobs
  • Developed data visualization dashboards and reports using tools like Tableau, Power BI, or Qlik to facilitate data-driven decision-making
  • Conducted data quality assessments and implemented improvement initiatives to enhance data accuracy and reliability
  • Conducted workshops and training sessions to educate client teams on data analytics tools, techniques, and best practices
  • Created detailed documentation for data workflows, ETL processes, and data architecture.

Intern/Hadoop Developer

E - Emphasys
08.2017 - 04.2018
  • Developed and maintained robust data pipelines using Apache Hadoop within a Spark environment on AWS
  • Utilized Apache Hive for data warehousing and analysis in Spark projects on AWS
  • Managed Hadoop clusters efficiently on AWS using Apache Ambari
  • Assisted in the design, development, and implementation of Hadoop-based solutions for data ingestion, processing, and analysis
  • Supported senior developers in optimizing Hadoop cluster performance and capacity planning
  • Participated in troubleshooting and debugging Hadoop ecosystem components such as HDFS, YARN, and MapReduce jobs
  • Contributed to the development and maintenance of ETL processes to integrate data from various sources into Hadoop using tools like Sqoop and Flume
  • Collaborated with data scientists and analysts to understand data requirements and develop scalable solutions
  • Assisted in writing and optimizing Pig, Hive, and Spark scripts for data transformations and analytics
  • Contributed to the creation and execution of data quality checks and data validation processes within Hadoop environments
  • Worked with Hadoop cluster administrators to deploy patches and upgrades, ensuring system security and stability
  • Documented technical specifications, system procedures, and user guides for Hadoop applications and processes.

Education

Masters in Computer/Information Technology services administration and management -

Concordia University, Saint Paul

Bachelor of Technology, Computer Science and Engineering -

NRI INSTITUTE OF TECHNOLOGY

Skills

  • Power BI
  • Tableau
  • SQL
  • Python
  • R
  • Java
  • Hadoop
  • Spark
  • Kafka
  • Hive
  • HBase
  • Azure Data Factory
  • Informatica
  • Talend
  • Apache NiFi
  • SSIS

Projects

Predictive Maintenance System for Industrial Equipment, 02/2023, Integrated models into operational workflows using Azure Functions., Implemented Azure AutoML to train predictive maintenance models., Developed ETL pipelines to extract sensor data from Azure Blob. Anomaly Detection System for IoT Sensor Data, 04/2023, Designed scalable data ingestion pipeline using Azure IoT Hub., Utilized Azure Machine Learning for anomaly detection model training., Integrated real-time alerting with Azure Functions for anomaly notifications.

Timeline

Azure Data Engineer

Bottomline Technologies
08.2021 - 07.2022

Azure Data Engineer

Capgemini
01.2020 - 07.2021

Data Consultant

Accenture
05.2018 - 12.2019

Intern/Hadoop Developer

E - Emphasys
08.2017 - 04.2018

Masters in Computer/Information Technology services administration and management -

Concordia University, Saint Paul

Bachelor of Technology, Computer Science and Engineering -

NRI INSTITUTE OF TECHNOLOGY
MANNAVA VINOD KUMAR