Summary
Overview
Work History
Education
Skills
Certification
Timeline
Accomplishments
Generic
Naveen Kumar

Naveen Kumar

Nashville,TN

Summary

Results-driven data professional with over seven years of experience and dual master’s degrees in Data Science and Artificial Intelligence from Arkansas State University. Proven track record in streamlining ETL processes, optimizing data workflows, and implementing robust CI/CD pipelines to enhance deployment efficiency. Successfully led the development of a customer model project while delivering three additional production-ready projects on schedule, consistently meeting all project specifications. Committed to leveraging analytical skills and technical expertise to drive data-driven decision-making and operational excellence.

Overview

8
8
years of professional experience
1
1
Certificate

Work History

Senior Data Engineer (CVS)

TCS
02.2024 - Current
  • Decreased computational costs by 20% by optimizing model performance and streamlining data processing workflows.
  • Implemented a scalable data pipeline that processed and analyzed terabytes of data, reducing computation time by 30% compared to previous methods.
  • Implemented GDPR-compliant data governance practices, enhancing data security and reducing compliance risk by 30% across multiple ad-related projects.
  • Architected and optimized data models for ad performance analysis, leading to a 50% improvement in data retrieval speed and a 20% increase in targeting accuracy.
  • Collaborated with data scientists and machine learning engineers to create data products that improved predictive accuracy for ad targeting models by 15%, resulting in a 25% boost in campaign effectiveness.
  • Designed and Implemented a real time analytics system using Apache Kafka and Flint to monitor ad performance metrics, reducing decision-making time by 50% and improving ad placement efficiency by 30%.
  • Leveraged machine learning algorithms developed ad targeting models, resulting in a 20% improvement in audience segmentation and a 15% increase in ad engagement.
  • Developed a scalable data infrastructure using AWS and Apache Spark, capable of processing petabytes of ad data, which improved data processing speed by 50% and enables real-time ad performance insights.
  • Led the automation of deployment pipelines with CI/CD tools like Bamboo and Github, decreasing deployment time by 70% and increasing team productivity by 60%.

Data Engineer (AETNA)

Tekinvaderz
08.2021 - 01.2024
  • Automated data ETL processes by configuring and deploying AWS Glue services and implementing AWS Lambda functions and achieved a 70% reduction in data processing time.
  • Architected and automated deployment pipelines using Bamboo and GitLab CI/CD tools; decreased average deployment time by 70% and accelerated system performance by 25%.
  • Developed a data ingestion pipeline using Infoworks and GCP for a healthcare project that enhanced data retrieval and analysis time by 50%, leading to better patient outcomes. Integrated sources like Salesforce, Kafka, Hadoop to Infoworks and migrated to GCP.
  • Spearheaded the migration of 500+ tables from Hive/Hadoop to Big Query in GCP and implemented schema design changes and achieved a 50% improvement in data processing efficiency.
  • Scheduled CRON jobs as a part of manual trigger to pipelines and maintained ran daily, weekly, and quarterly pipelines.
  • Advised the implementation of AWS Secret Manager for 7 different departments and improved team efficiency by 40% and increased time for customer-facing tasks by 30%.
  • Developed a real time data processing pipeline for analyzing ad performance, resulting in a 20% improvement in targeting accuracy.
  • Developed and Automated Scripts for Merge Queries, which resulted in a productivity increase of over by 60% by reducing manual data integration tasks and minimizing errors.

Graduate Assistant

Arkansas State University
06.2020 - 08.2021
  • Ensured data accuracy, consistency, and reliability by analyzing data from different sources, and utilized data cleansing techniques to eliminate errors, resulting in an increase in data accuracy.
  • Directed Tableau dashboard implementation project for a 6-month period, steering a team of 2 developers and reducing report creation time by 50%.
  • Analyzed data sets and utilized data mapping tools to identify and rectify data discrepancies across systems.
  • Identified and eliminated data inconsistencies by creating and implementing a comprehensive data validation strategy which dropped data errors by 50% and accelerated data integrity by 20%.
  • Built SQL queries and created test data for Salesforce and Informatica cloud mappings to improve the accuracy of data migration and aided in the reduction of data faults by 40% and saved 15+ hours per week.

Data Engineer

Infosys Limited
02.2018 - 12.2019
  • Mastered comprehensive Python programming language and libraries through a rigorous 3-month training program; developed and deployed an image recognition algorithm that reduced manual image sorting by 80%.
  • Guided HQL best practices for the team to improve data accuracy and reduce query errors, increased team efficiency.
  • Authored a series of SQL Joins and Views to combine data from sources, which provided a 360-degree view of customer behavior and led to a 25% increase in customer lifetime value.
  • Acquired proficiency in SQL and data structures through a self-paced online course. Analyzed and optimized a 5 million-record database, reducing query response time by 70%.
  • Learned the fundamentals of Hive and Hadoop clusters through a 6-week intensive training program and started a data pipeline that processed 50+GB of data daily, reducing processing time by 60%.
  • Gained hands-on experience with AWS services like EC2, S3, and DynamoDB through a 6-month apprenticeship. Architected and deployed a web application that handled 1000+ concurrent users, reducing infrastructure costs by 40%.

Education

MS - Data Science and Artificial Intelligence

Campbellsville University
Louisville, KY
10.2025

MS - Computer Science

Arkansas State University
Jonesboro, GA
08.2021

Bachelor of Technology - Mechanical Engineering

Hindustan University
Chennai
06.2017

Skills

  • Tools/ Technologies/ Techniques: Python (NumPy, Panda, PySpark, Data Structures, Matplotlib, Cassandra, Pyscopg2), SQL, R, Oracle, MS- SQL Server, Snowflake, PostgreSQL, Apache Cassandra (CQL), Hive, AWS Glue, Lamba, S3, Step Functions, ECS, Secret Manager, Power BI, GIT, SSIS, Airflow, Microsoft Visio, SAS (JMP & Studio), RESTful services, Flask API, Exploratory data analysis, ETL, Data warehousing, Data modeling and mining, Predictive modeling, SDLC, Agile, DBT,ELT

Certification

GCP – Associate Cloud Engineer.

Timeline

Senior Data Engineer (CVS)

TCS
02.2024 - Current

Data Engineer (AETNA)

Tekinvaderz
08.2021 - 01.2024

Graduate Assistant

Arkansas State University
06.2020 - 08.2021

Data Engineer

Infosys Limited
02.2018 - 12.2019

MS - Data Science and Artificial Intelligence

Campbellsville University

MS - Computer Science

Arkansas State University

Bachelor of Technology - Mechanical Engineering

Hindustan University

Accomplishments

· Optimized enterprise data pipelines to enhance computational performance by 30% and reduce operational costs by 20% across large-scale healthcare systems.

· Spearheaded cloud migration initiatives, transitioning 500+ legacy Hive and Hadoop datasets to GCP BigQuery, improving scalability and query efficiency by 50%.

· Designed and implemented real-time analytics frameworks leveraging Kafka and Flint, accelerating business decision-making speed by 50%.

· Streamlined CI/CD delivery pipelines through GitLab and Bamboo automation, reducing release cycle time by 70% and increasing deployment reliability.

· Strengthened data governance and security posture by deploying AWS Secret Manager across multiple business units, improving compliance and data protection by 40%.

· Championed business intelligence transformation by introducing interactive Tableau dashboards, cutting manual reporting time by 50%.

· Partnered with Machine Learning teams to enhance ad targeting precision by 15%, resulting in a 25% uplift in campaign ROI.