Summary
Overview
Work History
Education
Skills
Accomplishments
Affiliations
Languages
Interests
Timeline
Generic

Praneeth Isukapalli

Wendell,NC

Summary

Results-focused data professional equipped for impactful contributions. Expertise in designing, building, and optimizing complex data pipelines and ETL processes. Strong in SQL, Python, and cloud platforms, ensuring seamless data integration and robust data solutions. Known for excelling in collaborative environments, adapting swiftly to evolving needs, and driving team success.

Overview

7
7
years of professional experience

Work History

Data Engineer

CapitalOne
03.2025 - Current
  • Designed and built scalable data pipelines using PySpark to process large structured and unstructured datasets.
  • Created end-to-end ETL workflows using AWS Glue to ingest, clean, and transform data from S3, RDS, and streaming sources.
  • Wrote complex SQL queries (joins, CTEs, window functions) for analytics, reporting, and data validation.
  • Developed Python-based ETL frameworks using modular, reusable functions and object-oriented design.
  • Used Python libraries such as pandas, numpy, boto3, json, and logging for data processing and AWS integration.
  • Built Python scripts for automated data validation, schema checks, null handling, and reconciliation reports.
  • Implemented Python utilities for file ingestion, metadata extraction, and error handling across pipelines.
  • Automated batch and near-real-time workflows using Python to reduce manual intervention and failures.
  • Built Python-based preprocessing pipelines to clean and prepare text data for LLM and GenAI applications.
  • Generated embeddings and metadata using Python for RAG pipelines and vector database ingestion.
  • Used Copilot AI and GenAI tools to accelerate Python and Spark development.
  • Built and optimized Hive external tables on S3 using Parquet and ORC formats.
  • Designed real-time ingestion pipelines using AWS Kinesis for analytics and AI workloads.
  • Developed AWS Lambda functions using Python for event-driven processing and automation.
  • Tuned Spark jobs by optimizing partitions, caching, and shuffle operations.
  • Implemented strong data quality and monitoring using Python-based checks and CloudWatch.
  • Ensured secure access using AWS IAM roles and supported CI/CD pipelines and documentation.
  • Technologies Used: Python, SQL, PySpark, Spark, pandas, numpy, boto3, AWS Glue, S3, Lambda, Kinesis, Athena, Hive, GenAI, LLMs, RAG, Vector Databases, Copilot AI, Git, CloudWatch, Linux

Data Engineer

Walmart
02.2023 - 02.2025
  • Designed and built large-scale ETL pipelines using PySpark, Python, and AWS Glue for retail datasets.
  • Ingested data from POS systems, online platforms, vendor feeds, and internal applications into an S3 data lake.
  • Wrote optimized SQL queries to support reporting, forecasting, and analytics teams.
  • Developed Python-based data ingestion and transformation scripts for batch and streaming pipelines.
  • Used pandas and PySpark together to perform data cleansing, aggregations, and business rule validations.
  • Created Python utilities for automated file validation, duplicate detection, and exception reporting.
  • Implemented Python-driven data quality frameworks including schema validation and reconciliation logic.
  • Built scalable Spark transformations for sales metrics, inventory movement, and store-level insights.
  • Prepared structured and unstructured datasets using Python for GenAI and LLM-based analytics use cases.
  • Built RAG-ready datasets by generating embeddings and enriching metadata using Python workflows.
  • Designed near-real-time ingestion using AWS Kinesis for online orders and pricing events.
  • Developed event-driven workflows using AWS Lambda (Python) and Step Functions for orchestration.
  • Created Athena tables and views to enable self-service analytics.
  • Improved Spark performance using optimized joins, partitioning, caching, and efficient file formats.
  • Migrated legacy Hadoop and Hive pipelines to AWS Glue and PySpark.
  • Monitored and resolved pipeline failures using Python logs, CloudWatch, and Airflow.
  • Documented Python modules, ETL logic, and operational runbooks.
  • Technologies Used: Python, SQL, Apache Spark, PySpark, Hadoop, Hive, HDFS, AWS S3, AWS Glue, AWS Glue Workflows, AWS EMR, AWS Lambda, AWS Step Functions, AWS Kinesis, AWS Athena, DynamoDB, Airflow, Terraform, Databricks, Docker, Parquet, ORC, Git, Jenkins/GitLab CI/CD, CloudWatch, Linux/Unix

Software Development Engineer

Amazon, Alexa Speech
07.2022 - 01.2023
  • Spearheaded the enhancement of the Context Model Service component in Automatic Speech Recognition, focusing on building contextual artifacts to significantly enhance recognition. The team's key objectives centered around achieving heightened accuracy and minimizing latency.
  • Implemented critical code changes to a cross-region dual-write client, ensuring the enforcement of regionalized startup behavior. This involved intricate modifications to optimize the client's functionality across multiple regions.
  • Led an investigation into a bug within the SQS queue DeleterWorker, proactively identifying and rectifying the issue. Additionally, introduced metrics to monitor this specific type of workflow, enhancing the team's ability to diagnose and address similar challenges promptly.
  • Actively participated in on-call support responsibilities, demonstrating proficiency in deploying changes to the CI/CD pipeline. Monitored logs through CloudWatch, employing a keen eye for detail to identify and address root causes of issues, contributing to seamless system performance.
  • Applied AWS expertise to create and manage snapshots, volumes, and security groups for both public and private instances within EC2, demonstrating adept knowledge in cloud infrastructure.
  • Provided steadfast maintenance and support for AWS infrastructure, catering to both client production and QA/Development needs, ensuring seamless operations.
  • Utilized AWS S3 buckets as a strategic solution for online backup data archiving. Leveraged Amazon Web Services (AWS) tools like RDS, CloudWatch, and CloudFront for the efficient deployment of code across diverse environments.
  • Exhibited hands-on proficiency in working with Continuous Integration (CI) build-automation tools, including Maven, Jenkins, and Apache Ant, streamlining development processes, and ensuring the reliability of code integration.
  • Held responsibility for the robust maintenance and expansion of AWS Cloud Services infrastructure, employing AWS tools such as SNS and SQS to optimize communication and resource allocation within the cloud environment.
  • Technologies Used: Java, Google Guice, Git and Brazil, Apollo, Linux commands, CloudWatch etc.

DevOps Cloud Engineer

Origin Hubs Inc
01.2022 - 05.2022
  • Comprehensive exposure to full-stack Java programming, leveraging skills in both front-end and back-end development.
  • Spearheaded the creation of pivotal web page features using ReactJS, including the successful implementation of backlog clearing functionalities and the integration of a CI/CD pipeline approval feature.
  • Demonstrated proficiency with ReactJS's latest features such as Error Boundaries, Fragments, and new lifecycle methods, enhancing code robustness and efficiency.
  • Applied React fundamentals, managing props, state, keys, refs, events, and utilizing the fetch API for server communication.
  • Implemented effective routing using React-Router, employing Browser Router for web applications and Native Router for mobile platforms.
  • Integrated various third-party React components, including react-video, react-typeahead, react-tags input, react-DataGrid, react-calendar, react-tabs, and react-autocomplete.
  • Developed cloud microservices, implementing the back end using Spring Boot, contributing to a modularized and scalable architecture.
  • Employed Microservice architecture with Spring Boot, modularizing code, implementing REST APIs with Spring Rest, and integrating Swagger API for comprehensive documentation.
  • Exhibited a solid understanding of implementing and consuming SOAP and RESTful web services, contributing to seamless communication between components.
  • Implemented domain base classes utilizing Spring Data and leveraged Java 8 features like Functional Interfaces, Stream API, Time API, and Concurrency API.
  • Developed a feature-rich UI using HTML5, CSS3, JSP, and JavaScript, ensuring interactive cross-browser functionality and a sophisticated user interface.
  • Utilized Spring Boot-based services for RESTful interactions and Apache Kafka message brokers, implementing the DAO layer with Spring and Hibernate.
  • Applied various Spring modules including Spring Core, Spring Security, Spring AOP, Spring MVC, Spring JDBC, and Spring Batch, contributing to the comprehensive development of the application.
  • Technologies Used: React.js, Redux, React-Router, ES6, ES7, Webpack, Babel, Gerrit, Git, Jenkins, Maven, AWS, HTML5, CSS3, Bootstrap, Spring 4.x, Spring Boot, Microservices architecture.

Associate Software Engineer

FMIMI Pvt Ltd
10.2019 - 11.2020
  • Involved in the design, implementation, and deployment of Full Software Development Life Cycle (SDLC) of the project.
  • Implemented microservices architecture using Spring boot Framework and Spring REST to create REST endpoints and secured the API's by implementing Oauth2 token-based authentication/authorization scheme using Spring Security.
  • Worked on writing unit test cases using Junit and Mockito and continuous integration testing using Jenkins in a TDD approach and used Swagger for REST API documentation.
  • The REST-based Microservices were designed and implemented using Spring Boot, Spring Data with JPA, Spring AOP, and Spring DAO to offer complete CRUD capabilities.
  • Developed CI/CD pipelines with Jenkins on a Kubernetes environment to build, test, and deploy.
  • Technologies Used: Java, Spring Boot, Spring, Hibernate, RESTFUL, Log 4J, JSON, XML, Mongo dB, GIT, Postman, Swagger, JENKINS, JUnit, Mockito, MAVEN, Lambda, EC2, S3.

Associate Software Engineer

CGI
08.2018 - 09.2019
  • Adept in Java, Python, and C++, I have demonstrated proficiency by successfully delivering 10 projects with a strong emphasis on high code accuracy and on-time delivery.
  • Played an active role in the e-commerce team, contributing significantly to the design of software algorithms aimed at enhancing product-user matching. This strategic involvement resulted in a notable 20% increase in sales revenue over a 6-month period.
  • Efficiently harnessed software development tools like Git, JIRA, and IntelliJ to streamline development processes. This approach not only improved development speed but also led to a substantial reduction in bug density across 8 projects.
  • Led code reviews for 15 projects, actively seeking and incorporating valuable feedback from Senior Engineers and team members. This collaborative effort resulted in a tangible reduction in code defects and a marked improvement in overall code quality.
  • Successfully crafted custom, scalable code for 12 applications, showcasing a measurable improvement in application performance and a reduction in server resource consumption.
  • Technologies Used: Java, Python, C++, Git, JIRA, IntelliJ

Education

Master of Science - Computer Science

University of Central Missouri (UCM)
Lee’s Summit, MO
05.2022

Bachelor of Technologies - Computer Science and Engineering

Aditya Engineering College
Kakinada, India
05.2019

Skills

  • Programming Languages: C, Java, Python
  • Web Technologies: HTML5, CSS, JavaScript, jQuery
  • Frameworks: Spring Boot, Spring, Hibernate, JPA
  • Database Technologies: MySQL, PostgreSQL, DynamoDB
  • Source Code Management: Git, GitHub, Gitlab
  • Web Server: Apache, Tomcat
  • Testing Frameworks: Junit, Mockito, TestNG
  • Cloud and DevOps Tools: AWS, Azure, Docker, Kubernetes, Jenkins, Puppet, Chef, Splunk, Terraform, Ansible

Accomplishments

  • Collaborated with a team of 6 engineers in the development of scalable AWS-based data pipelines for analytics and GenAI use cases.
  • Documented and resolved data pipeline failures and schema mismatch issues, which led to improved data reliability and reduced ETL job failures by 30%.
  • Supervised and mentored 3 junior engineers by reviewing code, validating data pipelines, and supporting production deployments.
  • Achieved faster development and reduced coding errors by introducing Git-based CI/CD and Copilot AI for ETL and Spark development tasks.
  • Achieved timely and accurate delivery by completing large-scale data transformations and validations with high accuracy and efficiency.
  • Resolved data quality and reporting issues through validation testing and collaboration with downstream analytics and business users.
  • Achieved improved system stability through effectively helping with production support, on-call issue resolution, and root cause analysis.
  • Used Microsoft Excel to develop inventory and reconciliation tracking spreadsheets for validating ingested retail and supply-chain data.

Affiliations

  • Toastmasters
  • Lions Club

Languages

English
Full Professional
Hindi
Full Professional

Interests

  • Cooking
  • Gym Workouts
  • Music
  • App Development

Timeline

Data Engineer

CapitalOne
03.2025 - Current

Data Engineer

Walmart
02.2023 - 02.2025

Software Development Engineer

Amazon, Alexa Speech
07.2022 - 01.2023

DevOps Cloud Engineer

Origin Hubs Inc
01.2022 - 05.2022

Associate Software Engineer

FMIMI Pvt Ltd
10.2019 - 11.2020

Associate Software Engineer

CGI
08.2018 - 09.2019

Bachelor of Technologies - Computer Science and Engineering

Aditya Engineering College

Master of Science - Computer Science

University of Central Missouri (UCM)
Praneeth Isukapalli