Summary
Overview
Work History
Education
Skills
Key Skills and Expertise:
Certification
Timeline
Generic

Ponakala Sai Nandan

Dallas,TX

Summary

Experienced Data Engineer with 4+ years of expertise in Data Analysis and Data Engineering, including working with large datasets (structured and unstructured), data acquisition, data validation, predictive modeling, and data visualization. Proficient in designing and deploying business intelligence, analytics, and reporting solutions using advanced data analysis techniques and software development skills. Adept at using databases, programming languages, and business intelligence platforms to build scalable data pipelines and actionable dashboards.

Overview

5
5
years of professional experience
1
1
Certification

Work History

Data Engineer

Capital One
Plano, TX
07.2022 - Current

1. Project: Cloud Data Migration & Analysis for Dealer Data Transition.

Description:

Worked the end-to-end dealer data migration from the SHAW system to the AWS-based POTF platform, utilizing Snowflake for data warehousing and analysis. Built and automated scalable ETL processes using Python, Arrow, and AWS services to ensure seamless data ingestion. Developed advanced Tableau dashboards for visualizing migration progress and business insights. Worked closely with stakeholders to ensure technical solutions aligned with business objectives, improving data accessibility and insights.

Responsibilities:

  • Designed and automated data pipelines using Python and AWS Lambda, leveraging Arrow to handle large-scale data ingestion and transformation processes for the migration to Snowflake.
  • Utilized AWS services such as S3, Lambda, and EC2 for scalable data storage and processing, optimizing resource utilization to reduce operational costs.
  • Executed complex SQL queries in Snowflake to compare, transform, and validate large datasets. Optimized query performance and managed cloud-based data warehouses.
  • Built and maintained interactive Tableau dashboards to visualize dealer data comparisons, performance metrics, and migration health, providing real-time insights to stakeholders.
  • Enhanced Python scripts to optimize ETL jobs, reduce processing times, and improve overall system efficiency.
  • Deployed monitoring solutions to track the success of scheduled Arrow jobs, proactively identifying and resolving data loading issues.
  • Led efforts in data validation, ensuring accuracy and consistency in the migration of dealer data across platforms.
  • Acted as a key liaison between business users and development teams, translating complex business requirements into scalable technical solutions.
  • Contributed to daily GitHub workflows, handling code commits, issue tracking, testing, and maintaining version control for migration scripts and ETL processes.

Technologies & Tools: Python, AWS (S3, Lambda, EC2), Snowflake, SQL, Tableau, Arrow, GitHub, ETL Pipelines, Data Warehousing, Cloud Data Migration.

2. Project: Data Migration & Report Development for Snowflake and Tableau.

Description:

Worked on Snowflake for generating Tableau reports, collaborating closely with business stakeholders to gather requirements and define functionality. Developed performance reports that offered high-level overviews with the ability to drill down into specific details. Leveraged Python and AWS cloud services to automate processes and ensure efficient data management and report generation.

Responsibilities:

  • Managed the migration of data from Salesforce into Snowflake across different stages (D0 One Lake, D1 Snowflake, D2a) to prepare datasets for Tableau report generation.
  • Automated data ingestion and transformation processes using Python scripts, AWS Lambda, and Snowflake for scalable and efficient data loading.
  • Curated and transformed datasets using Snowflake's bulk loading and exchange tools, ensuring data integrity and performance optimization.
  • Developed detailed Tableau dashboards and reports, utilizing data blending techniques and best practices for effective data visualization and insights delivery. Integrated multiple data sources in Power BI for additional reporting needs.
  • Improved query performance in Snowflake by optimizing SQL queries, removing redundant columns, and implementing efficient joins and indexing strategies.
  • Created dashboards, scorecards, views, and pivot tables for business intelligence, ensuring that reports were user-friendly and aligned with business requirements.
  • Utilized AWS services (S3, Lambda, EC2) to manage data storage and processing, streamlining data migration and reducing operational costs.
  • Worked closely with business users, project managers, and subject matter experts to gather requirements and ensure that all reports met business needs and provided actionable insights.

Environment: AWS (S3, Lambda, EC2), Python, Snowflake, Tableau, Power BI, SQL Server, MySQL, Oracle, Teradata, SSIS, Agile Methodologies, C#, JavaScript, MS Office, MS Visual Studio, SVN.

Data Engineer

Verizon
Chennai, Tamil Nadu
11.2019 - 06.2021

● Worked with team on developing and maintaining ETL (Extract, Transform, Load) pipelines using tools like Python, SQL to automate data extraction from various sources and load it into data storage systems.

● Worked on managing data in cloud-based data warehouses like Amazon Redshift, Snowflake ensuring the data is optimized for performance and accessibility.

● Lead meetings with business partners to define functional needs, which leads to the creation of system specifications and applications.

● Gained experience in cleaning and transforming raw data using tools like Pandas, PySpark ensuring consistency and quality across datasets.

● exposed to cloud services like AWS (S3, EC2, Lambda, RDS), Azure to set up scalable data pipelines and storage solutions in the cloud.

● Worked on creating basic data visualizations and reports using Tableau, Power BI.

● Gained Knowledge on version control systems like Git/GitHub to manage code, scripts, and pipeline configurations.

● Collaborate with cross-functional teams using tools like Jira, Confluence, and Slack to track tasks and projects, ensuring smooth communication and teamwork.

Environment: Windows 7, Linux, Tableau Desktop (10.x/2018.x/2019.x), Tableau Server (10.x/2018.x/2019.x), Microsoft SQL Server, Apache Hadoop, DB2, Informatica, Python, Java,Git, GitHub, JIRA, Agile,and Microsoft Excel.

Education

Bachelor of Science - Computer Science

SRM Institute of Technology
Chennai

Master of Science - Computer Science

University of New Haven
West Haven, CT

Skills

Cloud Technologies:

AWS (S3, EC2, Lambda, Redshift), Azure

Data Warehousing & Databases

Snowflake, Amazon Redshift, Oracle, SQL Server

Programming & Scripting Languages

Python, SQL , Java/J2EE, R, SAS, Unix Shell Scripting (Bash)

ETL Tools

Python (for ETL automation), Informatica, Alteryx

BI &Reporting Tools:

Tableau (Server & Desktop), Power BI, Advanced MS Excel

Operating Systems

Windows, Linux (RHEL, CentOS)

Analytical Tools

R, SAS Enterprise Guide, SAS E-Miner, Adv MS Excel

Key Skills and Expertise:

  • Expertise in building data pipelines using Python, SQL, and Big Data technologies like Hadoop, Hive. Strong understanding of ETL processes, data validation, and transformation across diverse datasets.
  • Experienced in working with AWS cloud services such as EC2, S3, RDS, VPC, CloudWatch, Route 53, EBS, AMI, SNS, and more, for building and managing scalable data solutions in cloud environments.
  • Proficient in Tableau, Power BI, and SQL Server Reporting Services (SSRS) to create on-demand and scheduled reports for business analysis, ensuring data-driven decision-making.
  • Strong knowledge of databases such as Oracle, SQL Server, and MySQL, including experience in PL/SQL for building efficient stored procedures, functions, packages, triggers, and optimizing queries using techniques like Bulk Binds, Indexing, Joins, and Set Operators.
  • Advanced skills in Python and R for data manipulation, statistical analysis, and automation. Expertise in using Python libraries such as Pandas, NumPy, SciPy and matplotlib for data processing and visualization.
  • Adept in data modelling techniques to design optimized database architectures. Skilled in SQL and PL/SQL to handle complex data transformations, error handling, and data extraction across multiple sources.
  • Knowledgeable in Hadoop and Hive for managing large-scale datasets and conducting data analysis in distributed computing environments.
  • Expert in creating visually compelling and interactive dashboards in Tableau and Power BI that provide actionable insights and facilitate decision-making for stakeholders.
  • Proven track record of working closely with clients to understand their data requirements, translating business needs into technical solutions, and delivering meaningful reports and dashboards.
  • Thorough understanding of Software Development Life Cycle (SDLC) methodologies such as Agile and Waterfall, ensuring efficient project execution from requirement gathering to deployment.
  • Gathered Semi-structured data from S3 and relational structured data from RDS and kept data sets into centralized metadata Catalog using AWS GLUE. Extracted the datasets and loaded them into Kinesis streams.

Certification

  • HackerRank SQL certification
  • Hackerank Python Certification

Timeline

Data Engineer

Capital One
07.2022 - Current

Data Engineer

Verizon
11.2019 - 06.2021

Bachelor of Science - Computer Science

SRM Institute of Technology

Master of Science - Computer Science

University of New Haven
Ponakala Sai Nandan