Summary
Overview
Work History
Education
Skills
Certification
Timeline
Generic

Yasaswini Alluru

United States

Summary

Data Engineer with over 5 years of experience specializing in JAVA, PYTHON, and Machine Learning. Expertise in Big Data technologies and the Hadoop Ecosystem, alongside strong skills in Cloud Engineering and Data Warehousing. Proficient in AWS services such as EC2, S3, and EMR, with a solid track record of developing Spark applications using Spark SQL and Spark Streaming. Advanced knowledge of Python libraries including Pandas and NumPy for data manipulation and analysis.

Overview

5
5
years of professional experience
1
1
Certification

Work History

Data Engineer

Entergy
01.2023 - Current
  • Executed full lifecycle (SDLC) of data warehouse projects, focusing on dimensional data modeling.
  • Constructed data warehouse structures by developing facts, dimensions, and aggregate tables using Star and Snowflake schemas.
  • Designed and deployed scalable, fault-tolerant systems on AWS to enhance operational reliability.
  • Leveraged various AWS services including S3, EC2, EKS, Glue, Athena, RedShift, EMR, and SNS for diverse project needs.
  • Implemented scalable, cost-effective data pipelines with AWS Glue and S3 to ensure efficient data processing.
  • Extracted data from multiple source systems into Glue Catalog using Glue Crawlers to create organized datasets.
  • Automated daily production tasks through shell scripts and Python scripts to increase efficiency.
  • Developed ETL pipelines for data warehouse integration using Python and Snowflake SQL queries.

Data Engineer

Wells Fargo
India
06.2021 - 07.2022
  • Designed and implemented large-scale data intelligence solutions on big data platforms.
  • Developed Spark applications utilizing JAVA and Python for enhanced processing capabilities.
  • Created scalable data pipelines using AWS Glue, S3, Lambda, Redshift, EMR, and Athena.
  • Engineered ETL processes in AWS Glue to migrate data from external sources into Redshift.
  • Collaborated with marketing teams to define data requirements and optimize Martech tool utilization.
  • Crafted multiple recipes in Glue Data Brew for use in various ETL jobs.
  • Analyzed business requirements with users and developers to produce comprehensive BRD and FSD documents.
  • Translated business needs into functional specifications using workflow and sequence diagrams.

Data Engineer Intern

I-WEAVE SOLUTIONS PRIVATE LIMITED
India
02.2020 - 05.2021
  • Earned placement in competitive technical training program, demonstrating strong aptitude for software development.
  • Achieved proficiency in Java, Python, PySpark, Tableau, Hive QL, Hadoop, SQL Server, SDLC, ETL, and text-mining processes.
  • Developed and maintained transformation logic using Informatica mappings and workflows.
  • Extracted data from diverse sources including databases, flat files, and APIs.
  • Automated data pipeline jobs on daily and monthly schedules using Autosys.
  • Created dashboards and visualizations with Tableau using calculated fields and custom hierarchies.
  • Imported and exported data between MySQL and Hadoop clusters utilizing Sqoop.
  • Contributed to full life cycle implementation of ETL processes while designing data warehouse structures.

Education

Master of Science - Information Studies

Trine University
Detroit, MI
12-2023

Bachelor of Technology - Computer Science And Engineering

JNTU Anantapur
India
07-2021

Skills

Python

Apache Spark

Bash / Python / PowerShell

SQL

pyspark

Scala

Certification

  • AWS Certified Developer - Associate

Timeline

Data Engineer

Entergy
01.2023 - Current

Data Engineer

Wells Fargo
06.2021 - 07.2022

Data Engineer Intern

I-WEAVE SOLUTIONS PRIVATE LIMITED
02.2020 - 05.2021

Master of Science - Information Studies

Trine University

Bachelor of Technology - Computer Science And Engineering

JNTU Anantapur
Yasaswini Alluru