Summary
Overview
Work History
Education
Skills
Timeline
Generic

Sri Rasagna

Parsippany,NJ

Summary

Certified Data Engineer with AWS and Databricks certifications, and 4+ years of experience in data engineering, cloud technologies, and big data. Proficient in building scalable data pipelines, optimizing workflows, and leveraging AWS, Databricks, and Spark to deliver impactful business solutions. Detail-oriented and collaborative problem-solver with a strong focus on efficiency and innovation. Ready to tackle challenges and deliver impactful solutions for a high-caliber technology company.

Overview

3
3
years of professional experience

Work History

DB Engineer

SageIT INC
Barclays, NJ
05.2024 - Current
  • Worked on preparation and submission of annual CCAR (Comprehensive Capital Analysis and Review) reports, ensuring timely and accurate data sourcing, processing, and model building for financial projections
  • Managed and optimized data flows from ROCP (Risk and Operations Control Platform) to Oracle SQL databases, handling large-scale data extraction, transformation, and loading (ETL) processes for downstream analytical models
  • Collaborated with cross-functional teams to build and enhance models that generate business-wide projections for CCAR reports, using robust and scalable data pipelines
  • Developed and maintained automated Python scripts to process and move wholesale credit data and trade wind-down information, facilitating efficient and accurate data movement from SQL tables to Athena for business analytics
  • Utilized TeamCity for managing job deployments and continuous integration pipelines, ensuring smooth and reliable code and process updates across the data environment
  • Ensured data integrity and compliance by implementing rigorous validation checks and documentation for all data-related tasks, meeting regulatory requirements for financial reporting

Data Engineer

Capital One
09.2023 - 04.2024
  • Company Overview: Remote
  • Developed a real-time message processing pipeline to pull transactional data from AWS SQS and S3 buckets
  • Implemented Lambda functions to filter legitimate debit transactions, parsing incoming transactional messages for valid entries
  • Developed custom logic to identify and create unique transaction IDs, improving data consistency and accuracy by 25%
  • Integrated with partner offers from Macy's and CVS by assigning relevant offers to eligible debit transactions based on customer profiles and transaction details.
  • This led to a 15% increase in offer engagement and customer redemption rates
  • Designed and deployed SQS triggers to automatically process messages from S3 buckets as soon as new data was available, allowing for seamless and scalable data ingestion.
  • This event-driven architecture improved processing efficiency and reduced message backlog by 40%.
  • Automated the flow of processed transaction data to a central payout table, ensuring accurate and timely updates of transaction details for downstream

Data Engineer

PetSmart
03.2022 - 08.2023
  • Company Overview: Remote
  • Engineered end-to-end ETL pipelines, sourcing data from MEF Integration and TIBCO systems, and automating data ingestion from AWS S3 buckets in both JSON and CSV formats
  • Streamlined the process, reducing manual intervention by 30% and improving the overall efficiency of data intake
  • Utilized Py Spark to perform complex data transformations and aggregations, processing everyday retail transaction records daily
  • Optimized data partitioning strategies and applied caching techniques, leading to a 40% improvement in processing times for large datasets.
  • Managed the transformation of raw data into Delta Lake tables within Databricks, ensuring high-performance data processing, and compliance with ACID principles.
  • Implemented incremental loading strategies that increased data throughput by 25%, enabling near real-time data access for downstream systems.
  • Developed, automated, and managed Databricks workflows, ensuring seamless job execution and efficient scheduling.
  • Led the migration of critical data from Netezza to Delta Lake within Databricks, transferring over 500 GB of data with minimal downtime.
  • The migration resulted in a 35% improvement in query performance and a 50% reduction in storage costs due to Delta Lake’s optimized storage format
  • Designed and implemented Snow pipes for automated ingestion and transformation of data into Snowflake, creating a consumption layer for business intelligence.
  • This integration facilitated the creation of materialized views and improved data refresh cycles by 40%, empowering real-time decision-making via MicroStrategy cubes.

Education

Masters - Data Science

Western Michigan University
12-2021

Bachelors - Computer Science Engineering

JNTUH
04-2018

Skills

  • Databricks
  • Data Pipelines
  • AWS
  • GCP
  • Python
  • SAS
  • Spark
  • Oracle DB
  • Neptune DB
  • MySQL
  • PostgreSQL
  • Git
  • Bitbucket
  • Shell scripting
  • PowerShell
  • Kafka
  • AutoLoader

Timeline

DB Engineer

SageIT INC
05.2024 - Current

Data Engineer

Capital One
09.2023 - 04.2024

Data Engineer

PetSmart
03.2022 - 08.2023

Masters - Data Science

Western Michigan University

Bachelors - Computer Science Engineering

JNTUH
Sri Rasagna