Overview
Work History
Education
Skills
Certification
Projects
Timeline
Generic

Reshank Vasala

Scottsdale,AZ

Overview

6
6
years of professional experience
1
1
Certification

Work History

Data Engineer

Delton Solutions
11.2024 - Current
  • Designed and implemented ETL pipelines to process, transform, and load large datasets using tools like Apache Airflow, and PySpark, ensuring efficient data integration across systems.
  • Developed and maintained data warehouses and data lakes using on-premise technologies like Apache Hive and PostgreSQL, ensuring data accessibility, consistency, and compliance.
  • Built and optimized complex SQL queries and relational database models, improving query performance by up to 40% for analytics dashboards.
  • Integrated streaming data pipelines using tools like Apache Kafka and Apache Flink, enabling real-time analytics and event-driven workflows.
  • Deployed scalable data pipelines using Dockerized environments and managed orchestration through Kubernetes clusters.
  • Conducted extensive data profiling and quality checks, implementing robust monitoring frameworks and error-handling mechanisms to reduce data inconsistencies by 30%.
  • Collaborated with data scientists and analysts to create efficient pipelines that supported predictive models and advanced analytics workflows.
  • Established and maintained data governance standards, ensuring compliance with data security regulations and improving metadata management across systems.

Data Engineer

Athene Contract, Broad Peak INC
05.2023 - 09.2024
  • Developed and optimized ETL pipelines using Azure Data Factory, reducing data processing times by 30% by leveraging data flow transformations, custom logging, and monitoring, ensuring high-quality data for business analysis.
  • Managed and scaled a data warehouse on Azure Synapse Analytics, improving real-time query performance for business stakeholders by implementing best practices for data partitioning, indexing, and resource management.
  • Automated data workflows with Python and Bash scripts, streamlining data ingestion from multiple sources, performing transformations, and generating reports, enabling the team to deliver insights more quickly and accurately.
  • Integrated Azure Data Lake with Data bricks for advanced big data processing, enabling the company to handle and analyze large volumes of unstructured data from multiple sources with greater accessibility and flexibility.
  • Implemented cost-saving strategies, reducing cloud resource usage by 25% by optimizing resource scaling, rightsizing compute instances, and automating the scheduling of resource-intensive jobs, all while maintaining high system performance and reliability.

Junior Data Engineer

DePaul University
07.2021 - 11.2022
  • Built ETL pipelines using Azure Data Factory for seamless data ingestion from APIs and SQL databases into Azure Blob Storage, ensuring high-quality, real-time data availability for analytics teams.
  • Improved SQL query performance by optimizing complex queries and indexing strategies, and implemented efficient schema designs to reduce query times by 40%, enhancing data retrieval speeds and user experience in cloud-based data warehouse operations.
  • Automated data quality checks with Python scripts, which significantly improved data accuracy by identifying anomalies and inconsistencies, resulting in a 20% reduction in data errors and improved reporting integrity.
  • Collaborated with the cloud infrastructure team to provision Azure resources using ARM templates and Azure DevOps, which reduced the environment setup and deployment times by 50%, ensuring rapid scalability and continuous integration for cloud applications.

Data Engineering Intern

Contract: Airtel
01.2019 - 07.2020
  • Developed Python scripts to automate data ingestion from various file formats (CSV, JSON, XML) into Azure Blob Storage, streamlining the data pipeline process and improving overall efficiency.
  • Assisted in designing initial ETL workflows using Azure Data Factory, optimizing the cloud integration process to automate data movement across multiple systems and sources.
  • Conducted exploratory data analysis (EDA) on transactional data to generate actionable insights, identifying patterns in customer behavior that informed marketing strategies and improved targeting.
  • Collaborated with data scientists and business analysts to define requirements and structure datasets for predictive analytics models, increasing the quality and relevance of data used.
  • Integrated Azure SQL Database with Azure Blob Storage to store and retrieve structured and unstructured data, enabling seamless querying and data access.
  • Assisted in developing Python-based data cleaning and transformation scripts, ensuring high data quality for reporting and analysis purposes.
  • Automated data validation checks using SQL and Python, reducing errors in data processing and ensuring consistency across data pipelines.
  • Tools/Environment: Python, Azure Blob Storage, Azure Data Factory, SQL, Azure SQL Database, Jupiter Notebooks, Power BI
  • Participated in team projects, demonstrating an ability to work collaboratively and effectively.

Education

Master of Science - Information Systems

Depaul University
Chicago, IL
11-2022

Skills

  • ● Cloud Platforms: Microsoft Azure, Azure Data Factory, Azure Synapse Analytics, Azure SQL Database, Azure Blob Storage, Azure Data Lake
  • Data Engineering: ETL Pipelines, Data Transformation, Data Modeling, Data Warehousing, Apache Spark
  • Programming & Scripting: Python (Pandas, NumPy), Shell Scripting (Bash, PowerShell)
  • DevOps & Automation: Azure DevOps, Git, ARM Templates, Terraform
  • Data Visualization: Power BI, Soft Skills: Analytical Thinking, Problem-Solving, Collaboration, Stakeholder Communication
  • Data pipeline control

Certification

  • Microsoft Certified: Azure Data Engineer Associate (DP-203)
  • Microsoft Certified: Azure Fundamentals (AZ-900)
  • Microsoft Certified: Azure Administrator Associate (AZ-104)

Projects

Real-Time Data Processing with Azure and Databricks:

  • Built a real-time data pipeline using Azure Event Hubs to capture and process streaming data., Integrated Databricks with Azure Synapse Analytics to perform real-time transformations and store data for analytics.
  • Utilized PySpark to process and aggregate over 1,000 events per second with high fault tolerance.
  • Designed visual dashboards on Power BI to present real-time analytics and business insights., and Azure Event Hubs, Databricks, PySpark, Azure Synapse Analytics, Power BI


Scalable ETL Pipeline for E-Commerce Analytics: 

  • Developed an ETL pipeline using Azure Data Factory to extract, transform, and load data from multiple sources into Azure Data Lake.
  • Implemented data transformation logic in Python, improving data accuracy and reducing processing time by 40%.
  • Deployed the pipeline in a production environment, processing 5+ GB of e-commerce data daily., Automated the monitoring and error-handling processes to ensure consistent data flow and integrity., Azure Data Factory, Python, Azure Data Lake

Timeline

Data Engineer

Delton Solutions
11.2024 - Current

Data Engineer

Athene Contract, Broad Peak INC
05.2023 - 09.2024

Junior Data Engineer

DePaul University
07.2021 - 11.2022

Data Engineering Intern

Contract: Airtel
01.2019 - 07.2020

Master of Science - Information Systems

Depaul University
Reshank Vasala