Summary
Overview
Work History
Education
Skills
Certification
Timeline
Generic

Sai Ram Reddy Sudini

Summary

Enthusiastic and dedicated professional with robust experience in software development and Designing Data Engineering applications. Expertise in leveraging cloud technologies across both Azure and AWS platforms to streamline and scale data operations. Proficient in driving DevOps best practices through CI/CD pipeline construction, automation, and orchestration using tools such as ARM templates, BICEP, Azure Devops. Committed to optimizing deployment processes and enhancing infrastructure as code initiatives, seeking to bring strong analytical and problem-solving skills to an Azure DevOps role. Expert in Building Robust Pipelines Using Azure Data Factory, Azure Data Bricks, Logic Apps, Azure Synapse and Azure Fabrics.

Overview

9
9
years of professional experience
1
1
Certification

Work History

Sr. Data Engineer / DataOps Developer

Sumitomo Mitsui Banking Corporation
New Jersey
07.2022 - Current
  • Architected and managed robust CI/CD pipelines using Azure DevOps, facilitating seamless deployment cycles and continuous integration across both banking and healthcare sectors to enhance code deployment strategies and system responsiveness
  • Led the design and execution of secure data solutions within Azure, leveraging Azure Data Factory and Databricks to develop sophisticated ETL pipelines and scalable data models
  • Instituted IaC practices using ARM templates and BICEP, automating cloud infrastructure provisioning to ensure consistency and repeatability across diverse environments
  • Spearheaded the migration and integration of data storage solutions to Azure, optimizing for high availability, disaster recovery, and regulatory compliance
  • Crafted meticulous PowerShell and Bash scripts to automate routine data operations, significantly enhancing operational efficiency and the organization’s ability to respond rapidly to changing market conditions
  • Developed intuitive BI dashboards and reports, providing real-time financial and health insights to business analysts and executives
  • Conducted comprehensive monitoring, tuning, and troubleshooting of Azure Databricks clusters
  • Collaborated closely with business stakeholders to gather requirements and craft data visualization strategies.

Sr. Data Engineer / DataOps Developer

Kaiser Permanente
Pleasanton, California
06.2020 - 07.2022
  • Architected and managed continuous integration and delivery (CI/CD) pipelines using Azure DevOps, enhancing code deployment strategies and enabling rapid, reliable releases to Azure cloud environments
  • Automated end-to-end data operations with Azure Data Factory, crafting sophisticated ETL pipelines that integrate seamlessly with Azure SQL, Blob Storage, and Azure Synapse Analytics, thereby improving data fidelity and system responsiveness
  • Implemented Infrastructure as Code (IaC) using ARM templates and BICEP to provision and manage Azure resources, ensuring consistent, repeatable deployments across development, testing, and production environments
  • Utilized Azure Databricks for developing scalable data processing applications in PySpark and Spark-SQL, extracting and transforming diverse datasets into actionable insights, and optimizing resource allocation for peak performance
  • Conducted thorough monitoring, tuning, and troubleshooting of Azure Databricks clusters, leveraging Spark’s in-memory capabilities for efficient batch and stream processing
  • Optimized Spark applications by tuning parameters such as batch intervals, levels of parallelism, and memory management, resulting in significant performance improvements and cost savings
  • Developed JSON scripts to automate the deployment of data processing workflows in Azure Data Factory, orchestrating SQL activities for improved data management and analytics
  • Gained proficiency in writing complex SQL scripts for automating data movement and transformation tasks, thus reducing manual overhead and increasing operational efficiency
  • Partnered with business stakeholders to elicit requirements and provide expert guidance on data visualization techniques, leading successful training initiatives on self-service BI tools to empower end-users
  • Engineered resilient distributed systems for handling vast data volumes, establishing robust data pipelines, and creating a comprehensive analytics platform to track and analyze key performance metrics.

Data Engineer

Visa
Austin, Texas
05.2019 - 06.2020
  • Experience in working with Azure Cloud platform (Databricks, Data Lake, Blob, Data Factory, Azure functions)
  • Building the pipelines to copy the data from source to destination in Azure Data Factory
  • Worked on creating dependencies of activities in Azure Data factory
  • Transformed data in Azure Data Factory with the ADF Transformations
  • Worked with Azure Databricks (Spark) for data preparation using Pyspark and load the data into Azure Blob Storage
  • Extensively used Databricks notebooks for interactive analysis using Spark API’s
  • Developed Spark applications using Pyspark and Spark-SQL for data extraction, transformation and aggregation from multiple file formats for analyzing & transforming the data to uncover insights into the customer usage patterns
  • Developed SQL Server Stored Procedures, UDF’s and created Views based on the business requirements
  • Developed Azure Functions that support serverless computing to implement the business logic using C# language and debugged the Azure Functions using Insomnia locally
  • Implemented Azure DevOps Services for project management tasks, source code repository, continuous integration and deployment.

Data Engineer

Amex
Phoenix, Arizona
09.2018 - 05.2019
  • Worked with extensive data sets in Big Data to uncover pattern, problem & unleash value for the Enterprise
  • Worked with internal and external data sources on improving data accuracy / coverage and generate recommendation on the process flow to accomplish the goal
  • Ingestion of various types of data feeds from SOR and use-case perspective into Cornerstone 3.0 platform
  • Re-engineered legacy IDN FastTrack process to get the Bloomberg data directly from source to the CS3.0
  • Converted legacy Shell scripts to Map-Reduce jobs in a distributed manner without performing any kind of processing on the Edge node to eliminate the burden
  • Created Spark applications for data preprocessing for greater performance
  • Developed Spark code and Spark-SQL/streaming for faster testing and processing of data
  • Experience in creating spark applications using RDD, Data frames
  • Worked extensively on hive to analyses the data and create reports for data quality
  • Implemented Partitioning, Dynamic Partitions and Buckets in HIVE for increasing performance benefit and helping in organizing data in a logical fashion
  • Written Hive queries for data analysis to meet the business requirements and Designed and developed User Defined Function (UDF) for Hive
  • Involved in creating Hive tables (Managed tables and External tables), loading and analyzing data using hive queries
  • Good knowledge about the configuration management tools like SVN/CVS/GitHub
  • Experience in configuring Event Engine nodes to import and export the data from Teradata to HDFS and vice-versa
  • Worked with source to get the history data as well as BAU data from IDN Teradata to the Cornerstone platform and migrated also feeds from CS2.0
  • Expert in creating the nodes in Event Engine as per the use-case requirement to automate the process for the BAU data flow
  • Exported the Event Engine nodes created in the silver environment to the IDN repository in BitBucket and created DaVinci package to migrate it to Platinum
  • Worked with FDP team to create a secured flow to get the data from KAFKA Queue to CS3.0
  • Expert in creating the SFTP Connection to the internal and external source to get data in secured manner without any breakage
  • Handle the production Incidents assigned to our workgroup promptly and fix the bugs or route it to the respective teams and optimized the SLA’s.

Hadoop Developer

Anthem
Norfolk, Virginia
01.2017 - 09.2018
  • Designed and developed scalable and cost-effective architecture in AWS Big Data services for data life cycle of collection, ingestion, storage, processing, and visualization
  • Involved in creating End-to-End data pipeline within distributed environment using the Big data tools, Spark framework and Tableau for data visualization
  • Ensure that application continues to function normally through software maintenance and testing in production environment
  • Leverage Spark features such as In-Memory processing, Distributed Cache, Broadcast, Accumulators, Map side Joins to implement data preprocessing pipelines with minimal latency
  • Implemented real-time solutions for Money Movement and transactional data using Kafka, Spark Streaming, HBase
  • The project also includes a spread of big data tools and programming languages like Sqoop, Python, Oozie etc
  • Worked on scheduling Oozie workflow engine to run multiple jobs
  • Experience in creating python topology script to generate cloud formation template for creating the EMR cluster in AWS
  • Good knowledge on AWS Services like EC2, EMR, S3, Service Catalog, and Cloud Watch
  • Experience in using Spark-SQL to handle structured data from Hive in AWSEMR Platform (M4.Xlarge,M5.12Xlarge clusters)
  • Exploring with Spark, improving performance and optimization of the existing algorithms in Hadoop using Spark Context, Spark-SQL, Data Frame, and Pair RDD's
  • Experienced in handling large datasets using Partitions, Spark in Memory capabilities, Broadcasts in Spark, Effective & efficient Joins, Transformations and other during ingestion process itself
  • Experienced in optimizing Hive queries, joins to handle different data sets
  • Involved in creating Hive tables (Managed tables and External tables), loading and analyzing data using hive queries
  • Actively involved in code review and bug fixing for improving the performance
  • Good experience in handling data manipulation using python Scripts
  • Involved in development, building, testing, and deploy to Hadoop cluster in distributed mode
  • Created Splunk dashboard to capture the logs for end to end process of data ingestion
  • Written unit test cases for Pyspark code for CICD process
  • Good knowledge about the configuration management tools like GitHub and Bamboo(CICD).

Software Developer

Vitech System
Hyderabad, Telangana
01.2015 - 08.2016
  • Review the system requirements and attending requirements meetings with analysts and users
  • Involved in the life cycle of the project from documentation to unit testing making development as priority
  • Developed web pages using Struts framework, JSP, XML, JavaScript, HTML/ DHTML and CSS, configure struts application, use tag library
  • Used Apache Struts framework includes the integrated AJAX
  • Played major role in designing & developing JSP pages and XML reports
  • Developed Servlets and custom tags for JSP pages
  • Developed few module Web pages using Springs IOC and Hibernate
  • Designed and developed dynamic pages using HTML, CSS- layout techniques, Java script
  • Took the various challenges in the enhancement and completed them on time
  • Extensive Used Exception handling and Multi-threading for the optimum performance of the application
  • Involved in design and implemented (SOA, SOAP) next generation system on distributed platform
  • Extensively used XSL as a XML parsing mechanism for showing Dynamic Web Pages in HTML format
  • Implemented SOAP protocol to get the requests from the outside System
  • Used CVS as a source control for code changes
  • Used ANT scripts to build the project and JUnit to develop unit test cases
  • Developed coding using SQL, PL/SQL, Queries, Joins, Views, Procedures/Functions, Triggers and Packages
  • Provided development support for System Testing, Product Testing, User Acceptance Testing, Data Conversion Testing, Load Testing, and Production.

Education

Master of Science - Information Technology

Southern Arkansas University
Magnolia, AR

Skills

  • Azure Data Factory, Azure Databricks
  • Azure Functions
  • Azure logic Apps
  • Azure Devops , CICD
  • Python
  • API Development
  • Continuous integration
  • Data Modeling
  • Data Security
  • Python Programming
  • Performance Tuning
  • Data Migration
  • Database Design
  • Data Analysis
  • SQL and Databases
  • Big data technologies
  • Problem-Solving

Certification

  • Azure Data Engineer Associate

Timeline

Sr. Data Engineer / DataOps Developer

Sumitomo Mitsui Banking Corporation
07.2022 - Current

Sr. Data Engineer / DataOps Developer

Kaiser Permanente
06.2020 - 07.2022

Data Engineer

Visa
05.2019 - 06.2020

Data Engineer

Amex
09.2018 - 05.2019

Hadoop Developer

Anthem
01.2017 - 09.2018

Software Developer

Vitech System
01.2015 - 08.2016

Master of Science - Information Technology

Southern Arkansas University
Sai Ram Reddy Sudini