Summary
Overview
Work History
Education
Skills
Awards/Accomplishments
Certification
Timeline
Generic
RAM Vuruvakili

RAM Vuruvakili

Piscataway,NJ

Summary

Versatile Data Engineer with extensive experience in Banking, Credit, Healthcare, and Capital Markets. Proficient in utilizing Hadoop and Spark to optimize data operations, resulting in substantial improvements in data validation and developer productivity. Expertise in AWS Services combined with strong collaboration and problem-solving skills ensures a commitment to operational excellence and quality outcomes. Holds certification as Databricks Certified Data Engineer Associate

Overview

9
9
years of professional experience
1
1
Certification

Work History

Data Engineer

Barclays
Whippany, NJ, USA
05.2023 - Current

Treasury Data Source streamlines the data from various upstreams and leverages Nifi and Spark to further process the data according to the requirements specified by multiple regulatory bodies across the USA and UK.

  • Delivered user stories focused on regulatory business reports essential to the LTP through collaboration with Product Owners, Data Architects, and Business Analysts.
  • Boosted developer productivity by developing utility for automating Spark job or the pipeline executions to ease the development and testing efforts.
  • Developed an utility app that can help the developers and business owners to validate the data that can further ease approval process.
  • Ensured seamless continuity of operations by managing data quality promptly by collaborating with Run The Bank team.
  • Demonstrated proficiency in handling diverse file formats like Parquet, Avro, ctrl, avsc, and hql to enhance data integration.
  • Executed optimizations to boost the performance and efficiency of Spark applications.
  • Optimized TDS codebase leveraging DevOps tools like GIT, BitBucket, TeamCity, and SonarQube.

AWS & Big Data Developer

TAVANT Technologies
Hyderabad, Telangana, India
03.2020 - 04.2023

Experian Ascend provides its clients, a businesslike platform to run their campaigns on US population, that allows them effectively provide Credit Line Increase, Credit Line Decrease & Balance Transfer features to their customer.

  • Got good experience in AWS Services and worked on AWS Lambda, S3, RDS, EMR, Cloudwatch, SNS, SES & Step Functions
  • Mostly worked on the data pipeline automation using AWS using Lambda, Step Functions
  • Efficiently storing the data in S3 with minimal cost using S3 Lifecycles
  • Worked on sampling, validating the data using Spark/Scala
  • Writing Scala Test classes for Unit Testing and following TDD development approach
  • Worked on writing the AWS Cloud formation templates to manage the AWS resources and their dependencies
  • Providing the users an UI to schedule/trigger the production jobs using AWS Service Catalog and validating their input
  • Worked on Cloudwatch rules updating considering Day Light Saving and enabling alarms on AWS Resources
  • Allowing multiple runs on single day so that the client can test their strategies before implementing them in production
  • Validating the dependent jobs for the production run and not allowing the run in case of failures/issues
  • Worked on archiving the input & business artifacts to S3 using AWS Lambda
  • Worked as release manager, which involves deploying code changes to production
  • Working with the team on issues and fixing them

Big Data Developer

GENPACT
Hyderabad, Telangana, India
09.2016 - 03.2020

GE Healthcare Machine Data Platform mainly focuses on processing the XML, JSON & Text data which comes from machines in the form of logs. The machines can be any of the wide varieties of XRAY, CT, MR, Ultrasound and so on.

This data is processed based on the user requirements and finally loaded into Hive, Postgres & Greenplum databases. This loaded data is used by the data analytics team for further processing or analytics, which helps in better understanding of their machines and point of business expansion.

  • Got good experience in Big Data Development and worked on Hadoop, Hive, Spark/Java, Shell Scripting, Sqoop, Postgres & Greenplum development in the last 3 years
  • Analyzing the project requirements and developing the code as per requirements
  • Worked majorly on parsing the XML, JSON and Text data
  • Developing JAVA UDFs to process this data using Spark/Hive Processing
  • Developing Hive Scripts to load the data in Hive and HDFS
  • Developing Green Plum YAML files to load data into Green Plum
  • Developing Shell Scripts for process & load automation
  • Worked on exporting/importing data from/to HDFS/HIVE/RDBMS using sqoop
  • Job scheduling using Oozie/Shell Scripting
  • Got good knowledge and experience on Spark execution & Development
  • Working on Agile and Waterfall models
  • Did a POC on Spark File Streaming
  • Involved in Hortonworks to Amazon AWS cluster migration
  • Developing the unit test cases for every development activity and documentation
  • Continuously working with the team on Production issues and fixing them
  • Deployed code on Hadoop Edge Nodes and made test runs

Hyperion Admin

GENPACT
Hyderabad, Telangana
06.2016 - 09.2016

GE Capital is one of the biggest leasing businesses of the Europe. Capital uses Hyperion application for Budgeting, forecasting and closing reporting purposes.

  • Worked as Hyperion Administrator, where I mostly worked on Security (Hyperion Shared Services) and Hierarchy Maintenance
  • Involved in designing security filters and assigning them to end-users
  • Involved in optimization of outline, Data load and calc scripts
  • Involved in analyzing the financial data and tying it with oracle GL

Education

Bachelor of Technology - Computer Science& Engineering

JNTU ANANTAPUR
Andhra Pradesh, India
05.2016

Skills

  • Hadoop
  • Spark
  • Scala
  • Shell scripting
  • Hive/Impala
  • Python
  • Databricks
  • AWS Services
  • Postgres/Greenplum
  • Jenkins
  • Maven/SBT
  • MongoDB/DynamoDB

Awards/Accomplishments

  • Got recognized multiple times in Tavant's quarterly Appreciation Awards for my work in the project.
  • Wonderful appreciations from my team, for my work in the current project during Tavant Pulse.
  • Runner-ups during Experian hackathon.
  • Got bronze award for overall performance in Genpact during the year 2019.

Certification

  • Databricks Certified Data Engineer Associate

Timeline

Data Engineer

Barclays
05.2023 - Current

AWS & Big Data Developer

TAVANT Technologies
03.2020 - 04.2023

Big Data Developer

GENPACT
09.2016 - 03.2020

Hyperion Admin

GENPACT
06.2016 - 09.2016

Bachelor of Technology - Computer Science& Engineering

JNTU ANANTAPUR
RAM Vuruvakili