Summary
Overview
Work History
Education
Skills
Certification
Key Skills And Achievements
Linkdin
github
Timeline
Generic

Uday Routhu

Summary

Results-oriented IT professional with a decade of comprehensive experience in the data processing domain. Proven expertise in various roles, including key developer and solution designer. Adept at analyzing data using technologies such as HiveQL, Pig Latin, HBase, and custom Map Reduce programs in Java. Key Skills and Achievements: Proficient in loading and analyzing large datasets using Hadoop frameworks, including Map Reduce, HDFS, Pig, Hive, Sqoop, and Spark. Strong engineering background complemented by close collaboration with business customers, ensuring a deep understanding of business needs, objectives, and expectations in data projects. Extensive work on database applications utilizing DB2 UDB, Oracle, SQLPlus, PL/SQL, Development and implementation of ETL processes. Hands-on experience in application development using Java, RDBMS, and Linux shell scripting. Senior Java Developer with expertise in web/intranet and client/server technologies, employing Java, J2EE, Servlets, JSP, JSF, EJB, JDBC, and SQL. Familiarity with Snowflake utilities (SnowSQL) and proficiency in implementing Big Data models using Python. In-depth knowledge of the software development life cycle (SDLC) and database design. Expertise in technical documentation preparation, including detailed data workflow descriptions, solution design and architecture, and technical requirements specifications. Experienced in Agile engineering practices, with a focus on implementing complete CI/CD processes, from design to development. Proven ability to prioritize tasks effectively and provide accurate estimations. Conducts thorough code reviews and maintains direct communication with customers.

Overview

10
10
years of professional experience
1
1
Certification

Work History

Quantitative Analytics Senior

Freddie Mac
McLean
11.2021 - Current
  • Transformed intricate techniques and functional requirements into detailed design specifications
  • Engineered a robust framework for extracting data from RDMS to BDAP, employing diverse technology stacks
  • Communicated key development decisions and shared their implications with the PRD team
  • Rebuild of legacy on-premise Oracle-based data warehouse to a data lake based on AWS Cloud
  • Defines and implements Data Governance Management tools
  • Assesses data repositories for compliance with data governance policies and standards
  • Works with all areas of the business to ensure data quality and integrity (DQ reports)
  • Develop and maintain ETL processes to move data from source system (Oracle) to Snowflake
  • Develop RDS process for address Scrubbing
  • Development and implementation of ETL pipelines according to the PRD design and architecture (BDAP/Hadoop, AWS)
  • Developed and implemented data quality metrics and automated data quality checks, resulting in a 50% reduction in data errors and improved data accuracy
  • Prepared detailed technical documentation, encompassing data workflow descriptions, solution design, architecture, and technical requirements specifications
  • Conducted data profiling to assess and enhance data quality
  • Collaborated with professional services teams, system architects, and product managers to deliver high-quality technical solutions meeting business requirements and regulatory standards
  • Develop and implement procedures covering data management, model development, testing, technical implementation, model execution and monitoring
  • Participated in the preparation and execution of Proof of Concepts (POCs) and demos.

Software Engineer

Compunnel Software Group, Inc
Plainsboro, NJ
07.2018 - 11.2021
  • Developed and implemented ETL processes to improve data quality, resulting in a 20% increase in data accuracy
  • Held responsibilities for economic and financial modeling, showcasing a multifaceted skill set
  • Conducted business analysis to support the deployment of ETL processes, ensuring alignment with regulatory requirements
  • Designed, architected, and built a robust data pipeline capable of ingesting Terabytes of healthcare data from various EMR and EHR systems to the point of care
  • Utilized Map Reduce and Spark frameworks to process and transform large volumes of data into the Hadoop File System
  • Demonstrated a deep understanding of data analytics, emerging technologies, and modern econometric/statistics estimation techniques
  • Worked extensively with large datasets, showcasing proficiency in Matlab, Tableau, and C++ programming
  • Implemented Solr Search/Indexing using data sourced from HBase, enhancing data accessibility
  • Cleansed, manipulated, and analyzed large datasets, encompassing both structured and unstructured data formats (XMLs, JSONs, PDFs) on the Hadoop platform
  • Developed data profiling, deduping logic, matching logic for analysis
  • Processed large volumes and varieties of data, including both structured and unstructured formats (XMLs, JSONs, PDFs), with a focus on parallel processing
  • Managed multiple priorities effectively, consistently meeting deadlines.

Software Engineer

nDimensions
Livonia, MI
03.2017 - 06.2018
  • Developed and implemented ETL processes, yielding a significant 20% increase in data accuracy by improving data quality
  • Collaborated with data scientists to devise and implement data pipelines, resulting in an impressive 40% increase in data availability
  • Conducted thorough data analysis to identify patterns and trends in customer behavior, providing valuable insights for decision-making
  • Translated intricate techniques and functional requirements into detailed designs, showcasing a strong analytical and problem-solving skill set
  • Designed a framework for converting existing mappings to PySpark (Python and Spark) Jobs, enhancing efficiency and scalability
  • Analyzed current programs, focusing on performance, diagnosis, and troubleshooting of problematic programs
  • Designed effective solutions for problematic programming using Spark Data Frame
  • Worked on transferring data between Relational databases and HDFS, and vice versa, leveraging the Hadoop ecosystem (Sqoop)
  • Developed custom Spark User-Defined Functions (UDFs) by extending generic abstract UDFs, implementing an efficient initialization process
  • Implemented Spark using Python, leveraging Data Frames and Spark SQL API for faster and more streamlined data processing.

Software Engineer

MICROINFOINC
Dallas, Texas
02.2014 - 06.2017
  • Assisted in the development and implementation of ETL processes, contributing to streamlined data workflows
  • Conducted data cleaning and preparation tasks, ensuring data quality and integrity
  • Collaborated with data engineers to develop data pipelines, focusing on improving data quality and accessibility
  • Communicated relevant development decisions and shared their implications on development work
  • Coordinated with other technical staff involved in development technologies and decisions
  • Managed data extraction jobs and built new data pipelines from various structured and unstructured sources
  • Composed shell scripts for extracting information from MySQL servers to HDFS
  • Converted hard and complex ETL processes into Hadoop using ecosystems
  • Performed testing of software prototypes and facilitated the transfer to the operational team
  • Analyzed current programs for performance, diagnosis, and troubleshooting, designing effective solutions using Spark Data Frame
  • Demonstrated proficiency in writing Spark applications using both Python and Scala
  • Worked on different file formats (OrcFile, TextFile) and various compression codecs (GZIP, SNAPPY, LZO)
  • Implemented Spark using Scala, leveraging DataFrames and Spark SQL API for efficient data processing
  • Designed and built complete ETL processes for migrating and transforming data for Staging, EDW, and Data Marts using AWS API and S3
  • Wrote SAS scripts for comparing data sets
  • Migrated tables from RDBMS into Hive tables using SQOOP, later transitioning to AWS S3 API for moving on-prem data to S3
  • Experienced in writing test cases and implementing unit test cases using testing frameworks like Junit (Scala).

Education

Masters of Science in management information systems -

Bellevue University

Masters of computer applications -

Andhra University

Skills

  • Hadoop frameworks
  • Map Reduce
  • HDFS
  • Pig
  • Hive
  • Sqoop
  • Spark
  • DB2 UDB
  • Oracle
  • SQLPlus
  • PL/SQL
  • ETL processes
  • Java
  • RDBMS
  • Linux shell scripting
  • Web/intranet technologies
  • Client/server technologies
  • Java J2EE Servlets JSP JSF EJB JDBC SQL
  • Snowflake utilities (SnowSQL)
  • Big Data models using Python
  • Software development life cycle (SDLC)
  • Database design
  • Technical documentation preparation
  • Data workflow descriptions solution design and architecture technical requirements specifications
  • Agile engineering practices
  • CI/CD processes

Certification

Certificate in Data Science, Colorado Technical University (Coding Dojo), https://app.diplomasafe.com/en-US/diploma/d71e90bf261680ca33d6aa94e8e96b0b9fb494adf

SQL For Data Science https://coursera.org/share/7ea21e0747b32ad1001ff9ff156e55d9

Key Skills And Achievements

  • Proficient in loading and analyzing large datasets using Hadoop frameworks, including Map Reduce, HDFS, Pig, Hive, Sqoop, and Spark.
  • Strong engineering background complemented by close collaboration with business customers, ensuring a deep understanding of business needs, objectives, and expectations in data projects.
  • Extensive work on database applications utilizing DB2 UDB, Oracle, SQLPlus, PL/SQL
  • Development and implementation of ETL processes.
  • Hands-on experience in application development using Java, RDBMS, and Linux shell scripting.
  • Senior Java Developer with expertise in web/intranet and client/server technologies, employing Java, J2EE, Servlets, JSP, JSF, EJB, JDBC, and SQL.
  • Familiarity with Snowflake utilities (SnowSQL) and proficiency in implementing Big Data models using Python.
  • In-depth knowledge of the software development life cycle (SDLC) and database design.
  • Expertise in technical documentation preparation, including detailed data workflow descriptions, solution design and architecture, and technical requirements specifications.
  • Experienced in Agile engineering practices, with a focus on implementing complete CI/CD processes, from design to development.
  • Proven ability to prioritize tasks effectively and provide accurate estimations.
  • Conducts thorough code reviews and maintains direct communication with customers.

Linkdin

  • https://www.linkedin.com/in/uday-routhu-4130b4104/

github

  • https://github.com/uday-routhu

Timeline

Quantitative Analytics Senior

Freddie Mac
11.2021 - Current

Software Engineer

Compunnel Software Group, Inc
07.2018 - 11.2021

Software Engineer

nDimensions
03.2017 - 06.2018

Software Engineer

MICROINFOINC
02.2014 - 06.2017

Masters of Science in management information systems -

Bellevue University

Masters of computer applications -

Andhra University
Uday Routhu