Accomplished Principal Data Engineer with extensive experience in leading large-scale transformative data projects and mentoring multiple teams on cutting-edge cloud technologies. Proven expertise in data application migration from on-premises to cloud environments, with a strong focus on AWS Cloud. Demonstrated ability to guide teams of enterprise data architects and data engineers in implementing robust data solutions. Adept at designing and managing comprehensive Data Warehousing and ETL processes. Skilled in handling big data processing and real-time data processing using data streaming technologies. Proficient in integrating disparate systems through APIs, ensuring seamless data flow and connectivity across the systems. Detail-oriented data engineer designs, develops and maintains highly scalable, secure and reliable data structures. Accustomed to working closely with system architects, software architects and design analysts to understand business or industry requirements to develop comprehensive data models. Proficient at developing database architectural strategies at the modeling, design and implementation stages.
Overview
19
19
years of professional experience
Work History
Principal Data Engineer
Fannie Mae
07.2015 - Current
Successfully led teams responsible for migrating all Data Warehouse Applications and Operational Data Stores to AWS cloud.
Spearheaded implementation of best practices for data processing on AWS Cloud, ensuring efficient and reliable data operations.
Provided comprehensive strategy for parallel execution of cloud and on-premises systems during transition period, ensuring business continuity and minimal disruption.
Trained multiple data engineers on Enterprise Data Lake and related technologies, fostering knowledgeable and skilled team.
Defined robust data governance rules and standards to ensure safe and compliant operation of data processes on cloud.
Successfully lead ETL development team and created robust and scalable ETL solutions to meet Data Warehouse business requirements on Cloud
Migrated legacy systems to modern big-data technologies, improving performance and scalability while minimizing business disruption.
Collaborated with cross-functional teams for seamless integration of data sources into company''s data ecosystem.
Hands-on knowledge of Data warehousing ETL solutions using Staging Areas, ODS and Data Marts
Strong experience in Dimensional modeling, Solution architecture, Technical architecture, Data Profiling, Data cleansing technologies and ETL testing strategies
Currently working on Uniform Loan Delivery Project which is ongoing effort by Freddie Mac and Fannie Mae to enhance the accuracy and quality of loan data each GSE requires at loan delivery
Implemented Data Streaming based solution using to process real-time data from streams to insights
Implemented Data Privacy standards to ensure Non-Public information is secure and handled properly
Worked on detailed ETL design for the Data Mart project and presented to all the stakeholders including Business Team, System Testing Team, UAT and Operations Team to ensure that the design meets the expectations of all the stakeholders and completely satisfy the business and operational requirements
Responsible for handling issues raised by all the developers, coordinating with business teams and ensuring quality of Agile deliverables
Lead Enterprise Data Architect
Blue Cross Blue Shield
03.2008 - 07.2015
Designed and implemented scalable, high-performance data architecture solutions to support enterprise-level applications and analytics.
Led the design and development of data integration and ETL processes to consolidate data from disparate sources into a unified data platform.
Leveraged big data technologies (e.g., Hadoop, Spark) to create robust data processing frameworks, enhancing the organization's ability to handle large datasets efficiently.
Architected and deployed cloud-based data platforms (AWS, Azure, Google Cloud), optimizing data storage, processing, and retrieval.
Established and enforced data security protocols and compliance measures (e.g., GDPR, HIPAA) to protect sensitive information and adhere to regulatory requirements.
Enabled advanced analytics and business intelligence by designing data warehouses, data lakes, lake house and analytics platforms that support real-time and batch processing.
Evaluated and selected data technologies, tools, and platforms to meet evolving business needs and technological advancements.
Sr. ETL Developer
22nd Century Technologies Inc.
11.2005 - 03.2008
Worked with Fannie Mae Ab Initio Support Team (FAST) to implement ETL standards and Best Practices in the development process
Extensively worked on improving the performance of Ab Initio graphs by using Various Ab Initio Performance Techniques
Suggested multiple architecture and design approaches to development teams, like CLMRE (Credit Loss Management Reporting Environment), ACH (Automated Clearing House), Capital Markets PDS, GL Datamart project, for scalable solution and performance enhancements
Participated in code and design reviews for most of the Restatement projects to ensure the ETL design/code is in compliance with the company AI standards to ensure that graphs are designed and developed to meet the expectations in terms of Maintainability, Auditability, Code Optimization, Standards, Documentation, Restartability and Recoverability
Developed many generic re-usable Ab Initio graphs that can be used in multiple jobs to perform similar ETL activities like Generic Extract and Generic Load graphs
Extensive knowledge of Ab Initio multifile system and partition components for better performance when dealing with large volumes of data
Worked as part of ETL Administration Team to create projects on development EME and ETL servers in compliance with corporate AI standards
Worked with multiple AI application Teams to implement ETL standards and Best Practices in the development process
Wrote UNIX script to check the AI standards for the graphs
The script is useful in enforcing AI standards and helps in creating consistent and more maintainable graphs
The script works on deployed graphs and checks for standards like MAX-CORE, World write Access on output files, max-memory parameter, phases, checkpoints and other standard statements required to put in the start and end script of the graph/project as part of corporate AI standard
Wrote UNIX script to create standard data directories for serial and multifile system
The script sources the sandbox project parameters and uses global .abinitiorc file to figure out the location and directories to be created in development, test, acceptance and production environments
Extensively used EME commands to resolve Developer issues while using EME
Worked with AI Corporation on multiple AI Software internal problems like GDE performance issue, memory leak issue while unloading large volume of data using Input table component
Involved in providing Ab Initio/ETL training for multiple development Team members, and other activities associated with developing individual and Team skills, abilities and knowledge of the Ab Initio
Automated multiple spreadsheets using Microsoft Visual Basic for Applications to programmatically generate Ab Initio Code and Design Review documents in MS Word
Programmed Excel spreadsheets to generate in-house metrics report and Ab Initio ETL Server statistics reports with charts.
Operational Risk Principal (Director) – Single Family Strategy and Governance at FANNIE MAEOperational Risk Principal (Director) – Single Family Strategy and Governance at FANNIE MAE