Summary
Overview
Work History
Education
Skills
Timeline
Generic

Rahul Vaidya

Chesterfield,VA

Summary

Highly skilled Senior Data Architect/Analyst with 20+ years of experience known for delivering robust data solutions and driving strategic initiatives. Proven expertise in creating scalable data models and enhancing data governance frameworks to support business objectives. Collaborative team player with focus on achieving measurable outcomes and adapting to evolving project requirements. Skilled in data integration and cloud computing technologies.

Overview

23
23
years of professional experience

Work History

Senior Data Architect

Magellan Health
05.2020 - Current
  • Developed an enterprise data architecture plan to ensure data is consistent among various departments and secure across multiple application and software systems
  • Designed a logical and physical data model for the organization's database system, ensuring scalability and maintainability
  • This helped facilitate the smooth handshaking of various applications with respect to their tables and relationships
  • Implemented ETL processes to integrate disparate datasets into the organization's big-data platform thus reducing the approximate cost of around $90,000/ year for using legacy server system and their yearly maintenance
  • Determined optimal approaches like using the Tableau tool and AWS S3 buckets for storing, managing, analyzing, and visualizing large volumes of complex data
  • Performed detailed analysis of existing data sources to determine their suitability for integration into the new architecture per the data migration plan
  • Assisted in setting up and managing Snowflake accounts, including creating databases, schemas, and user roles to ensure secure and organized data storage
  • Developed and managed enterprise-wide data analytics environments
  • Conducted performance-tuning activities by optimizing queries and indexes or recommending any suitable upgrades if necessary
  • Developed Python scripts to automate data extraction from various sources, such as databases and web APIs thus eliminating the need for production support teams to be on the clock for any failures.
  • Monitored Snowflake usage and performance metrics through the Snowflake UI and queries, identifying opportunities to improve query execution efficiency
  • Executed SQL queries in Snowflake to retrieve, analyze, and manipulate data, supporting day-to-day reporting and data exploration tasks
  • Created data visualizations and reports for management, showcasing findings and recommendations thus eliminating the old ways of sending data to customers in Excel formats and which were non interactive

Senior Data Architect

Capital One
10.2018 - 05.2020
  • Developed complex SQL queries to analyze large datasets for trends and insights thus reducing manual effort for cumbersome analysis of Excel sheet data
  • Automated daily reporting processes with Python scripting techniques thus reducing man hours for manually generating the reports after hours further reducing cost
  • Created unit tests for existing and new Python code using the PyTest framework to get the needed objective along with the right accuracy
  • Troubleshot, debugged, and fixed issues related to existing Python codebase
  • Utilized Pandas library for data analysis, manipulation, and visualization purposes
  • Employed Git version control system for tracking changes in source code repositories over time so that there is proper coordination of check-in and checkout of codes when there were multiple developers
  • Developed basic data pipelines in Databricks using PySpark to clean, transform, and prepare datasets for further analysis
  • Created and managed notebooks in Databricks to visualize data insights and generate reports for business stakeholders
  • Supported data integration tasks by connecting Databricks to cloud storage solutions like Azure Data Lake or AWS S3 for seamless data ingestion
  • Participated in troubleshooting and debugging SQL queries and pipelines, resolving errors and ensuring data accuracy in Snowflake workflows
  • Performed data loading and unloading tasks using SnowSQL and integrated cloud storage platforms like AWS S3 and Azure Blob Storage to facilitate seamless data movement

Senior Data Analyst

CarMax
10.2017 - 09.2018
  • Created data visualizations in Tableau to present findings to 20 state stakeholders thus helping them to get the past, present, and future behavior of market trends for car sales
  • Developed complex SQL queries to analyze large datasets for trends and insights combined with the above Tableau reports thus eliminating the old non-interactive ways of Excel data analysis
  • Established KPI metrics and monitored the performance of data analysis projects thus giving a new outlook to the stakeholders about the different interactive ways to analyze data
  • Provided data-driven recommendations for marketing campaigns and product development initiatives based on historical data sales, customer behavior, and demographic information
  • This way was able to keep the cost of the marketing campaigns well within the projected budget thus saving significant cost
  • Collaborated with cross-functional teams across departments such as Sales, Marketing, and Finance on the various relational database systems and developed shell script jobs
  • Analyzed and tracked data to prepare forecasts and identify trends
  • Analyzed competitive market strategies through related product, market or share trends and developed visualizations to compare the competitors and built in-house strategic techniques for enhancing our data against competitor data

Senior Data Analyst

Capital One Bank
03.2015 - 10.2017
  • Developed shell scripts to automate and streamline system administration tasks thus eliminating 15 on call resources as part of production support teams monitoring the jobs at night time and who had to run them manually at various intervals at night time thus eliminating night time support resources saving significant cost savings
  • Managed source control repositories like GIT and Subversion for versioning code bases thus keeping the up-to-date latest code and no conflict between the multiple developers working at the same time
  • Developed Python scripts to automate data extraction from various sources, such as databases thus reducing/eliminating manual intervention
  • Performed exploratory data analysis using Pandas, Numpy, and Matplotlib libraries in Python
  • Created dashboards in Tableau to monitor key performance indicators
  • Utilized SQL queries to extract relevant information from relational databases

Project Manager

Amgen
11.2013 - 03.2015
  • Analyzed data from various sources such as surveys, interviews, and focus groups to inform decision-making related to projects
  • Presented findings from data analysis in a clear, concise manner that was easily understood by stakeholders at all levels of the organization preventing the stakeholders from unnecessarily purchasing legacy servers, software, and developers for traditional ways of data analysis giving them a projected saving of $235,000 per year
  • Negotiated contracts with vendors and suppliers based on company objectives and budget constraints having a projected saving of almost $45,000/ year till the lifetime of the project
  • Ensured compliance with applicable laws, regulations, policies, and procedures throughout the life cycle of each project eliminating any regulatory fines from the federal government
  • Facilitated resolution of conflicts between team members when necessary thus delivering every project with smooth accuracy, budget and well within timelines thus getting rewarded as 'Most Valued Manager' for 3 quarters in a row

Data Architect

Capital One
10.2011 - 10.2013
  • Developed complex SQL queries to analyze large datasets for trends and insights and removal of duplicates, bad characters thus maintaining data quality
  • Automated daily reporting processes and also created data models for functional and technical requirements
  • Worked on PL/SQL and Unix Shell scripts for automating jobs and scheduled them in cron thus removing dependency of 5 on call resources who used manual corrections
  • Created unit tests for existing and new Python code using the PyTest framework to get the needed objective along with the right accuracy
  • Troubleshoot, debugged, and fixed issues related to existing Python codebase
  • Worked with basic data analysis using Pandas library for manipulation, and visualization purposes
  • Used Confluence, JIRA, and Sharepoint for tracking the sprint and documents

ETL Architect / ETL Team Lead

Genworth Financial
03.2004 - 10.2011
  • Developed and implemented ETL processes to extract, transform, and load data into a central repository
  • Designed, built, and maintained efficient ETL architectures for various source systems and also worked in parallel with cost-saving initiatives
  • Created technical design documents for all ETL development projects and developed reusable mappings and applets to avoid developing duplicate code thus saving cost and time
  • Monitored performance of ETL jobs daily using automated tools such as SQL Server Agent Jobs or Unix Shell Scripts
  • Scheduled these jobs in cron so we could eliminate the manual intervention from the night shift resources thus saving significant costs
  • Decommissioned 2 legacy servers from 3 which were housing similar relational information on them which could easily be stored on one of the 3 thus achieving a significant cost saving of $325,000
  • Provided leadership and guidance to team members, ensuring that tasks were completed on time and to a high standard making my team the most efficient team 2 years in a row
  • Conducted performance appraisals for team members, providing feedback and identifying areas of improvement

Software Engineer

Mahindra & Mahindra
03.2002 - 06.2004
  • Analyzed user requirements to develop software solutions and created technical specifications
  • Conducted unit tests on code modules to verify the accuracy and functionality of program logic
  • Monitored system performance and troubleshoot issues in a production environment
  • Collaborated with cross-functional teams to ensure quality assurance throughout the development process
  • Created detailed design documents, test plans, user manuals, release notes, and other related documentation

Education

Bachelors in Engineering - Electronics And Telecommunication

Mumbai University
Mumbai India
11.2002

Skills

  • Proficient in Business Intelligence Tools
  • Proficient in Python Programming
  • Change Management Techniques
  • ETL Workflow Design
  • Data Architecture, Integration and Modeling Skills
  • PL/SQL Development and Performance tuning
  • Data Warehousing Solutions
  • Informatica and Snaplogic Development
  • Tableau Dashboard Development
  • SDLC and Agile
  • UNIX Shell Programming and Perl Programming
  • Data Quality Assurance and Data Analysis
  • Project Management
  • Proficient in Python
  • Data Migration Execution
  • Relational Database Knowledge

Timeline

Senior Data Architect

Magellan Health
05.2020 - Current

Senior Data Architect

Capital One
10.2018 - 05.2020

Senior Data Analyst

CarMax
10.2017 - 09.2018

Senior Data Analyst

Capital One Bank
03.2015 - 10.2017

Project Manager

Amgen
11.2013 - 03.2015

Data Architect

Capital One
10.2011 - 10.2013

ETL Architect / ETL Team Lead

Genworth Financial
03.2004 - 10.2011

Software Engineer

Mahindra & Mahindra
03.2002 - 06.2004

Bachelors in Engineering - Electronics And Telecommunication

Mumbai University
Rahul Vaidya