Senior ETL Developer with 12 years of experience in the field of Information Technology focusing on Data warehousing with Employers like Markel, Cognizant Technology Solutions, UST Global, IBM/Damian Consulting, Accenture/Artech Consulting, Dexian Signature Consulting Primary skillsets include Microsoft Azure Cloud computing(Data Factory, Databricks, Synapse, Data Lake storage, Keyvaults), Informatica Power center, PowerShell, Python, Automate BPA tool, GitHub, SQL Server, Teradata / BTEQ , Unix, Oracle, Hadoop (TDCH)/ Hive, Control M, Tivoli Proven ability in working with major US based clients like Markel(Specialty Insurance), Kaiser Permanente(Healthcare), Elevance Health/Anthem(Healthcare), Abbott Laboratories (Lifescience), Corelogic (Banking and Finance)
Overview
13
13
years of professional experience
1
1
Certification
Work History
Senior Software Engineer
Markel Insurance(Data Services)
Henrico, VA
01.2025 - Current
Worked as ETL Developer/Lead on Agile or scrum teams for complete delivery of solutions
Technologies that involved Extraction, Transform and data Load (ETL) using Microsoft Azure Cloud computing (Data Factory, Databricks, Synapse), Informatica, PowerShell, Python, SQL Server, Automate BPA Scheduler
Understand business requirements, provide the Design flow, build and unit test Data Solutions and application software in a multi-technology environment.
Analysis of Business Requirements
Microsoft Azure Cloud computing (Data Factory, Databricks, Synapse, DataLake Storage) : Designed and developed Data Factory pipelines, triggers using Linked services/Datasets , Keyvaults, managing resources to store credentials, creation of Managed Private endpoints to automate solutions
PowerShell and Python : Perform database/synapse and ETL operations, folder/file operations like Archival, Creation and Deletion, incorporating, email and logger functionalities
Informatica : analyze, design, develop, deploy Slowly Changing Dimension Mappings , workflows and monitor jobs
Automate BPA Scheduler tool : to schedule and integrate fully developed ETL artifacts and develop and deploy fully automated jobs
Worked on Performance Tuning, parameterizing and Query optimization of various ETL jobs
Align with design practices in support of the technology and architecture in support of the Data Services tools and platform.
Undertakes complete design of complex software applications or components applying agreed standards, patterns and tools.
Assists as part of a team in the design of components of larger software systems.
Collaborates in reviews of work with others as appropriate.
Applies agreed standards and tools, to achieve a well-engineered result.
Applies data analysis, design, modelling, and quality assurance techniques, based upon a detailed understanding of business processes, to establish, modify or maintain data structures and associated components (entity descriptions, relationship descriptions, attribute definitions).
Advises database designers and other application development team members on the details of data structures and associated components.
Defines the software modules needed for an integration build and produces a build definition for each generation of the software.
Produces integration test specifications, conducts tests and records and reports on outcomes.
Diagnoses faults and records and reports on the results of tests.
Produces system integration reports.
Installs or removes hardware and/or software, using supplied installation instructions and tools including, where appropriate, handover to the client.
Conducts tests, corrects malfunctions, and documents results in accordance with agreed procedures.
Reviews requirements and specifications, and defines test conditions.
Designs test cases and test scripts under own direction, mapping back to pre-determined criteria, recording and reporting outcomes.
Analyses and reports test activities and results.
Identifies and reports issues and risks associated with own work.
Maintains professional and technical knowledge by attending educational workshops; reviewing professional publications; establishing personal networks; benchmarking state-of-the-art practices; participating in professional societies.
Identifies and resolves issues with applications
Create and maintain system documentation including entity, component/logical, and physical diagrams.
Design repeatable, automated build and deployment strategies to ensure simple, successful deployments across environments.
Participate in production support activities.
Identify enhancements and innovation opportunities within the data Services organization.
Mentor junior developers for performing system and technical tasks
Environment: Azure Data Factory, Azure Databricks, Azure Synapse, Azure Data Lake Storage, Informatica Power Center 11.4.0, SQL Server, Powershell, Python, Automate BPA, Git, Confluence
Data Engineer
Dexian Consulting/Markel Insurance(Data Services)
Henrico, VA
03.2021 - 12.2024
Worked on complete SDLC from Extraction, Transformation and Loading (ETL) of data using Informatica.
Provide the Design flow and data modeling based on the Analysis of Business Requirements
Client interaction for the requirement clarifications and documenting the business and mapping logics accordingly
Prepare the Technical Design documentations once the scope is defined and requirements are analyzed
Suggest changes in the mapping documents on technical grounds
Design and develop the system architecture that involved Membership and Claims subject areas
Implemented Slowly changing dimensions, to load the data from stage to DW layer.
The main Transformations used for the extraction and manipulation Expression, Aggregator, Lookup, Router, Filter, Sequence generator, Update Strategy, etc. to transform and cleanse the data according to business rules.
Prepare the Mapping documents as per the source Flat files received on a weekly basis
Using Workflow Manager for Workflow/worklets and Session Management, database connection management and Scheduling of jobs to be run in the batch process.
Captured data error records corrected and loaded into target system.
Worked on Performance Tuning and optimization of the Sessions, Mappings, and Transformations that helped us to reach the goal of decreasing the time taken for ETL process.
Involved in process Improvement techniques and Automation of manual intervention of tasks
Stored Procedures, Indexing were used to performance tune the high volume of data
Identify tools and technologies to be used in the project as well as reusable objects that could be customized for the project.
Usage of SQL joins for subquery implementations in the mappings for lookup, source qualifier transformations; creation of Table Partitions, Global Temporary tables,etc
Involved in monitoring the workflows and in optimizing the load times
Used Change Data Capture (CDC) to simplify ETL in data warehouse applications
Develop, schedule and migration of Control M jobs for ETL artifacts
Analysis of existing BTEQ scripts for new logic modification
Review the code and prepare the objects for migration
Error handling and Debugging of any performance bottlenecks
Rational Clear Case was used to Controlling versions of all files & Folders (Check-out, Check-in) for deployment of the artifacts
Participate in meetings and provide updates on project progress
Creation of Teradata BTEQ scripts for loading the data into the stage and DW tables
Performance Tuning done using the usage of Volatile tables, Global temporary tables and derived queries
Unix scripting usage for the creating the backup and compressing the source files, pre and post session operations
Perform Unit Testing and prepare Unit test documentations
Fully participate in System Integration Testing of ETL jobs resolving issues/concerns
Used Agile methodology where I was involved in scrum events like Sprint planning meetings, grooming the user stories, facilitated Story Point discussions to analyze the level of effort on project specifications, created technical tasks under user stories assigned, tracked them to closure.
Drive solution delivery by closely working with business users using all phases of SDLC Waterfall methodology
Given Technical clarifications and assigned work to off-shore team and monitored their work status daily
Prepare/maintain all documentation relevant to the development process in the Confluence and linked them to the corresponding user stories in JIRA.
Environment: Informatica Power Center 9.6, Teradata, Oracle, UNIX, Control M, JIRA, Confluence, IBM Rational Clearcase
Data Engineer
Kaiser Permanente/Accenture/Artech
Remote, VA
07.2020 - 12.2020
Worked on the end-to end SDLC phases that involved Extraction, Transform and data Load (ETL) using Informatica.
Provide the Design flow and data modeling based on the Analysis of Business Requirements
Translated the business requirements into functional specifications and technical design documents to visualize the KPMC flow, integrate different regional membership data
The mapping documents are reviewed and observations are sent to the Solution Consulting Team
Prepare the Technical Design documentations once the scope and requirements are finalized
Design and developed the system architecture using membership related data marts by providing accurate data, view of core membership information from across all regions
Data load from Flat file and Oracle tables are performed on the basis of 3 layer process –
Extract mapping: to load the data from Stage tables to Confirming tables and Load Mapping (SCD Type2): to load data from Confirming table to DW tables / Views.
Complex Crosswalk mappings were also developed to store the Source system keys of different Legacy and KPMC extracts
Perform History and incremental loads for Legacy and KPMC extracts
Worked on Performance Tuning, parameterizing and optimization of the Sessions, Mappings, and Transformations that helped us to reach the goal of decreasing the time taken for ETL process.
Usage of complex SQL joins for joining high volume tables, implementing Line numbers for a set of key columns, implementations in the mappings for lookup, source qualifier transformations; creation of Table Partitions, Global Temporary tables, etc
Review the code and prepare the objects for migration
Perform Unit Testing and prepare Unit test documentations ,worked with SIT, UAT and Prod support Teams for defect resolutions and its implementation with the help of Release management teams
Automation performed to deploy the ETL code using batch script where the parameters involved are the Deployment group, , Source Repository, Target Repository, email address, Request number
Auto load Balancing ETL artifacts created for count , date, null, duplicate checks
Error handling tables created to capture the Referential integrity check, duplicate check
Unix scripting for the renaming the source file, creating the backup , copying the Source file for ETL processing and compressing it, pre and post session operations, auto deployment, etc
Schedule the job using Tivoli for ETL artifacts
Given Technical clarifications and assigned tasks to off-shore team and monitored their work status daily
Leading the new end-to-end POC environment creation from ETL Team
Environment: Informatica Power Center 10.2.0, Oracle, UNIX, Tivoli, JIRA, Confluence, JAZZ
ETL Developer
Anthem/UST Global
Remote, VA
10.2017 - 01.2018
Involved in daily scrum meetings for agile methodology project specifications, created technical tasks under user stories assigned, tracked them to closure.
Prepare/maintain all documentation relevant to the development process in the Confluence and linked them to the corresponding user stories in JIRA.
Developed flow diagrams for Business process and designed the ETL processes using Informatica to load data from sources like Teradata, Flat Files to target database.
Extensively used ETL to load data from Flat files which involved both fixed width as well as Delimited files
Worked Informatica power center tools like source analyzer, mapping designer, mapplet and transformations
Developed complex Informatica mappings to load the data from various sources using different transformations like source qualifier, lookup, expression, aggregator, joiner, filter, router, sequence generator transformations
Extensive worked on Informatica 9.6.1 (mappings, sessions, workflows) in creating various mappings focusing on SCD1 and SCD2 implementation.
Developed Re-Usable Transformations, Re-Usable Mapplets, sessions and workflows for designed mappings.
Environment: Informatica Power Center 8.5.1, Teradata, Oracle 11g, UNIX, Windows
ETL Developer
Cognizant Technology Solutions
Kerala, Bangalore
12.2011 - 10.2017
Worked with the business users to provide the deliverables using Waterfall methodology
Involved in Sprint planning meetings, grooming the user stories, facilitated Story Point discussions to analyze the level of effort on project specifications, created technical tasks under user stories assigned, tracked them to closure using Agile/ Scrum methodology
Maintain all documentation relevant to the development process in the Confluence and linked them to the corresponding user stories in JIRA.
Provide the Design flow and data modeling based on the Analysis of Business Requirements
Converted the business rules into technical specifications and developed the ETL code for extract,transform and load into the Dimension/Fact tables in the Data warehouse
Worked on performance tuning and optimization of the Sessions, Mappings, and Transformations that helped us to reach the goal of decreasing the time taken for ETL process Code development, creation of database objects and ETL mappings and Deployment of the objects
Designed and developed ETL Mappings using Informatica to extract data from various sources like Salesforce, Oracle and to load the data into the target database
Involved in preparing the Source to Target Mapping documentation
Captured error records, corrected and loaded into the target system
Usage of Stored Procedures, Indexing, Sequences, Synonyms, Collect stats to tune the performance of complex queries
Debugging and error handling of ETL mappings and sessions
Monitoring failed sessions and workflows
Created UNIX scripts to perform operations like sending post-session completion email notifications, zip the flat files, remove and touch files
Perform Code Reviews and validate if the checklist is met
Review the code and prepare the objects for migration
Worked in Process Improvement/Automation of the project.
Participate in meetings and provide updates on project progress
Perform Unit Testing and prepare Unit test documentations
Schedule and monitor jobs using Autosys Job Scheduling
Involved in Maintenance activities like provide KT sessions to the Support Team, Bug fixing, Code & Data Analysis
Direct Interaction with Client Business team for the requirements and documenting the business logics and mapping logics accordingly
Involved in knowledge transitions from Development teams and Lights On team regarding changes to existing applications in Production.
Environment: Informatica Power Center, Oracle, DB2, Autosys, UNIX, Windows
Education
B.Tech - Information Technology
University of Kerala
05.2011
Secondary Education - 12th - CBSE
St. Thomas Central School
Trivandrum
03.2007
Primary Education - 10th - CBSE
St. Thomas Central School
Trivandrum
03.2005
Skills
Microsoft Azure Cloud computing
Data Factory
Databricks
Synapse
Data Lake storage
Keyvaults
Informatica Power center
PowerShell
GitHub
Python
Automate BPA tool
SQL Server
Teradata
BTEQ
Unix
Oracle
Hadoop
Hive
Control M
Tivoli
Azure
Informatica Power Center Client
Teradata SQL Assistant
Toad
Citrix
Putty
Autosys
Jira
Confluence
DB Visualiser
WinSCP
MKS Integrity Client
IBM Rational Clearcase
JAZZ
Bitbucket
DB2
Shell Scripting
Certification
Cognizant Certified Professional (CCP) in Data Warehousing from Cognizant Technology Solutions
Cognizant Certified Professional (CCP) in Informatica from Cognizant Technology Solutions
Cognizant Certified Professional (CCP) in Oracle, Unix from Cognizant Technology Solutions
LEARNING PATH: Complete Roadway to Informatica Powercenter 9 from Udemy
Accomplishments
Received Certificate of Appreciation ‘Rainmaker-Go-Getter’ Award for Delivery Excellence in the Enterprise Information Management Domain in (CTS Organization)
Recognition by Abbott Laboratories Client for intra-account Automation contest for Process improvement (CTS Organization)