Summary
Overview
Work History
Education
Skills
Certification
Accomplishments
Timeline
Generic

APARNA RUBY VENUKUMAR

VA

Summary

Senior ETL Developer with 12 years of experience in the field of Information Technology focusing on Data warehousing with Employers like Markel, Cognizant Technology Solutions, UST Global, IBM/Damian Consulting, Accenture/Artech Consulting, Dexian Signature Consulting Primary skillsets include Microsoft Azure Cloud computing(Data Factory, Databricks, Synapse, Data Lake storage, Keyvaults), Informatica Power center, PowerShell, Python, Automate BPA tool, GitHub, SQL Server, Teradata / BTEQ , Unix, Oracle, Hadoop (TDCH)/ Hive, Control M, Tivoli Proven ability in working with major US based clients like Markel(Specialty Insurance), Kaiser Permanente(Healthcare), Elevance Health/Anthem(Healthcare), Abbott Laboratories (Lifescience), Corelogic (Banking and Finance)

Overview

13
13
years of professional experience
1
1
Certification

Work History

Senior Software Engineer

Markel Insurance(Data Services)
Henrico, VA
01.2025 - Current
  • Worked as ETL Developer/Lead on Agile or scrum teams for complete delivery of solutions
  • Technologies that involved Extraction, Transform and data Load (ETL) using Microsoft Azure Cloud computing (Data Factory, Databricks, Synapse), Informatica, PowerShell, Python, SQL Server, Automate BPA Scheduler
  • Understand business requirements, provide the Design flow, build and unit test Data Solutions and application software in a multi-technology environment.
  • Analysis of Business Requirements
  • Microsoft Azure Cloud computing (Data Factory, Databricks, Synapse, DataLake Storage) : Designed and developed Data Factory pipelines, triggers using Linked services/Datasets , Keyvaults, managing resources to store credentials, creation of Managed Private endpoints to automate solutions
  • PowerShell and Python : Perform database/synapse and ETL operations, folder/file operations like Archival, Creation and Deletion, incorporating, email and logger functionalities
  • Informatica : analyze, design, develop, deploy Slowly Changing Dimension Mappings , workflows and monitor jobs
  • Automate BPA Scheduler tool : to schedule and integrate fully developed ETL artifacts and develop and deploy fully automated jobs
  • Worked on Performance Tuning, parameterizing and Query optimization of various ETL jobs
  • Align with design practices in support of the technology and architecture in support of the Data Services tools and platform.
  • Undertakes complete design of complex software applications or components applying agreed standards, patterns and tools.
  • Assists as part of a team in the design of components of larger software systems.
  • Collaborates in reviews of work with others as appropriate.
  • Designs, codes, verifies, tests, documents, amends and refactors moderately complex programs/scripts.
  • Applies agreed standards and tools, to achieve a well-engineered result.
  • Applies data analysis, design, modelling, and quality assurance techniques, based upon a detailed understanding of business processes, to establish, modify or maintain data structures and associated components (entity descriptions, relationship descriptions, attribute definitions).
  • Advises database designers and other application development team members on the details of data structures and associated components.
  • Defines the software modules needed for an integration build and produces a build definition for each generation of the software.
  • Produces integration test specifications, conducts tests and records and reports on outcomes.
  • Diagnoses faults and records and reports on the results of tests.
  • Produces system integration reports.
  • Installs or removes hardware and/or software, using supplied installation instructions and tools including, where appropriate, handover to the client.
  • Conducts tests, corrects malfunctions, and documents results in accordance with agreed procedures.
  • Reviews requirements and specifications, and defines test conditions.
  • Designs test cases and test scripts under own direction, mapping back to pre-determined criteria, recording and reporting outcomes.
  • Analyses and reports test activities and results.
  • Identifies and reports issues and risks associated with own work.
  • Maintains professional and technical knowledge by attending educational workshops; reviewing professional publications; establishing personal networks; benchmarking state-of-the-art practices; participating in professional societies.
  • Identifies and resolves issues with applications
  • Create and maintain system documentation including entity, component/logical, and physical diagrams.
  • Design repeatable, automated build and deployment strategies to ensure simple, successful deployments across environments.
  • Participate in production support activities.
  • Identify enhancements and innovation opportunities within the data Services organization.
  • Mentor junior developers for performing system and technical tasks
  • Environment: Azure Data Factory, Azure Databricks, Azure Synapse, Azure Data Lake Storage, Informatica Power Center 11.4.0, SQL Server, Powershell, Python, Automate BPA, Git, Confluence

Data Engineer

Dexian Consulting/Markel Insurance(Data Services)
Henrico, VA
03.2021 - 12.2024
  • Worked on complete SDLC from Extraction, Transformation and Loading (ETL) of data using Informatica.
  • Provide the Design flow and data modeling based on the Analysis of Business Requirements
  • Client interaction for the requirement clarifications and documenting the business and mapping logics accordingly
  • Prepare the Technical Design documentations once the scope is defined and requirements are analyzed
  • Suggest changes in the mapping documents on technical grounds
  • Design and develop the system architecture that involved Membership and Claims subject areas
  • Implemented Slowly changing dimensions, to load the data from stage to DW layer.
  • The main Transformations used for the extraction and manipulation Expression, Aggregator, Lookup, Router, Filter, Sequence generator, Update Strategy, etc. to transform and cleanse the data according to business rules.
  • Prepare the Mapping documents as per the source Flat files received on a weekly basis
  • Using Workflow Manager for Workflow/worklets and Session Management, database connection management and Scheduling of jobs to be run in the batch process.
  • Captured data error records corrected and loaded into target system.
  • Worked on Performance Tuning and optimization of the Sessions, Mappings, and Transformations that helped us to reach the goal of decreasing the time taken for ETL process.
  • Involved in process Improvement techniques and Automation of manual intervention of tasks
  • Stored Procedures, Indexing were used to performance tune the high volume of data
  • Identify tools and technologies to be used in the project as well as reusable objects that could be customized for the project.
  • Usage of SQL joins for subquery implementations in the mappings for lookup, source qualifier transformations; creation of Table Partitions, Global Temporary tables,etc
  • Involved in monitoring the workflows and in optimizing the load times
  • Used Change Data Capture (CDC) to simplify ETL in data warehouse applications
  • Develop, schedule and migration of Control M jobs for ETL artifacts
  • Analysis of existing BTEQ scripts for new logic modification
  • Review the code and prepare the objects for migration
  • Error handling and Debugging of any performance bottlenecks
  • Rational Clear Case was used to Controlling versions of all files & Folders (Check-out, Check-in) for deployment of the artifacts
  • Participate in meetings and provide updates on project progress
  • Creation of Teradata BTEQ scripts for loading the data into the stage and DW tables
  • Performance Tuning done using the usage of Volatile tables, Global temporary tables and derived queries
  • Unix scripting usage for the creating the backup and compressing the source files, pre and post session operations
  • Perform Unit Testing and prepare Unit test documentations
  • Fully participate in System Integration Testing of ETL jobs resolving issues/concerns
  • Used Agile methodology where I was involved in scrum events like Sprint planning meetings, grooming the user stories, facilitated Story Point discussions to analyze the level of effort on project specifications, created technical tasks under user stories assigned, tracked them to closure.
  • Drive solution delivery by closely working with business users using all phases of SDLC Waterfall methodology
  • Given Technical clarifications and assigned work to off-shore team and monitored their work status daily
  • Prepare/maintain all documentation relevant to the development process in the Confluence and linked them to the corresponding user stories in JIRA.
  • Environment: Informatica Power Center 9.6, Teradata, Oracle, UNIX, Control M, JIRA, Confluence, IBM Rational Clearcase

Data Engineer

Kaiser Permanente/Accenture/Artech
Remote, VA
07.2020 - 12.2020
  • Worked on the end-to end SDLC phases that involved Extraction, Transform and data Load (ETL) using Informatica.
  • Provide the Design flow and data modeling based on the Analysis of Business Requirements
  • Translated the business requirements into functional specifications and technical design documents to visualize the KPMC flow, integrate different regional membership data
  • The mapping documents are reviewed and observations are sent to the Solution Consulting Team
  • Prepare the Technical Design documentations once the scope and requirements are finalized
  • Design and developed the system architecture using membership related data marts by providing accurate data, view of core membership information from across all regions
  • Data load from Flat file and Oracle tables are performed on the basis of 3 layer process –
  • Extract mapping: to load the data from Stage tables to Confirming tables and Load Mapping (SCD Type2): to load data from Confirming table to DW tables / Views.
  • Complex Crosswalk mappings were also developed to store the Source system keys of different Legacy and KPMC extracts
  • Perform History and incremental loads for Legacy and KPMC extracts
  • Worked on Performance Tuning, parameterizing and optimization of the Sessions, Mappings, and Transformations that helped us to reach the goal of decreasing the time taken for ETL process.
  • Usage of complex SQL joins for joining high volume tables, implementing Line numbers for a set of key columns, implementations in the mappings for lookup, source qualifier transformations; creation of Table Partitions, Global Temporary tables, etc
  • Review the code and prepare the objects for migration
  • Perform Unit Testing and prepare Unit test documentations ,worked with SIT, UAT and Prod support Teams for defect resolutions and its implementation with the help of Release management teams
  • Automation performed to deploy the ETL code using batch script where the parameters involved are the Deployment group, , Source Repository, Target Repository, email address, Request number
  • Auto load Balancing ETL artifacts created for count , date, null, duplicate checks
  • Error handling tables created to capture the Referential integrity check, duplicate check
  • Unix scripting for the renaming the source file, creating the backup , copying the Source file for ETL processing and compressing it, pre and post session operations, auto deployment, etc
  • Schedule the job using Tivoli for ETL artifacts
  • Given Technical clarifications and assigned tasks to off-shore team and monitored their work status daily
  • Leading the new end-to-end POC environment creation from ETL Team
  • Environment: Informatica Power Center 10.2.0, Oracle, UNIX, Tivoli, JIRA, Confluence, JAZZ

ETL Developer

Anthem/UST Global
Remote, VA
10.2017 - 01.2018
  • Involved in daily scrum meetings for agile methodology project specifications, created technical tasks under user stories assigned, tracked them to closure.
  • Prepare/maintain all documentation relevant to the development process in the Confluence and linked them to the corresponding user stories in JIRA.
  • Developed flow diagrams for Business process and designed the ETL processes using Informatica to load data from sources like Teradata, Flat Files to target database.
  • Extensively used ETL to load data from Flat files which involved both fixed width as well as Delimited files
  • Worked Informatica power center tools like source analyzer, mapping designer, mapplet and transformations
  • Developed complex Informatica mappings to load the data from various sources using different transformations like source qualifier, lookup, expression, aggregator, joiner, filter, router, sequence generator transformations
  • Extensive worked on Informatica 9.6.1 (mappings, sessions, workflows) in creating various mappings focusing on SCD1 and SCD2 implementation.
  • Developed Re-Usable Transformations, Re-Usable Mapplets, sessions and workflows for designed mappings.
  • Environment: Informatica Power Center 8.5.1, Teradata, Oracle 11g, UNIX, Windows

ETL Developer

Cognizant Technology Solutions
Kerala, Bangalore
12.2011 - 10.2017
  • Worked with the business users to provide the deliverables using Waterfall methodology
  • Involved in Sprint planning meetings, grooming the user stories, facilitated Story Point discussions to analyze the level of effort on project specifications, created technical tasks under user stories assigned, tracked them to closure using Agile/ Scrum methodology
  • Maintain all documentation relevant to the development process in the Confluence and linked them to the corresponding user stories in JIRA.
  • Provide the Design flow and data modeling based on the Analysis of Business Requirements
  • Converted the business rules into technical specifications and developed the ETL code for extract,transform and load into the Dimension/Fact tables in the Data warehouse
  • Worked on performance tuning and optimization of the Sessions, Mappings, and Transformations that helped us to reach the goal of decreasing the time taken for ETL process Code development, creation of database objects and ETL mappings and Deployment of the objects
  • Designed and developed ETL Mappings using Informatica to extract data from various sources like Salesforce, Oracle and to load the data into the target database
  • Involved in preparing the Source to Target Mapping documentation
  • Captured error records, corrected and loaded into the target system
  • Usage of Stored Procedures, Indexing, Sequences, Synonyms, Collect stats to tune the performance of complex queries
  • Debugging and error handling of ETL mappings and sessions
  • Monitoring failed sessions and workflows
  • Created UNIX scripts to perform operations like sending post-session completion email notifications, zip the flat files, remove and touch files
  • Perform Code Reviews and validate if the checklist is met
  • Review the code and prepare the objects for migration
  • Worked in Process Improvement/Automation of the project.
  • Participate in meetings and provide updates on project progress
  • Perform Unit Testing and prepare Unit test documentations
  • Schedule and monitor jobs using Autosys Job Scheduling
  • Involved in Maintenance activities like provide KT sessions to the Support Team, Bug fixing, Code & Data Analysis
  • Direct Interaction with Client Business team for the requirements and documenting the business logics and mapping logics accordingly
  • Involved in knowledge transitions from Development teams and Lights On team regarding changes to existing applications in Production.
  • Environment: Informatica Power Center, Oracle, DB2, Autosys, UNIX, Windows

Education

B.Tech - Information Technology

University of Kerala
05.2011

Secondary Education - 12th - CBSE

St. Thomas Central School
Trivandrum
03.2007

Primary Education - 10th - CBSE

St. Thomas Central School
Trivandrum
03.2005

Skills

  • Microsoft Azure Cloud computing
  • Data Factory
  • Databricks
  • Synapse
  • Data Lake storage
  • Keyvaults
  • Informatica Power center
  • PowerShell
  • GitHub
  • Python
  • Automate BPA tool
  • SQL Server
  • Teradata
  • BTEQ
  • Unix
  • Oracle
  • Hadoop
  • Hive
  • Control M
  • Tivoli
  • Azure
  • Informatica Power Center Client
  • Teradata SQL Assistant
  • Toad
  • Citrix
  • Putty
  • Autosys
  • Jira
  • Confluence
  • DB Visualiser
  • WinSCP
  • MKS Integrity Client
  • IBM Rational Clearcase
  • JAZZ
  • Bitbucket
  • DB2
  • Shell Scripting

Certification

  • Cognizant Certified Professional (CCP) in Data Warehousing from Cognizant Technology Solutions
  • Cognizant Certified Professional (CCP) in Informatica from Cognizant Technology Solutions
  • Cognizant Certified Professional (CCP) in Oracle, Unix from Cognizant Technology Solutions
  • LEARNING PATH: Complete Roadway to Informatica Powercenter 9 from Udemy

Accomplishments

  • Received Certificate of Appreciation ‘Rainmaker-Go-Getter’ Award for Delivery Excellence in the Enterprise Information Management Domain in (CTS Organization)
  • Recognition by Abbott Laboratories Client for intra-account Automation contest for Process improvement (CTS Organization)

Timeline

Senior Software Engineer

Markel Insurance(Data Services)
01.2025 - Current

Data Engineer

Dexian Consulting/Markel Insurance(Data Services)
03.2021 - 12.2024

Data Engineer

Kaiser Permanente/Accenture/Artech
07.2020 - 12.2020

ETL Developer

Anthem/UST Global
10.2017 - 01.2018

ETL Developer

Cognizant Technology Solutions
12.2011 - 10.2017

B.Tech - Information Technology

University of Kerala

Secondary Education - 12th - CBSE

St. Thomas Central School

Primary Education - 10th - CBSE

St. Thomas Central School
APARNA RUBY VENUKUMAR