Accomplished Data Engineer with expertise in cloud data architecture and ETL development, honed at Depository Trust & Clearing Corporation. Proven track record in designing scalable data migration solutions and implementing metadata governance frameworks. Adept in Python and collaborative problem-solving, driving impactful data-driven decisions across organizations.
Innovative Database Architect with a solid background in designing, implementing, and maintaining complex database solutions. Possess strong skills in data modeling, system scalability, and ensuring data integrity across diverse platforms. Effective communicator and collaborator who has contributed to enhancing database performance and supporting business objectives through strategic planning and technology integration.
Overview
6
6
years of professional experience
Work History
Data Engineer
Depository Trust & Clearing Corporation
Tampa
09.2023 - Current
Led multiple data migration projects by designing ingestion architectures for JSON data transfer to Snowflake using Apache Camel and integrated Collibra for metadata governance.
Developed scalable ingestion solutions with Apache Camel workflows for real-time cloud data access and analytics for a complete centralized solution for data.
Built ETL and ELT pipelines utilizing REST APIs and file-based patterns to populate Snowflake Data Warehouse for business intelligence reporting.
Created centralized data governance frameworks, including glossaries and quality rules, accessible to stakeholders.
Engineered ingestion applications with Shell scripting, Java, and Spark SQL to transform data for analytics.
Developed CI/CD pipelines using Jenkins and Bitbucket for automated deployment processes.
Maintained ingestion pipelines with JDBC connectors to extract data from DB2, PostgreSQL, Oracle, and APIs into Snowflake.
Designed custom ELT/ETL frameworks managing sensitive data with HashiCorp Vault for secure handling.
Applied Data Quality Rules using Ataccama and developed and completed a complete automated setup for Ataccama.
Data Engineer
Albertsons
Fremont
02.2022 - 08.2023
Designed and implemented Big Data analytics architectures to migrate JSON data from Snowflake and Azure Cosmos DB using Azure Databricks, Kafka consumers, and producers.
Developed scalable ETL pipelines with Azure Data Factory for migrating data from traditional sources to cloud environments.
Engineered Spark applications in Python, Java, and Spark SQL for processing and analyzing customer behavior, pricing trends, and inventory optimization.
Automated CI/CD pipelines with Jenkins and GitHub Actions for efficient deployments on on-prem Linux servers.
Migrated legacy Snowflake ETL workflows to Azure Synapse Analytics, performing performance analysis and tuning with PCF log analytics.
Built OLAP structures including fact and dimension tables, cubes, and MDX queries using SQL Server Analysis Services and SSRS.
Created SSIS packages for transforming multi-format flat files and Excel data into structured SQL databases.
Implemented complex ADF pipelines for seamless data movement between Azure SQL, Blob Storage, SQL Data Warehouse, and Talend-based pricing systems.
Data Administrator
University of East London, UEL
London
06.2019 - 07.2020
Collaborated on skills-matching projects, analyzing and categorizing resumes to identify student opportunities across multiple departments.
Supported development of user-centric data management tools, enhancing accessibility for academic staff.
Utilized Talend on Ubuntu Linux to build ETL pipelines for transforming and migrating academic data.
Assisted faculty in creating data-driven research models by preparing datasets and performing exploratory data analysis.
Analyzed student data to uncover trends in enrollment and resource allocation, enabling informed decision-making.
Conducted ad-hoc data analysis using Python and R, providing insights for research initiatives and policy design.
Delivered training sessions on data governance protocols, improving team efficiency and data handling accuracy.
Presented demonstrations on Hive and HDFS usage, establishing a foundation for an ETL/data engineering career.
Education
Master of Science - Computer Science
Rivier University
Nashua, NH
12-2021
Master of Science - Computer Science
University of East London
London
05-2020
Skills
Data migration using Enterprise Data Ingestion tool and Azure Data Factory (ADF)
ETL development
Cloud data architecture with both AWS and Azure cloud services
Data warehousing using Snowflake
Metadata governance using Collibra
Data quality management using Ataccama
Python, Spark, Java, SQL, NoSQL
Data visualization tools like Tableau, Amazon QuickSight, and Power BI
Database options like MySQL, PostgreSQL, MongoDB, Azure Cosmos DB, IBM DB2, and Data Lake
CICD tools with Bitbucket, GitHub, and Jenkins
Execution environments with EC2, Windows, and Unix/Linux