Highly skilled and dedicated Data Engineer with over 5 years of experience in software analysis, design, development, and implementation of Cloud and Big Data solutions. Proficient in leveraging technologies such as Big Query, Spark, Scala, Hadoop, and Oracle Database to build and maintain robust data pipelines.
Extensive expertise in developing data models, pipeline architectures, and providing ETL solutions for project models. Managed end-to-end operations of ETL data pipelines using Matillion on AWS Cloud Services and Azure Data Factory on Azure Cloud Services, ensuring seamless data ingestion, Data Processing/Transformation,Data Curation.
Proven ability to design and specify Informatica ETL processes, optimizing schema loading and performance. Skilled in ETL architecture design and implementation, consistently delivering high-performance solutions.
Certified in software engineering concepts, with hands-on experience in system design, application development, testing, and operational stability. Proficient in coding using modern programming languages and database querying languages, ensuring efficient and maintainable code.
Proficient in utilizing Python for data manipulation, analysis, and scripting, enabling efficient data processing and transformation. Skilled in PySpark, leveraging the power of Apache Spark for distributed data processing, machine learning, and real-time analytics. .
Strong troubleshooting and problem-solving skills, capable of identifying and resolving complex software issues. Exhibits excellent communication and collaboration skills, enabling effective teamwork and coordination.
A driven and motivated professional, consistently striving for excellence in data engineering, delivering scalable and high-quality solutions to meet business needs.
• Design and setup Enterprise Data Lake to provide support for various uses cases including Analytics, processing, storing, and Reporting of voluminous, rapidly changing data.
• Responsible for maintaining transactional data in the source by performing operations such as cleaning, transformation and ensuring Integrity in a relational environment by working closely with the stakeholders & solution architect.
• Creating SQL Plus scripts and packages to generate comprehensive reports.
• Developing and automating Shell scripts to streamline processes and eliminate manual tasks.
• Adapting existing logics or developing new ones to meet evolving customer requirements. Managing monthly data transfers from mainframe systems to Oracle databases.
• Collaborated with business stakeholders to analyze requirements and develop customized SQL logic, ensuring system alignment with evolving business needs.
• Proficient in Databricks data streaming tech architecture, with a strong understanding of building Analysis Services reporting models.
• Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems (RDBMS) and from RDBMS to HDFS.
• Experienced in connecting to various data sources in Databricks, importing data, and transforming it for Business Intelligence purposes.
• Leverage Azure Databricks to migrate on-premises data to the cloud, optimizing data processing and analytics capabilities.
• Develop and execute data pipelines using Azure Databricks to transform and load data into cloud-based data warehouses or data lakes.
GPA: 4.0/4.0
GPA: 71.72