Having 10+ years professional career in IT - System Analysis, Design and Development, Migration, Production support and Maintenance projects. Very good experience in full Software Development Life Cycle (SDLC), Agile Methodology, Azure Cloud services (PaaS and IaaS), Azure DevOps, Bigdata, Hadoop eco system, Data Warehousing concepts and Abinitio. 6+ Years Experience in Azure Data Implementation services and Bigdata Spark Ecosystem (Azure SQL, Blob storage, Azure Synapse Analytics, Azure Data Lake Storage, Azure Databricks, Event Hubs, Azure Data Factory, Logic App,IOT devices, Cosmos DB, Postgre SQL, Azure DevOps,Spark Scala, Sqoop and PySpark) Certified professional of DP-200 & DP-201 Designing and implementing an Azure Data solution. Very Good experience in migrating the end-to-end Oracle, Netezza DWH and AbInitio ETL applications into Azure. Good experience in creating Generic Framework pipelines and Transformations using ADF, PySpark and Azure Databricks Notebook. Good Experience in migrating the MySQL, SQL server on-premises DBs to Azure Cloud using Azure DMS and SSMS. Extensive experience in building the ETL/ELT (Extract Loading and Transform) pipelines using ADF,PySpark and Azure Databricks with different file formats(csv, Json , delta, Parquet, Avro and text files) Expertise with the tools in Hadoop Ecosystem including Hive, HDFS, MapReduce, Sqoop, Spark and Yarn. Hands-on experience in designing and maintaining Online Transaction Processing (OLTP) databases. Knowledgeable in schema design for transactional databases to ensure data integrity and consistency. Skilled in managing and optimizing databases to handle high transaction volumes in real-time. Can work parallelly in both GCP and Azure clouds coherently. Have Extensive Experience in IT data analytics projects, Hands on experience in migrating on premise ETLs to Google Cloud Platform (GCP) using cloud native tools such as BIG query, Cloud Data Proc, Google Cloud Storage, Composer. Good experience in building Databricks Lake house, creating data frames and data ingestion modules to ingest the data into HDFS from heterogeneous data sources using PySpark and Sqoop. Very good experience in Data Bricks Cluster Configuration & Azure Synapse Cost Management, creating HDInsight cluster and Storage Account with End-to-End environment for running the jobs. Expertise in data warehousing concepts, ELT/ETL programming using ADF, Abinitio and other data warehouse lifecycle tasks such as data analysis, transforming, debugging/testing, and data loading across target systems. Good experience of full life cycle implementation of DWBI application, Star Schema and Snowflake model. Having strong interpersonal, problem solving & team skills, quick learner, able to meet deadlines, ability to interact with people at various levels. Experience in using build/deploy tools such as Jenkins and Azure Devops CI/CD Extensive experience in Onsite-Offshore model in terms of estimation, work allocation, code review, System Integration Testing, User Acceptance testing, deployment planning and Postproduction warranty support. Extensive knowledge of business process and functioning of Health Care, Telecom, Mortgage, Financial, Retail and Insurance sectors. Proficient in performance tuning of SQL queries for OLTP environments. Very good experience in Coordinating different teams to troubleshoot the critical issues without impacting the BAU process.
· Developed Spark applications using Databricks, pyspark and Spark-SQL for data extraction, transformation, and aggregation from multiple file formats for analyzing & transforming the data to uncover insights into the customer usage patterns.
· Developed Spark applications using Databricks, Python and Spark and Spark-SQL for data extraction, transformation, and aggregation from multiple file formats for analyzing & transforming the data to uncover insights into the customer usage patterns.