Over 5+ years of experience in Data Engineering, Data Pipeline Design, Development, and Implementation as a Data Engineer/Data Developer and Data Modeler. Strong experience in Software Development Life Cycle (SDLC) including requirements Analysis, Design Specification, and Testing as per Cycle in both Waterfall and Agile methodologies. Experience developing SPARK applications using Spark tools like RDD transformations, Spark core, Spark MLlib, Spark Streaming, and Spark SQL. Experience refactoring the existing spark batch process for different logs written in Scala. Experience in Hadoop Ecosystem components Map - Reduce, HDFS, Yarn/MRv2, Hive, HDFS, HBase, Spark, Kafka, Sqoop, Flume, Avro, Sqoop, AWS, Avro, Solr and Zookeeper. Experience developing applications using Map Reduce to analyze Big Data with different file formats. Experience with Apache Spark components including SPARK CORE, SPARK SQL, SPARK STREAMING, and SPARK MLLIB. Experience in data analysis using Hive, and Impala. Experience working on creating and running Docker images with multiple micro-services. Experience in Data Modeling and ETL processes in data warehouse environments such as star schema, and snowflake schema. Experience structural modifications using Map-Reduce, and Hive and analyze data using visualization/reporting tools (Tableau). Experience working with GitHub/Git source and version control systems. Experience in Azure Cloud, Azure Data Factory, Azure Data Lake Storage, Azure Synapse Analytics, Azure Analytical services, Azure Cosmos NO SQL DB, Azure HD Insight Big Data Technologies (Hadoop and Apache Spark), and Data bricks. Experience designing Azure Cloud Architecture and Implementation plans for hosting complex application workloads on MS Azure. Experience in Amazon Web Services (AWS) concepts like EC2, S3, EMR, ElasticCache, DynamoDB, Redshift, Aurora. Experience developing scripts using Python or Shell Scrips to Extract, Load, and Transform. Experience in developing JSON scripts for deploying the pipeline in Azure Data Factory (ADF), which processes the data using the Cosmos Activity. Experienced in code repositories like GitHub. Experience in using SQOOP to import and export data from RDBMS to HDFS and Hive. Experience in multiple databases like MongoDB, Cassandra, MySQL, ORACLE, and MS SQL Server. Experience in agile software development methodology. Ability to work effectively in cross-functional team environments, excellent communication and interpersonal skills. Excellent communication skills, interpersonal skills, problem-solving skills a very good team player along with a can-do attitude and ability to effectively communicate with all levels of the organization such as technical, management, and customers.