Proficient IT professional with over 9 years of experience, specialized in Data Engineering, Big Data, Back-end Engineering, Ecosystem- Data Acquisition, Ingestion, Modelling, Storage Analysis, Integration, Data Processing. Extensive Experience in working with GCP/AWS Databricks, Synapse Analytics, Azure Data Factory, Stream Analytics, AWS/Azure Analysis Services, Data Lake, Azure Storage, Azure SQL Database, SQL Data Warehouse, Azure Cosmos DB. Expertise in working with Azure services like HDInsight, Application Insights, Azure Monitoring, Azure AD, Function apps, Logic apps, Event Hubs, Iot hubs, Storage Explorer, Key Vault. Strong working experience with SQL and NoSQL databases (Azure Cosmos DB, MongoDB, HBase, Cassandra), data modeling, tuning, disaster recovery, backup and creating data pipelines. Have extensive experience in creating pipeline jobs, schedule triggers using Azure dat. Have good experience designing cloud-based solutions in Azure by creating Azure SQL database, setting up Elastic pool jobs and designing tabular models in Azure analysis services. Strong knowledge in working with ETL methods for data extraction, transformation and loading in corporate-wide ETL Solutions and Data Warehouse tools for reporting and data analysis. Experience in all phases of Data Warehouse development like requirements gathering, design, development, implementation, testing, and documentation. Acquired profound knowledge in developing production ready Spark applications utilizing Spark Core, Spark Streaming, Spark SQL, DataFrames, Datasets and Spark-ML. Expertise in building PySpark and Spark applications for interactive analysis, batch processing and stream processing. Extensively used Spark Data Frames API over Cloudera platform to perform analytics on Hive data and also used Spark Dataframe Operations to perform required Validations in the data. Strong Hadoop and platform support experience with all the entire suite of tools and services in major Hadoop Distributions – Cloudera, Azure HDInsight, AWS and Hortonworks. Hands on experience in using Hadoop ecosystem components like Hadoop, Hive, Pig, Sqoop, HBase, Cassandra, Spark, Spark Streaming, Spark SQL, Oozie, Zookeeper, Kafka, Airflow, Flume, MapReduce framework, Yarn. Skilled in using Azure authentication and authorization and experience in using Visualization tools like Tableau, Power BI. Basic hands-on experience working with Kusto. Strong knowledge in working with ETL methods for data extraction, transformation and loading in corporate-wide ETL Solutions and Data Warehouse tools for reporting and data analysis. Capable of using AWS utilities such as EMR, S3 and cloud watch to run and monitor Hadoop and spark jobs on Amazon Web Services (AWS). Working knowledge with Amazon S3, Amazon EC2, AWS Kinesis to provide a complete solution for computing, query processing, and storage across a wide range of applications. Experienced in working with micro batching to ingest millions of files on Snowflake cloud when files arrive at the staging area. Ingested data into Snowflake cloud data warehouse using Snowpipe. Good experience on maintaining version control using code versioning tools & Azure Devops. Skilled in creating dashboards in Power BI, Tableau and in Jupyter notebooks using (Matplotlib, Seaborn). Experience in configuring and monitoring data processing and data storage solutions. Experience in importing and exporting the data using Sqoop from HDFS to Relational Database Systems and from Relational Database Systems to HDFS. Workflows (Requirement study, Analysis, Design, Coding, Testing, Deployment, and Maintenance) in Event-driven and client/Server application development. Have expertise in Cloud Cost Management and Optimization techniques involved in projects.