Around 8+ Years of IT background in delivering end-to-end data analytics solutions.
GCP Proficiency: Mastery over GCP services such as BigQuery, Google Cloud Storage (GCS) buckets, Cloud Functions, and Dataflow for seamless data analytics solutions.
GCP Command Line Tools: Solid grasp of Cloud Shell, GSUTIL, and BQ command line utilities for efficient management of GCP resources.
Data-driven setting fostering collaborations to maintain the edge necessary for delivering products and services built on innovative technologies like Artificial Intelligence and Machine Learning.
Data Pipeline Execution: Successfully designing and executing data pipelines within the GCP platform, optimizing data flow strategies for insights generation.
Containerization and Kubernetes: Proficient handling of Docker and Kubernetes for efficient containerization and orchestration, particularly experienced in managing deployments using Google Kubernetes Engine (GKE).
Azure DevOps Expertise: Proficiency in building reusable YAML pipelines, creating CI/CD pipelines using Azure DevOps, and implementing Git flow branching strategies.
Azure Cloud Services: Solid understanding of Azure services including Databricks, Data Factory, Data Lake, and Function Apps for effective data management and analytics.
Efficient Data Integration: Expertise in designing and deploying SSIS packages for data extraction, transformation, and loading into Azure SQL Database and Data Lake Storage.
Hadoop Proficiency: Strong support experience across major Hadoop distributions - Cloudera, Amazon EMR, Azure HDInsight, Hortonworks, utilizing tools such as HDFS, MapReduce, Spark, Kafka, Hive, and more.
Real-time Data Solutions: Proficient in building real-time data pipelines and analytics using Azure components like Data Factory, HDInsight, and Stream Analytics.
API Development and Integration: Experienced in developing highly scalable and resilient RESTful APIs, ETL solutions, and third-party platform integrations within the GCP ecosystem.
AWS Cloud Services: Proficiency in AWS cloud services like EC2, S3, Glue, Athena, DynamoDB, RedShift, and hands-on experience with Hadoop ecosystem tools.
Legacy Data Migration: Led successful migration projects from Teradata to AWS Redshift and on-premises to AWS Cloud, ensuring seamless SQL database migration to GCP's Data Lake, BigQuery, and other relevant services.
AWS Cloud-Based Pipelines: Utilization of AWS services like EMR, Lambda, and Redshift to develop cloud-based pipelines and Spark applications.
Azure Data Services: ETL expertise using Azure Data Factory, T-SQL, Spark SQL, and U-SQL Azure Data Lake Analytics, along with ingestion and processing within Azure Databricks.
GCP Data Services Excellence: Expertise in GCP data services including Cloud Composer for orchestrating data tasks and profound ETL experience using GCP services like Dataflow.
DevOps and Scripting Proficiency: Skilled in PowerShell scripting, Bash, YAML, JSON, GIT, Rest API, and Azure Resource Management (ARM) templates for effective pipeline management.
Data Visualization and Analysis: Proficient in creating data visualizations using Python, Scala, Tableau, and developing Spark scripts for data transformation.
Big Data Ecosystem: Extensive experience with Amazon EC2, Azure Cloud, and Big Data tools like Hadoop, HDFS, MapReduce, Hive, HBase, Spark, Kafka, Flume, Avro, Sqoop, and PySpark.
Database Migration: Expertise in migrating SQL databases to AWS Redshift and Azure Data Lake, along with managing SQL databases, Parquet files, and parsing JSON formats.
Cloud Computing and Big Data Tools: Proficiency in AWS Cloud and Azure components, with working knowledge of Spark using Scala and PySpark.
Real-time Data Solutions: Building real-time data pipelines and analytics using AWS components and setting up workflows with tools like Apache Airflow and Oozie.
Database Expertise: Working with SQL Server and MySQL databases, skilled in setting up workflows with Apache Airflow.
IDE and Version Control: Proficient use of version control systems like Git, along with popular IDEs like PyCharm and IntelliJ for efficient code management.
Spark Streaming: Developing Spark streaming modules for RabbitMQ and Kafka data ingestion in Azure and GCP environments.
Windows Scripting and Cloud Containerization: Proficient in scripting and debugging within Windows environments, familiarity with container orchestration, Kubernetes, Docker, and Azure Kubernetes Service (AKS).