Over 5 years of experience in Data Engineering, specializing in GCP, and AWS with expertise in building scalable data lakes, ETL pipelines, real-time data processing, and cloud-based analytics. Designed and developed centralized data lakes on GCP, leveraging Cloud Storage, Dataproc, BigQuery, and BigTable for efficient data storage and processing. Built scalable ETL workflows using Cloud Dataflow, Dataproc with Spark, and Apache Airflow, automating data ingestion and transformation pipelines. Implemented real-time data ingestion architectures using Druid and Kafka on GCP, ensuring low-latency data processing for analytics and reporting. Collaborated with ML engineers to integrate data pipelines with AI/ML models. Strengthened data security and compliance by implementing GCP IAM policies, role-based access control, encryption, and data masking techniques. Orchestrated multi-source data ingestion pipelines using Cloud Composer and Cloud Dataproc, ensuring seamless integration with diverse data sources. Developed scalable, high-performance code in Python and Scala for complex data transformations and workflow automation. Designed and maintained AWS-based data solutions, developing data lakes on Amazon S3 and optimizing them with partitioning strategies and lifecycle policies. Built and managed ETL pipelines using AWS Glue, Lambda, and Apache Airflow, streamlining automated data transformations and processing. Implemented real-time streaming solutions with AWS Kinesis, Spark Streaming, and Apache Kafka, ensuring continuous data availability and processing. Worked on multi-terabyte data migrations from Oracle to AWS, storing optimized copies in Amazon Redshift for business intelligence and reporting. Optimized Redshift clusters, focusing on schema design, query performance tuning, and workload management for enhanced data analytics. Involved in data security on AWS by configuring IAM roles, S3 bucket policies, and AWS KMS encryption, ensuring regulatory compliance. Developed real-time dashboards and analytics solutions using Tableau, and AWS Athena improving data accessibility and business insights. Automated Tableau dashboard updates using Python and AWS Lambda, reducing manual effort and improving real-time reporting efficiency.