HDFS, MapReduce, Spark, Hive, Sqoop, Flume, Kafka, Oozie, Pig, HBase, Well-rounded IT experience covering various Big Data technologies, Spark, and database development, Demonstrated proficiency in Amazon Web Service (AWS) concepts, particularly in EMR and EC2, ensuring fast and efficient processing of Teradata Big Data Analytics, Proficient in Google Cloud Platform (GCP), adding versatility to cloud service utilization, Expertise in transforming intricate business requirements into analytical models, involving algorithm design, model building, and solution development for Data Mining, Data Acquisition, Data Preparation, Data Manipulation, Feature Engineering, and scalable Machine Learning Algorithms, Proven ability to develop solutions that scale across massive volumes of both structured and unstructured data, Experience in working with Hadoop distributions like Cloudera and Hortonworks, Excellent experience in end-to-end ETL processes, including Designing, Developing, Documenting, and Testing of ETL jobs and mappings in Server and Parallel jobs using Data Stage, Hands-on experience in Apache Spark, Spark Streaming, and Spark SQL, Familiarity with NoSQL databases like HBase, Cassandra, and MongoDB, Established and executed comprehensive Data Quality Governance Frameworks, ensuring decisions align with intended purposes through end-to-end processes, Expert in designing Server jobs using various types of stages like Sequential file, ODBC, Hashed file, Aggregator, Transformer, Sort, Link Partitioner, and Link Collector, Proficient in a wide range of Big Data Practices and Technologies, including HDFS, MapReduce, Hive, Pig, HBase, Sqoop, Oozie, Flume, Spark, and Kafka, Expertise in designing Parallel jobs, utilizing various stages like Join, Merge, Lookup, remove duplicates, Filter, Dataset, Lookup file set, Complex flat file, Modify, Aggregator, and XML, Demonstrated a commitment to continuous learning and skill enhancement throughout the career