Over10+ years of work experience in IT field, involved in all phases of software development lifecycle while working in different projects. Very strong experience in processing, analyzing large sets of structured, semi-structured and unstructured data and supporting systems application architecture. Extensive experience in writing Hadoop jobs for data analysis as per the business requirements using Hive and Pig. Expertise in creating Hive Internal/External Tables/Views using shared Meta store. Having Good knowledge of Pyspark and Scala. . Developed the Code using SCALA and pyspark. Having knowledge about serverless technique like AWS Lambda, Athena,AWS Batch,S3,EMR. Having Support Experience in Production Side. Have experience of working on Snowflake and Vertica data warehouse. Worked extensively on SQOOP to import and export data from RDBMS to HDFS and vice-versa. Proficient in big data ingestion and streaming tools like Sqoop, Kafka and Spark. Experience of working on data formats like Avro, Parquet. Hands on experience in Sequence files, RC files, Combiners, Counters, Dynamic Partitions, Bucketing for best practice and performance improvement. Experience creating real-time data streaming solutions using Apache Spark core, Spark SQL, Kafka, spark streaming and Apache Storm. Worked on Oozie to manage and schedule the jobs on Hadoop cluster. Done data Migration from abinitio,Informatica. Having the knowledge of PL/SQL queries. Knowledge of developing analytical components using Scala. Experience in managing and reviewing Hadoop log files. Worked with NoSQL database HBase to create tables and store data. Experience in setting up Hive, Pig, HBase, and SQOOP on Ubuntu Operating system. Strong experience in design and development of relational database concepts with multiple RDBMS databases including Oracle10g, MySQL, MS SQL Server & PL/SQL,Netezza. Proficient in using data visualization tools like Tableau, Raw and MS Excel. Experience in developing web interfaces using technologies like XML, HTML, DHTML and CSS. Implemented functions, stored procedures, triggers using PL/SQL. Good understanding of ETL processes and Data warehousing. Strong experience in writing UNIX shell scripts. Working in different projects provided exposure and good understanding of different phases in SDLC. Deploying the code via Jenkins/Bitbucket. Merging the code to master after validating to Bitbucket (similar to GIT).