Having around 10+ years of Software developer experience in Application development/Architecture and Data Analytics with specialization in Web Applications and Client-Server, Big data applications with Java and Big Data Technologies and expertise in Java, Scala, Python, Spark, Pyspark, Hadoop Map Reduce, in various industrial sectors including but not limited to banking, insurance, education, cloud and 3+ year of experience in AWS, Kafka, ElasticSearch, Devops and Linux Administration.
Strong Big data analytics experience using the Hadoop Ecosystem tools Map-Reduce, HDFS, Apache Spark, JupyterLab, AWS EMR, Jira, Glue.
Site Reliability Engineering responsibilities for Kafka platform that scales 2 GB/Sec and 20 million messages/sec.
Hands-on cloud platform experience with vast AWS services like AWS EMR, Glue, Redshift, SWF, S3, IAM, SQS, SNS, IAM, Cloudformation, Cloudwatch, DynamoDB.
Experienced with Spark improving the performance and optimization of the existing algorithms in Hadoop using Spark Context, Spark-SQL, Data Frame, Pair RDD's, Spark YARN.
Developed Big data applications using Apache Spark and Apache Hadoop, handling 200 million records at the smallest scale.
Hands-on experience in developing SPARK applications using Spark tools like RDD transformations, Spark core, Spark Dataframes, Spark Set Operations and Spark SQL.
Excellent Programming skills at a higher level of abstraction using Java and Python.
Experience in using D-Streams, Accumulator, Broadcast variables, JIRA, Rally, Remedy, RDD caching for Spark.
Working knowledge of Amazon Elastic Cloud Compute (EC2) infrastructure for computational tasks and Simple Storage Service (S3) as Storage mechanism.