Boasting over 6 years of robust experience as a Data Engineer, skilled across various domains including IT services and financial services. Adept in utilizing Hadoop ecosystem technologies such as Apache Hadoop, HDFS, MapReduce, Apache Hive, and Apache Pig for complex data processing. Proficient with Azure cloud services including Azure HDInsight, and Azure Stream Analytics, focusing on solutions that scale across data-heavy environments. Expert in Python scripting to automate and streamline data processes, enhancing efficiency and productivity. Demonstrated ability to manage Git for effective version control, ensuring smooth workflows in collaborative and distributed development environments. Versed in executing SQL-driven projects for database management, showcasing the ability to handle complex data manipulation and querying tasks. Experienced with a comprehensive suite of AWS services including S3, Redshift, and EMR, optimizing cloud-based data storage and processing. Capable of integrating MongoDB into data solutions, emphasizing performance in NoSQL database implementations. Specialized in Apache Kafka for effective real-time data streaming and processing within high-demand environments. Proficient in the deployment of SSIS and Apache Airflow for robust ETL processes, ensuring data integrity and timely delivery. Hands-on experience with AWS Schema Conversion Tool and Hadoop YARN, focusing on data schema standardization and resource management. Currently engaged with cutting-edge technologies such as Apache Spark, which facilitates large-scale data processing and analytics. Utilizing Azure SQL Database, ADLS, and ADF to develop advanced data handling and storage solutions in the cloud. Implementing Azure Cosmos DB for globally distributed database services, ensuring high availability and elastic scalability. Employing Terraform for efficient infrastructure as code applications, enabling reproducible and scalable cloud environments. Leveraging Snowflake for dynamic cloud data warehousing, enhancing data retrieval and analytics capabilities. Integrating Azure Active Directory to manage secure access and identity management within complex project environments. Utilizing PyTorch in financial models to predict and analyze data, leading to more informed decision-making processes. Focused on automating ETL and data processing tasks to improve operational efficiencies and reduce manual errors. Ensuring robust data security and compliance through stringent measures and regular audits in sensitive data environments. Dedicated to continuous professional development, keeping abreast with the latest data engineering tools and methodologies. Committed to mentoring junior data engineers, fostering a learning environment and sharing expert knowledge in data engineering.