I work as a Senior Data Engineer with 7 years of experience in designing and building data solutions that support large-scale, enterprise-grade analytics. In my current role, I’m part of a team responsible for developing and maintaining end-to-end ETL pipelines that power critical data platforms within our Organization. Our work involves integrating and transforming data from multiple sources, ensuring it is clean, reliable, and accessible for downstream applications. We build and manage data workflows using technologies like SQL, Python, Spark, and HQL, while leveraging platforms such as AWS (EMR, S3, Lambda), Azure Data Factory, Snowflake, and Hadoop for cloud-based storage, processing, and orchestration. My day-to-day responsibilities include writing and optimizing PySpark jobs, orchestrating ETL pipelines, and ensuring seamless data movement between systems, both cloud-native and on-premises. I work closely with cross-functional teams to ensure data consistency, optimize job performance, and align data workflows. I also take an active role in implementing data quality checks, improving pipeline reliability, and driving automation to enhance operational efficiency.
AWS Certified Developer - Associate