My proficiency revolves around developing and implementing robust ETL/ELT data pipelines. My expertise lies in seamlessly integrating diverse datasets encompassing Financial and healthcare data. My approach ensures that data is collected and meticulously transformed into actionable insights.Transitioning from data-centric environment with focus on developing efficient data solutions and optimizing workflows. Skilled in data architecture, database management, SQL, and Python, with track record of enhancing data-driven decision-making processes.
• Designed, developed, and overhauled a real-time lead prioritization predictive model to process data of existing customers and new prospects using dbt, snowpark
• Led end-to-end architecture on modern ELT data stack on dbt core and snowflake
• Reduced manual and hands-on tasks by 40+ hours per request by creating a tool that obtained data and created documentation using Python, GraphQL, REST APIs, and AWS Lambda
• Migrated and refactored complex Alteryx data pipelines using dbt, creating a more efficient approach to reverse ETL Salesforce data leading to a significant reduction in tech debt and enabling the team to build relevant and actionable KPIs and metrics
• Developed incremental and snapshot dbt models in snowflake, and orchestrated CI/CD data pipelines, which reduced the ELT run time from +12 hours to less than 20 minutes
• Engineered a python automation script that enabled the internal analytics team to reduce maintenance tasks on the Tableau server by 75%
• Reengineered existing ETL workflows to improve performance by identifying bottlenecks and optimizing code accordingly.