Senior Data Analyst with 5+ years of experience driving data-informed product and business decisions through experimentation, funnel analysis, KPI development, and executive reporting. Proven partner to Product, Engineering, and Design teams, delivering rigorous analytics, scalable dashboards, and clear insights that improve user engagement, conversion, and operational performance. Strong foundation in SQL, Python, experimentation analysis, analytics instrumentation, and data quality, with experience supporting high-growth, consumer-focused platforms.
Overview
6
6
years of professional experience
Work History
Senior Data Analyst
Alcon
Fort Worth, Texas
01.2025 - Current
Collected, integrated, and analysed large-scale pharmaceutical and healthcare supply chain data from internal systems and external platforms using SQL, Python, and RESTful APIs, ensuring end-to-end data availability for strategic decision support.
Built analytical frameworks to measure process optimization, conversion, and bottlenecks, translating insights into actionable recommendations.
Collaborated on data instrumentation and pipeline monitoring, identifying gaps in tracking and improving metric reliability.
Delivered concise, visually compelling insights to senior stakeholders, enabling confident, data-driven product and operational decisions.
Performed data cleansing, validation, and transformation using Pandas, NumPy, R, and SQL, improving the accuracy and reliability of regulatory and compliance reporting across global supply chain operations.
Designed and maintained interactive dashboards and inventory analytics visualizations using Tableau, Power BI, and Excel, enhancing operational transparency and stakeholder communication for distribution and product flow management.
Developed and validated SQL- and Python-based data models, ensuring high data accuracy for reporting and experimentation analysis.
Built scalable ETL and data ingestion pipelines using Apache Airflow, Talend, and AWS Glue, reducing data latency and improving timeliness of insights across pharmaceutical distribution channels.
Developed and deployed predictive analytics and machine learning models using Scikit-learn, TensorFlow, and Keras to forecast drug demand, optimize supply chain networks, and support strategic business planning.
Managed large-scale data lakes and warehousing environments using Amazon Redshift, Snowflake, S3, and Google BigQuery, ensuring high-performance querying and reliable global analytics for supply chain and commercial teams.
Processed structured and unstructured healthcare datasets using Informatica, SSIS, and Talend, and leveraged Apache Spark, Hadoop, and Hive for efficient large-dataset processing and data governance compliance.
Automated data preparation, reconciliation, and transformation workflows using Alteryx and Python, accelerating delivery of actionable insights for logistics, drug distribution, and sales forecasting.
Implemented CI/CD-driven analytics workflows and automated data quality checks using PowerShell, Python, and Jenkins, ensuring consistency and accuracy across global distribution networks.
Built real-time pharmaceutical supply chain monitoring dashboards using Looker and Google Data Studio integrated with BigQuery, and deployed Docker-based containerized analytics workflows to ensure scalability, reproducibility, and seamless collaboration across analytics teams.
TOOLS USED: SQL, Python, APIs, Pandas, NumPy, SAS, Tableau, Power BI, Excel, Amazon Redshift, S3, Google Big Query, Apache Airflow, Talend, AWS Glue, Scikit-learn, TensorFlow, Keras, Talend, Hadoop, HDFS, MapReduce
Data Analyst
Caterpillar Inc.
Irving, Texas
03.2023 - 12.2024
Analysed large volumes of operational, financial, and customer transaction data using SQL, Python, and R, extracting insights from Oracle and SQL Server to support decision-making across product lifecycle, aftermarket services, and financial services business units.
Built interactive and executive dashboards using Tableau, Power BI, and Looker to monitor KPIs such as equipment utilization, asset performance, loan portfolio health, and credit risk, enabling leadership to make data-driven strategic decisions.
Designed and automated scalable ETL pipelines using Talend, Informatica, AWS Glue, and Apache NiFi, integrating data from multiple ERP, telematics, and financial systems and leveraging Apache Spark for high-volume distributed processing and reporting.
Developed and deployed machine learning models using Scikit-learn, TensorFlow, and Keras for fraud detection, credit scoring, demand forecasting, and predictive maintenance, improving operational efficiency and risk mitigation across business units.
Built cloud-based financial and operational analytics pipelines using Amazon Redshift, Snowflake, and S3, incorporating Kafka-enabled real-time event streaming for proactive monitoring and automated alerting through SQL triggers and Jenkins CI/CD workflows.
Implemented modern data governance and metadata management using Collibra and Informatica Data Catalog, and deployed containerized analytics applications using Docker and Kubernetes, ensuring scalable, compliant, and secure data access across global teams.
Performed data cleansing, validation, and transformation using Python (Pandas, NumPy), R, and SQL, and developed scalable financial data models to support regulatory reporting, credit risk assessment, and executive decision-making.
Designed and delivered financial dashboards and reporting solutions using Tableau, Power BI, Looker, and Excel, presenting actionable insights on claims, portfolio risk, revenue trends, and business performance to stakeholders across MUFG business units.
Built and automated ETL & data integration pipelines using Apache Airflow, Talend, AWS Glue, SSIS, and Informatica, consolidating disparate financial datasets and ensuring timely data availability for analysis and compliance reporting.
Performed risk analytics, forecasting, and predictive modeling using SAS, Scikit-learn, TensorFlow, and Keras, enabling data-driven decisions for underwriting, claims management, credit scoring, and fraud monitoring.
Managed large-scale data warehouses and cloud-based analytics ecosystems using Snowflake, Amazon Redshift, S3, Hadoop, and Google BigQuery, ensuring secure, scalable storage, fast querying, and high-performance analytics for enterprise-wide finance teams.
TOOLS USED: Pandas, NumPy, SQL, Python, SAS, Tableau, Power BI, Excel, Amazon Redshift, S3, Google Big Query, ETL, Apache Airflow, Talend, AWS Glue, machine learning, Scikit-learn, TensorFlow, Keras, Informatica
Programmer Analyst
Pidilite Industries Limited
Mumbai
06.2020 - 08.2021
Utilized SQL and Python to extract, analyse, and model high-volume manufacturing, product performance, and customer data from ERP and laboratory systems, supporting data-driven decision-making for product development and production optimization.
Applied advanced statistical analysis in Python and R to evaluate raw material performance, product durability, and defect trends, enabling actionable insights for R&D teams, quality control, and supply chain optimization.
Designed and delivered interactive dashboards and visual analytics using Tableau and Power BI to track production KPIs, product lifecycle performance, market demand trends, and sales distribution, improving business transparency and cross-functional reporting.
Built and automated ETL pipelines and data workflows using Informatica, Alteryx, SSIS, and Apache Airflow, integrating plant-level, lab-level, and commercial datasets while ensuring data quality, integrity, and regulatory compliance.
Leveraged machine learning models with Scikit-learn and TensorFlow and Big Data tools like Hadoop and Spark to predict product demand, identify failure/defect patterns, and improve raw material planning-supporting cost efficiency and business continuity.