- Designed and deployed Retrieval-Augmented Generation (RAG) pipelines using OpenAI GPT-4, LangChain, and FAISS, delivering real-time insights from structured and unstructured datasets.
- Designed, developed, and deployed end-to-end machine learning and deep learning models using Python, TensorFlow, PyTorch, and XGBoost for classification and clustering problems.
- Fine-tuned LLMs using Hugging Face Transformers and PyTorch, improving model precision and reducing manual response time by 35%.
- Designed and deployed NLP pipelines for text classification, sentiment analysis, and summarization using Hugging Face Transformers, improving text understanding accuracy by 30%
- Automated ETL workflows using PySpark, Pandas, and NumPy, transforming multi-source data for ML model consumption and analytics.
- Designed and trained Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs) using Keras and TensorFlow, achieving >90% accuracy on image and sequence classification tasks.
- Built LLM-powered data retrieval pipelines using LangChain, enabling contextual insights from enterprise data stored in S3 and Redshift via Bedrock API orchestration.
- Built and deployed object detection and image segmentation models using OpenCV, Keras, and TensorFlow, enabling automated defect detection in manufacturing images
- Built scalable and modular ML pipelines leveraging SageMaker Processing, Training, and Model Registry for automated retraining and deployment.
- Built metadata-driven ETL frameworks in Python using configuration tables in Databricks SQL, reducing manual maintenance and enabling reuse across 100+ data sources.
- Collaborated with business stakeholders through Dataiku dashboards and insights, providing explainable model outputs and actionable recommendations.
- Developed modular MLOps pipelines on AWS for model training, validation, deployment, and monitoring using CI/CD workflows, Docker, and SageMaker endpoints.
- Built streamlined MLOps pipelines connecting LangChain-based RAG systems with AWS Bedrock, Glue, and Step Functions for production-ready AI integration.
- Built interactive analytics models and dashboards in Tableau, Power BI, and Databricks SQL, enabling self-service BI and operational insights.
- Collaborated with ML, DevOps, and compliance teams in an Agile environment (JIRA), delivering scalable AI-first solutions within strict enterprise and security standards.
- Optimized model performance through hyperparameter tuning, feature engineering, and GPU acceleration, reducing training time and improving accuracy.
Environment: Python, PySpark, Pandas, NumPy, Delta Lake, Azure Databricks, NLP, Event Hubs, Delta Live Tables, Unity Catalog, XGBoost, Power BI, SQL, Node.js, Express.js, React, Redux, GraphQL, Hugging Face Transformers, PyTorch, LangChain, GPT-4, FAISS, Docker, Kubernetes, Terraform, Jenkins, Azure DevOps, Great Expectations, Boto3, AWS CloudFormation, Celery, Redi, AWS Sagemaker, Keras.