Machine Learning Engineer specializing in Generative AI (LLM), NLP, Deep Learning and Computer Vision with expertise in developing machine learning pipelines and scalable micro-services. Skilled in model optimization, RESTful API development, and data analysis. Led projects achieving over 95% accuracy, resulting in significant business savings. Experienced in building AI infrastructure, collaborating on strategic projects, and contributing to applied research. Proficient in computer vision techniques and NLP models, with a background across various Industries and Business such as Retail, Banking and Capital Markets (Finance), Manufacturing, Logistics, and Healthcare. Fluent in Hindi and English, with basic Spanish proficiency.
Years of AI & Machine Learning experience
In my current role at Syndigo LLC, I built robust AI and ML infrastructure from the ground up. My journey has
been marked by impactful achievements:
Scaling Success: I’ve led large-scale ML projects within a tight-knit team of four, demonstrating ownership and
expertise. From inception to deployment, I’ve navigated the entire project lifecycle.
Problem Solver Extraordinaire: My toolkit includes data analysis, model development, ablation studies,
optimization, fine-tuning, RESTful API development and MLOps. I’ve turned complex challenges into elegant
solutions.
Strategic Orchestrator: Collaborating with key stakeholders and leaders, I’ve orchestrated requirements gathering
sessions and brainstorming marathons. My strategic project planning ensures alignment with business goals.
Cutting-Edge Research: The ML landscape is ever evolving, and I thrive on staying ahead. I’ve contributed to
applied research, bringing state-of-the-art solutions to the table—always mindful of practicality and relevance.
Mentorship and Knowledge Sharing: Guiding junior team members is my passion. I’ve hosted year-long companywide ML training sessions, fostering a culture of continuous learning.
LLM, NLP and Computer Vision Champion: I’ve led successful projects in Large Language Models (LLM), natural
language processing (NLP), speech synthesis and computer vision. Transforming data into insights is where I excel.
Seamless Integration: Whether it’s securing models or integrating ML services into existing tech stacks, I ensure
smooth transitions.
Worked on wide variety of target tasks such as classification, summarization, question answering and generation
on textual, audio, image/video data. Conducted and led applied research on various ML techniques/models to solve
the underlying business problems generating significant financial revenue and immense end-user satisfaction.
My experience is resonant but not limited to:
1. Computer Vision:
o Feature-Based Methods:
i. Proficient in SIFT, SURF, and ORB—classic techniques for feature extraction and matching.
ii. Leveraged these methods for tasks such as object recognition, image stitching, and scene
reconstruction.
o Siamese Networks:
i. Designed and trained Siamese neural networks for similarity learning.
ii. Applications include face verification, product verification, and tracking objects across frames.
o CNN (Convolutional Neural Networks): Extensive experience with CNN architectures for:
i. Image classification (e.g., ResNet, VGG, Inception).
ii. Object detection (e.g., YOLO, Faster R-CNN).
iii. Semantic segmentation (e.g., U-Net, DeepLab).
o Large Language Models/ Diffusion:
i. Achieved high quality image in-painting and out-painting using diffusion models like stable
diffusion (SDXL).
ii. Built pipelines to integrate GPT-Vision into products.
2. Audio and Multi-Modal Models:
o Audio Deep Learning:
i. Investigated WaveNet, MelGAN, and Tacotron for speech synthesis.
ii. Applied CRNNs for music genre classification.
o Multi-Modal Fusion:
i. Explored late fusion (concatenation) and early fusion (parallel processing).
ii. Integrated vision and audio features for tasks like scene recognition within videos.
3. Natural Language Processing (NLP):
o RNNs (Recurrent Neural Networks):
i. Developed and fine-tuned LSTMs and GRUs for sequential data modeling.
ii. Applied them to tasks such as sentiment analysis, language modeling, and time series prediction.
o Entity Matching:
i. Devised algorithms for entity resolution across large text corpora.
ii. Utilized techniques like TF-IDF, Jaccard similarity, and Levenshtein distance.
o Textual Entailment:
i. Constructed models to determine logical entailment between sentences.
ii. Incorporated attention mechanisms and word embeddings for context understanding.
o BERT Models:
i. Fine-tuned pre-trained BERT models for specific NLP tasks.
ii. Achieved state-of-the-art results in sentiment analysis and question answering.
o LLMs (Large Language Models)/Transformers:
i. Explored transformer architectures (e.g., GPT, T5) for various NLP applications.
ii. Implemented self-attention and multi-head attention layers.
4. ML Techniques:
o Quantization and Pruning: Optimized neural networks during deployment by reducing weight precision
(quantization) using LORA and QLORA and removing unnecessary connections (pruning) using PEFT.
Achieved a delicate balance between accuracy and efficiency.
o Neural Architecture Search (NAS): Achieved a delicate balance between accuracy and efficiency using
Vertex AI (Google). This required a lot of computation resources and budget.
o Core Frameworks: Proficient in both PyTorch and TensorFlow frameworks. Designed custom layers,
implemented complex loss functions, and appreciated their unique features.
o Hardware Optimization: Explored parallelism, memory hierarchy, and energy efficiency trade-offs
amongst various CPUs and GPUs, deploying models on relevant hardware. LPU’s and NPU’s are popular
alternatives but our requirements did not have a necessity on that hardware.
o Hyperparameter Tuning: Fine-tuned models using grid search, random search, and Bayesian
optimization. Strived for optimal accuracy-efficiency trade-offs.
o Knowledge Distillation: Compressed large models while preserving performance. Explored techniques
like temperature scaling and attention transfer to achieve compressed student neural networks.
5. Production grade software systems for AI:
o End-to-End Pipelines:
i. Architected robust pipelines handling data prep, training, and deployment.
ii. Ensured seamless integration into micro-services with version control, monitoring, logging and
securing.
iii. Utilized Azure Kubernetes clusters to create the production host environment.
o RESTful APIs:
i. Designed RESTful APIs enabling CRUD operations using Python, FastAPI and flask.
ii. Generated requirements and guidance for software developers on MLOps and DevOps.
Python
Building Serverless Applications (Coursera), 09/2021, 8JDNLAJ66XH4
Hiking
Fitness
Guitar
E-Sports
Reading
Building Serverless Applications (Coursera), 09/2021, 8JDNLAJ66XH4