Summary
Overview
Work History
Skills
Timeline
Generic

CHARLES TALATOTI

Summary

Experienced AI/ML and Data Engineer with 8 years of expertise in developing and deploying machine learning models, building scalable data pipelines, and working with cloud-based solutions. Experienced building end-to-end AI/ML solutions across deep learning, NLP, computer vision, and big data domains. Skilled in implementing end-to-end MLOps pipelines to automate the machine learning lifecycle, ensuring faster deployment, monitoring, and reproducibility. Proficient in Python, SQL, and ML frameworks such as TensorFlow, PyTorch, and Scikit-learn for model development and evaluation. Hands-on experience with data engineering tools like Apache Airflow, Apache Kafka, Databricks, and cloud services including AWS, Azure (Dataflow). Strong understanding of data preprocessing, feature engineering, model training, and real-time inference for use cases like predictive analytics, and customer segmentation. Experienced in CI/CD practices for machine learning and data workflows, enabling continuous integration and delivery of models in production. Passionate about solving real-world problems using AI, continuously learning and exploring new technologies in machine learning, cloud computing, and big data. Integrated MLOps practices into Agile workflows, ensuring model versioning, automated testing, and CI/CD pipelines align with sprint cycles. Expert in developing scalable ML models using TensorFlow, PyTorch, and Python for production-grade applications. Designed and deployed AI systems on cloud platforms like AWS, GCP, and Azure with CI/CD and containerized environments. Built large-scale data pipelines and ETL workflows to support real-time and batch ML model training and inference. Specialized in NLP tasks including text classification, entity recognition, summarization, and LLM-based GenAI solutions. Experienced in scientific and genomic data processing using Python, Spark, and statistical machine learning methods. Proficient in integrating ML services with web applications using Django, Flask, MySQL, REST APIs, and OpenCV. Strong skills in SQL, data warehousing, master data management, and dashboarding via Tableau and Power BI. Proven track record of optimizing algorithms and improving model performance through iterative testing and tuning. Collaborative team player with leadership experience, delivering AI solutions that align with tight deadlines with business impact.

Overview

7
7
years of professional experience

Work History

Senior ML/GenAI Engineer

BJC, St. Louis
01.2024 - Current
  • Developed a PDF Q&A chatbot with asynchronous processing, dynamic prompts, and FAISS-based similarity search.
  • Built a medical diagnostic assistant using LLaMA-3.1-8B with Pinecone for context-aware responses.
  • Improved diagnostic model reliability through rigorous dataset preparation and evaluation on dermatology datasets.
  • Created an identity card fraud detection system using custom image cropping and enhanced preprocessing logic.
  • Achieved 94% accuracy in ID frame detection by optimizing detection pipelines and analyzing performance metrics.
  • Deployed WhisperX and Pynanote-based transcription pipeline with summarization using BART-large.
  • Designed FastAPI and RabbitMQ-based real-time processing backend for scalable meeting transcription.
  • Developed and tested secure RESTful APIs for healthcare data exchange using Spring Boot.
  • Automated deployments with Jenkins and GitLab CI/CD and implemented API tests using Postman and JUnit.
  • Used Kubernetes for scalable deployment and resource management of AI services.
  • Implemented monitoring dashboards using AWS CloudWatch and Power BI for diagnostic and chatbot performance.
  • Reduced patient ID verification time by 40% and improved chatbot response accuracy by 30%.
  • Tech Stack: Python, FastAPI, Llama 3, HuggingFace, WhisperX, Pynanote, BART-large, Spring Boot, PostgreSQL, Pinecone, FAISS, AWS (EC2, Lambda, S3, DynamoDB, CloudWatch), Kubernetes, Jenkins, GitLab CI/CD, Power BI

Machine Learning Engineer

ARTIS Inc
08.2020 - 12.2022
  • Led the design, development, and refinement of machine learning models using diverse algorithms and techniques; built, tested, and deployed models for accuracy and performance.
  • Developed predictive models and implemented machine learning algorithms in Python to generate actionable insights and drive AI-powered decision-making.
  • Collected and processed real-time data using web scraping tools; built a sentiment analysis model leveraging multiple machine learning approaches.
  • Created data visualizations and stakeholder dashboards using Power BI for insights and reporting.
  • Trained custom object detection models using YOLO and developed interactive machine learning web applications with Streamlit.
  • Built a real-time sentiment trend monitoring interface using Streamlit for live data visualization.
  • Designed and optimized Retrieval-Augmented Generation (RAG) workflows to enhance output relevance through contextual data integration.
  • Collaborated in Agile Scrum teams to rapidly develop proof-of-concept (PoC) solutions and deliver iterations efficiently.
  • Researched and evaluated emerging LLM tools, models, and techniques to support PoC development and innovation.
  • Designed and implemented data pipelines for preprocessing raw training data to ensure model quality and readiness.
  • Stayed current with the latest machine learning research, tools, and best practices; continuously suggested improvements to workflows and architectures.
  • Integrated APIs with JupyterLab (Python), Redis/Kafka, and MongoDB to streamline data flow and processing.
  • Supported business development by aligning technical solutions with client requirements and contributing to proposal content.
  • Utilized Linux environments and NVIDIA GPUs to train models with maximum computational efficiency.
  • Managed data collection, annotation, and hyperparameter tuning to enhance model performance.
  • Architected robust data pipelines and developed models using advanced neural network frameworks.
  • Gathered, cleaned, and preprocessed datasets to ensure reliable and high-quality model input.
  • Tech Stack: Python, scikit-learn, PyTorch, TensorFlow, YOLO, Streamlit, Power BI, Kafka, Redis, MongoDB, PostgreSQL, Docker, Linux, CUDA, JupyterLab, FastAPI, Git, Agile/Scrum.

Data Engineer

Technomics Inc, Pune India
09.2018 - 08.2020
  • Engineered complex data engineering solutions on a cloud-based platform using AWS services such as S3, Glue, Spark, Kafka, Hadoop.
  • Designed and built AWS infrastructure components including S3, EC2, ECS, Kinesis, DynamoDB, SNS, SQS, Lambda, Redshift, Athena, and CloudWatch.
  • Collaborated closely with the Business team to implement data strategies, develop data flows, and create conceptual data models.
  • Optimized data ingestion processes, reducing preprocessing time by 30% and improving data quality.
  • Developed and maintained data lakes on AWS S3, integrating data from primary and secondary sources and enhancing flexibility and scalability of data evaluation processes.
  • Engineered data pipelines using Apache Spark, AWS EMR, and AWS Glue for efficient ETL of large-scale marketing data.
  • Leveraged AWS Lambda for serverless data processing pipelines and performed transformations using cloud.
  • Implemented solutions for processing and analyzing streaming data using Kafka and Spark Streaming.
  • Orchestrated complex data workflows using Apache Airflow DAGs to automate data pipeline executions and monitoring.
  • Collaborated with data architects to optimize data warehouses, integrating structured and unstructured data sources like CSV, JSON, and Parquet through ELT processes.
  • Utilized relational databases (MySQL, Oracle DB, Microsoft SQL Server) and NoSQL databases (MongoDB, Cassandra DB, HBase) for data storage and processing.
  • Integrated AWS S3 with Snowflake for data warehousing, supporting business analytics, reporting, and dashboarding.
  • Monitored system performance, troubleshooted issues, and optimized data pipelines to prevent bottlenecks.
  • Implemented version control using Git integrated with AWS Cloud for managing codebase changes.
  • Created interactive dashboards using Tableau, AWS Athena to visualize marketing analytics and derive actionable insights.
  • Tech Stack: AWS (S3, Glue, Spark, EC2, Lambda, Kinesis, Redshift, Athena, CloudWatch), Kafka, Hadoop, MongoDB, Cassandra, MySQL, Jenkins, Tableau, Airflow

Skills

  • Programming Languages: Python, SQL, NoSQL, JavaScript, HTML, CSS, Bash
  • Machine Learning & AI: AI Models, Machine Learning, Deep Learning, PyTorch, TensorFlow, Scikit-learn, LLMs
  • Natural Language Processing: NLP, Text-to-Speech, AI Chatbots, Transformers, Signal Processing
  • Data Engineering: PySpark, Data Preprocessing, Data Handling, Data Analysis, Redis, PostgreSQL, Pandas, NumPy
  • Cloud Platforms: AWS (S3, Lambda, SageMaker), GCP (BigQuery, Dataflow, Vertex AI), Azure (ADF, ML Studio)
  • MLOps & DevOps: CI/CD, Docker, Kubernetes, Model Deployment, MLFlow, Jenkins, Terraform
  • Data Visualization & Analysis: Microsoft Office, Data Analyst tools, Matplotlib, Seaborn, Quality Control
  • Database & Storage: SQL, PostgreSQL, NoSQL, Redis, BigQuery, Data Lakes
  • Security & Compliance: Encryption, Data Masking, RBAC, HIPAA/GDPR compliance
  • Project & Process Management: Jira, Agile/Scrum, Project Management, Training & Development
  • Web Technologies: HTML, CSS, Web Design
  • Automation & Testing: Selenium, PyTest, Unit Testing

Timeline

Senior ML/GenAI Engineer

BJC, St. Louis
01.2024 - Current

Machine Learning Engineer

ARTIS Inc
08.2020 - 12.2022

Data Engineer

Technomics Inc, Pune India
09.2018 - 08.2020
CHARLES TALATOTI