Summary
Overview
Work History
Education
Skills
Websites
Projects
Timeline
Generic

Vinith Kumar Guntupalli

Irving,TX

Summary

Overall, 5+ years of experience in Data Science and Big Data in the Business Industry with adept knowledge of Data Analytics, Machine Learning (ML), and Predictive Modeling. Machine Learning and Computer Vision expert with experience in image processing and analysis. Proven track record in developing and optimizing computer vision algorithms and ML models. Skilled in refining ML models to improve accuracy and collaborating with cross-functional teams to integrate AI/ML features into applications. Specialized in Text Analytics, developing Statistical Machine Learning and Data Mining solutions using R, Python, and Tableau. Proficient in Data Analytics, including Data Reporting, Ad-hoc Reporting, and OLAP reporting. Strong command over SQL, with experience in RDBMS (e.g., MySQL) and NoSQL databases (e.g., MongoDB). Skilled in using Spark MLlib utilities for machine learning tasks and various visualization tools like Tableau, SAS, QlikView, and Microsoft BI. Expertise in operations research techniques, mathematical programming, heuristic algorithms, and stochastic modeling. Hands-on experience in designing optimization models using Python and relevant packages such as PuLP, PYOMO, and CVXPY, with proficiency in CPLEX. Well-versed in TR components and NESS requirements into scalable analytical models, working with Hadoop ecosystem components, and designing captivating visualizations using Tableau. Knowledgeable in various machine learning algorithms and techniques, including Reinforcement Learning, LDA, Naive Bayes, Random Forests, Decision Trees, Linear and Logistic Regression, SVM, Clustering, and neural networks.

Overview

7
7
years of professional experience

Work History

AI/ML Engineer

Abbott
IL, USA
01.2023 - 09.2024
  • Developed and optimized Computer Vision algorithms for medical imaging applications
  • Improved the accuracy of ML models through continuous refinement and testing
  • Collaborated with cross-functional teams to integrate AI/ML features into healthcare applications
  • Implemented image processing techniques to enhance the quality of medical images
  • Presented technical findings and project updates to stakeholders
  • Designed and implemented data pipelines for financial data processing and analysis using Python and SQL
  • Conducted data integration from diverse financial systems including SAP, Oracle Financials, and QuickBooks to ensure comprehensive financial data consolidation
  • Developed ETL processes to extract, transform, and load financial data from various sources into a centralized data warehouse on Azure
  • Utilized Python and SQL for financial data parsing, cleansing, and validation to maintain data integrity
  • Performed comprehensive investigations and analysis, leveraging data analytics tools to detect patterns and trends
  • Created and deployed automated solutions to optimize risk assessment workflows, leading to a 20% improvement in operational efficiency
  • Implemented automation processes for Continuous Improvement/Continuous Deployment pipelines, optimizing workflows
  • Designed and implemented optimization models using Python and relevant optimization packages, including PuLP, PYOMO, CVXPY, and CPLEX
  • Conducted thorough analysis of optimization results to ensure accuracy and relevance, providing actionable insights for business improvement
  • Documented findings, solutions, and methodologies, and presented them to both technical and non-technical audiences
  • Built predictive models using machine learning algorithms to forecast financial trends and performance metrics
  • Created real-time dashboards and visualizations in Tableau to monitor key financial indicators and support decision-making
  • Designed and maintained financial data models to support business intelligence initiatives using Azure Analysis Services
  • Conducted data quality assessments and implemented data validation scripts to ensure the accuracy and reliability of financial data
  • Extensively used Cloudera Hadoop distributions within the project
  • Ingested data from various source systems like Teradata, MySQL, Oracle databases
  • Handled data in different file formats like Avro and Parquet
  • Worked on creating tabular models on Azure analysis services for meeting business reporting requirements
  • Created procedures in Azure SQL Datawarehouse and built the final aggregate tables for dashboards.

Data Engineer/ML Engineer

EPAM Systems
Hyderabad, India
09.2020 - 12.2021
  • Led the development of advanced image analysis algorithms for various client projects
  • Enhanced existing ML models, resulting in a 15% increase in accuracy
  • Worked closely with software development teams to integrate Computer Vision solutions into applications
  • Conducted research on the latest trends in Computer Vision and ML to inform project strategies
  • Mentored junior engineers on best practices in ML and Computer Vision
  • Analyze the data by performing Extraction, Transformation & Load function mentioned by clients to process and get accurate feedback that includes cause for warnings & failures
  • Ability to explain the cause for the loss of records when the data moves from import table to Transformation table
  • Modify the instance and purge (i.e., Clearing the data from the database) the data required by the specific client to get the clean validation pass for the files submitted
  • Developed and maintained Conceptual, Logical, and Physical Data Models using Erwin
  • Collaborated within agile scrum teams, employing tools like SSIS, Azure Data Factory, Spark Scala, and Hadoop
  • Applied mathematical programming techniques to formulate and solve optimization problems, contributing to improved business processes
  • Utilized heuristic and metaheuristic algorithms to address complex optimization challenges
  • Collaborated with various stakeholders to understand and address business needs through optimization modeling
  • Developed and implemented optimization models in Python, leveraging PuLP, PYOMO, CVXPY, and other relevant packages
  • Designed algorithms for incident alert extraction, reducing redundancy and improving efficiency
  • Utilized Python Notebooks in Data Bricks Workspace for data transformation requirements
  • Implemented mathematical models and optimization cores using tools like MSW DST and Sage maker
  • Processed the spend and goals data in Alteryx in such a way that it is suitable for reporting
  • Performed data imputation using Scikit-learn package in Python
  • Visualizing the data with the help of box plots and scatter plots to understand the distribution of data using Tableau, Python libraries
  • Involved in exhaustive documentation for the technical phase of the project and training materials for all data management functions
  • Conducted major stakeholder interviews involving SME's, Business Analysts, and other stakeholders
  • Applied concepts of R-squared, R.M.S.E, P-value, in the evaluation stage to extract interesting findings through comparisons
  • Perform exploratory data analysis, statistical analysis and build visualization report on ultra-learn data to understand instructor behavior patterns on grading
  • Define scope of projects based on gathered Business Requirements including documentation of constraints, assumptions, business impacts & project risks
  • Strong background in support documentation
  • Analysis and review of Business Requirement Documents.

Data Engineer

Virtusa
Hyderabad, India
01.2019 - 08.2020
  • Designed and implemented image processing algorithms for real-time applications
  • Collaborated with data scientists to develop and train ML models for object detection and recognition
  • Conducted extensive testing and validation of models to ensure high performance
  • Assisted in the deployment of AI/ML features into production environments
  • Provided technical support and troubleshooting for Computer Vision projects
  • Automated Python scripts to organize raw data from multiple Json, Excel files
  • Performed Data Profiling and implemented Data Quality checks using SSIS, SQL, and EXCEL
  • Involved in development of SSIS ETL packages
  • Performed Data Analysis and Data Profiling and worked on data transformations and data quality rules
  • Used Python and R to perform Exploratory Data Analysis (EDA) and used machine learning algorithms like K-means clustering and Random Forest as per the business requirement
  • Using Random Forest with Variable Importance we identified which variables are playing the key role in our product sales and demand
  • To better the performance of our model, we use Grid Search CV and Random Search CV for best parameters
  • Worked closely with the different teams to discuss various ideas/solutions, issues, and timelines
  • Identified and implemented innovative ideas to yield cost/effort savings
  • Taken initiative in evaluating and adapting new approaches from data science research
  • Built customized interactive reports and dashboards using Tableau
  • Mentoring the team in various tasks, ranging from development to root cause analysis
  • Analysis of functional and non-functional categorized data elements for data profiling and mapping from source to target data environment
  • Developed working documents to support findings and assign specific tasks
  • Involved with data profiling for multiple sources and answered complex business questions by providing data to business users
  • Worked with data investigation, discovery, and mapping tools to scan every single data record from many sources.

Data Analyst

Mouri Tech
Hyderabad, India
01.2018 - 12.2018
  • Worked with Data Analyst for requirements gathering business analysis and project coordination
  • Generated SQL scripts to extract relevant data and developed SSIS packages for data migration
  • Performed the Data Accuracy, Data Analysis, Data Quality checks before and after loading the data
  • Written SQL Scripts and PL/SQL Scripts to extract data from Database and for Testing Purposes
  • Created T/SQL statements (select, insert, update, delete) and stored procedures
  • Designed and developed ETL workflows and datasets in Alteryx
  • Used Alteryx for Data Preparation and then Tableau for Visualization and Reporting
  • Generated Data dictionary reports for publishing on the internal site and giving access to different users
  • Produced report using SQL Server Reporting Services (SSRS) and creating various types of reports
  • Worked on Data Mining and data validation to ensure the accuracy of the data between the warehouse and source systems
  • Worked in generating and documenting Metadata while designing OLTP and OLAP systems environment
  • Worked with data investigation, discovery, and mapping tools to scan every single data record from many sources
  • Assisted Reporting developers in building Reports using Crystal Reports
  • Performed ETL data management by developing code scripts with statistical programs SQL, Tableau
  • Performed Data Validation / Data Reconciliation between disparate source and target systems for various projects
  • Involved working on different set of layers of Business Intelligence Infrastructure
  • Created financial models using Excel Pivot Tables and formulas to develop best- and worst-case scenarios
  • Performed data mining on Claim’s data using very complex SQL queries and discovered claims pattern
  • Created DML code and statements for underlying & impacting databases
  • Extensively used ETL methodology for supporting data extraction, transformations and loading processing, in a complex EDW using Informatica
  • Perform data reconciliation between integrated systems
  • Metrics reporting, data mining and trends in helpdesk environment using Access
  • Written complex SQL queries for validating the data against different kinds of reports generated by Business Objects XIR2.

Education

Master of Science - Computer Science

Campbellsville University
Campbellsville, KY
10-2023

Bachelor of Science - Information Technology

Vasireddy Venkatadri Institute of Technology
India
05-2019

Skills

  • Python
  • SQL
  • PL/SQL
  • PySpark
  • Scala
  • Shell Scripting
  • MATLAB
  • Oracle
  • PostgreSQL
  • MySQL
  • SQL-Server
  • Teradata
  • TensorFlow
  • Keras
  • OpenCV
  • Scikit-Learn
  • Cassandra
  • MongoDB
  • MariaDB
  • HBase
  • Hadoop
  • HDFS
  • MapReduce
  • Pig
  • Sqoop
  • Spark
  • Impala
  • Cloudera
  • Hortonworks HDP
  • Spark SQL
  • Kafka
  • Spark-Streaming
  • Flink
  • Ambari
  • Apache Airflow
  • Dagster
  • Pentaho Data Integration
  • AWS Glue
  • Informatica power center
  • Azure Data Factory
  • Datadog
  • Grafana
  • ELK stack
  • Snowflake
  • Redshift
  • Microsoft Azure synapse analytics
  • Hive
  • DBT
  • Alteryx
  • Microsoft SSIS
  • SSAS
  • SSRS
  • Microsoft Power BI
  • Tableau
  • QlikView
  • Informatica 61
  • Visual Studio
  • PyCharm
  • Jupyter Notebook
  • Microsoft Azure – Azure Databricks
  • Data Lake
  • Blob Storage
  • SQL Database
  • Azure Synapse Analytics
  • Cosmos DB
  • Active Directory
  • Amazon AWS - EMR
  • EC2
  • EBS
  • RDS
  • S3
  • Athena
  • Glue
  • Elasticsearch
  • Lambda
  • SQS
  • DynamoDB
  • Kinesis

Projects

Medical Image Analysis: Developed algorithms to enhance and analyze MRI and CT scan images. Object Detection System: Created a real-time object detection system using YOLO and TensorFlow. Image Segmentation Tool: Designed an image segmentation tool for satellite imagery using U-Net architecture.

Timeline

AI/ML Engineer

Abbott
01.2023 - 09.2024

Data Engineer/ML Engineer

EPAM Systems
09.2020 - 12.2021

Data Engineer

Virtusa
01.2019 - 08.2020

Data Analyst

Mouri Tech
01.2018 - 12.2018

Master of Science - Computer Science

Campbellsville University

Bachelor of Science - Information Technology

Vasireddy Venkatadri Institute of Technology
Vinith Kumar Guntupalli