Possessing analytical and problem-solving skills, coupled with positive attitude. Well-versed in statistical analysis and data visualization, mastering Python and SQL. Capable of transforming data insights into actionable business strategies.
Overview
2
2
years of professional experience
Work History
Robotic Process Automation Developer
Booz Allen Hamilton, Inc.
06.2024 - Current
Worked closely with the Department of Veteran's Affairs to create and maintain automations using UiPath
Communicated proactively with stakeholders at every stage of project life cycles to keep them informed about progress or any potential delays
Utilized Python to clean, transform, and combine large text-based datasets to find new key-phrases for an existing automation
Identified 7 new key-phrases estimated to recapture a total of 580 working hours per year using Pandas, NLTK, and Rake
Created dashboards in Power BI and Splunk that displays a summary of our automations' performance for leadership, stakeholders, and the individual VAMCs
Created tool-tips in Power BI that breaks down certain visualizations, explains how certain statistics were calculated, and gives definitions for measures of central tendency
Utilized Agile development strategies including ceremonies
Data Scientist
Booz Allen Hamilton, Inc.
03.2024 - 06.2024
Partook in Booz Allen Hamilton's Technical Excellence Advanced Data Science 8-week cohort designed for machine learning and AI development
Gained an understanding of the importance in data governance and management as well as the role they play in AI development
Performed data visualization and exploratory data analysis using packages in Python such as Seaborn, Matplotlib, SciPy, and Pandas
Used deep learning frameworks such as Tensorflow and PyTorch
Practiced implementing various predictive machine learning models on real-world datasets such as K Nearest Neighbors, Logistic Regression, Support Vector Machines, and K Means
Undergraduate Researcher
Purdue University - The Data Mine
08.2023 - 12.2023
Collaborated with DriveOhio, a smart mobility company, in analyzing data collected from their autonomous vehicles to identify problems and reasonings behind their autopilot disengagements
Utilized Amazon Web Services (AWS), DynamoDB, NoSQL, and Python to access, clean, and transform the raw data from the autonomous vehicles
Acquired and plotted autopilot disengagement points on the routes the autonomous vehicles drive on using GitHub and Python (matplotlib)
Accessed National Oceanic and Atmospheric Administration (NOAA) weather data using their API endpoint for the routes' location to see if there was a correlation between autopilot disengagements and weather
Utilized Docker to display a three-dimensional video based off data recorded from an autonomous vehicle's Light Detection and Ranging (LiDAR) sensors giving a better understanding of how the Apollo system equipped to the vehicles identifies and maneuvers obstacles around it
Education
B.S. - Mathematics
Youngstown State University
Youngstown, OH
Skills
Machine Learning
Natural Language Processing
Statistics
Deep Learning
Artificial Intelligence
Data Analysis
Data Visualization
Power BI
Python
SQL
GitHub
Docker
Seaborn
Pandas
Scikit-Learn
TensorFlow
PyTorch
Clearance Level
Secret
Certifications Specialized Training
Youngstown State University Data Analytics Certification, 06/23
Technical Excellence Advanced Data Science Program, 04/24
Agentic RAG for Magic: The Gathering: Implemented the Crawl4AI web crawler library in Python for lightspeed web scraping; configured the crawler to look for specific material in the HTML code; prompt engineered gpt-4o to output a summary of the scraped material; chunked the summary into categories and stored it in a vector database; created a front-end for users to interface with using Streamlit; checks Supabase database for entries that best match the user's question and feeds the information into gpt-4o for summarization; provides answers to questions about rules, interactions, formats, and even deck building
Generative AI for Text Classification: Fine-tuned an NLP model, BERT, to summarize and categorize a given article into one of four different categories; imported the CNN/DailyMail dataset to train the model on; created functions to remove stop words, remove special characters, and normalize the text within the article; used the Tensorflow library to build the machine learning model framework, initialize the tokenizer, and to pad sequences up to the max token length; used the Transformers library to load a pre-trained BERT model and fine-tuned it on articles of varying lengths padding articles that were less than the max token count for the model and feeding it subsections of an article if they were larger than the max token counts; created Docker Files, Docker containers, a FastAPI endpoint, and a virtual environment for efficient redeployment of the fine-tuned model
Recognizing Handwritten Digits Using AI: Created a machine learning model that accurately recognizes handwritten digits between 0 and 9 with 98% accuracy, conducted data augmentation to aid the model in finding patterns, utilized convolution neural networks and dense layers to train our initial model using Keras, conducted model fine-tuning using Hyperopt and MLFlow
Predicting Customer Churn: Fit and evaluated multiple machine learning models to predict customer churn including Logistic Regression, Support Vector Machines, K Nearest Neighbors, and Random Forest; evaluated each model using a ROC AUC score to determine which model was the most accurate for our data set; parameter-tuned our Random Forest model using GridSearchCV; found relationships between features and target using graphs and central tendency measures