Mitch Alley

Software Engineer

Extensive experience in system design, data science, and building AI models.

Contact Me

Introduction

about

AI and data science professional with a strong software engineering background, currently pursuing a Master’s in AI at UT Austin. Experienced in leading projects and implementing ML solutions.

03+ Years
experience
20+ Completed
projects
04+ Companies
worked
(Including internships)

Technical Skills

Software Engineer

More than 2 years
Day-to-day stack

PYTHON

35%

SQL

25%

PYTHON LIBRARIES (PLT, TORCH, ETC.)

20%

JAVA

10%

HTML/CSS/JAVASCRIPT

10%

Data Analytics Intern

3 Months (Summer)
Day-to-day stack

SQL

50%

POWER BI

25%

PYTHON

15%

XGBOOST

10%

Data Analytics Intern

2 Months (Summer)
Day-to-day stack

PYTHON - PANDAS

45%

MATPLOTLIB

35%

XGBOOST

20%

Software Engineer

3 Months (Summer Internship)
Day-to-day stack

XCODE/SWIFT

80%

Python

10%

SQL

10%

Qualifications

Education
Work

M.S. in Artificial Intelligence

University of Texas at Austin
2024 - 2025

Large Language Model Customization

Nvidia DLI
2024

B.S. in Computer Science

University of Wisconsin-Madison
2018 - 2022

B.S. in Economics with Math Emphasis

University of Wisconsin-Madison
2018 - 2022

Software Engineer

JLG Industries
Jun 2022 - Present

Data Analytics Intern

Polaris Industries
May 2021 - Aug 2021

Data Analytics Intern

University of Wisconsin-Madison Economics Dept.
Jun 2020 - Aug 2020

Software Engineer

University of Wisconsin-Parkside App Factory Intern
Jun 2018 - Aug 2018

AI Projects

Mitigating Errors in NLI with ELECTRA-Small

View More

Mitigating Errors in NLI with ELECTRA-Small

  • Investigated the impact of spurious correlations and dataset artifacts on pre-trained models in NLI

  • Used contrast sets, checklist sets, and adversarial challenge sets to assess model behavior and identify weaknesses in performance on challenging examples.

  • Implemented methods like adversarial data augmentation to enhance model generalization and reduce bias from spurious correlations.

  • Conducted in-depth analysis to identify specific error types and implemented focused training strategies to address performance issues in targeted subsets of data.

  • Evaluated model performance with accuracy metrics and visualizations to understand the effectiveness of fixes and identify improvements in generalization to real-world scenarios.

Fact-checking Outputs from ChatGPT

View More

Fact-checking Outputs from ChatGPT

  • Analyzed and verified outputs from large language models (LLMs) like ChatGPT by conducting fact-checking against Wikipedia through two methods.

  • Method 1: Implemented a word overlap method for fact-checking, using metrics like cosine similarity and Jaccard similarity to predict supported vs. not supported facts.

  • Method 2: Applied a pre-trained DeBERTa-v3 model for textual entailment to classify facts as supported or not supported based on premise-hypothesis pairs.

  • Optimized the entailment model with pruning based on word overlap to improve performance and meet accuracy requirements.

  • Conducted error analysis on the entailment model's predictions, categorizing false positives and false negatives into fine-grained error types for further evaluation.

Transformer Language Modeling

View More

Transformer Language Modeling

  • Implemented a Transformer language model to predict character occurrences and next characters in sequences using PyTorch.

  • Built a simplified Transformer encoder with self-attention and positional encodings to predict character occurrences (before or before-and-after) in sequences.

  • Extended the model to predict the next character in a sequence, achieving a perplexity ≤ 7 and training within 10 minutes.

  • Evaluated performance on perplexity and log-likelihood, requiring the model to output valid probability distributions for predictions.

Feedforward Neural Networks, Word
Embeddings, and Generalization

View More

Feedforward Neural Networks, Word Embeddings, and Generalization

  • Implemented optimization with SGD for a quadratic function and explore step size tuning.

  • Built a Deep Averaging Network (DAN) using pre-trained word embeddings and fine-tune model parameters.

  • Implemented batching in the neural network and handle padding for sequences of varying lengths.

  • Enhanced model generalization by addressing misspelled words with spelling correction or prefix embeddings.

Sentiment Classification with Perceptron
and Logistic Regression

View More

Sentiment Classification with Perceptron and Logistic Regression

  • Implemented perceptron and logistic regression classifiers for binary sentiment classification using unigram features.

  • Explored feature extraction techniques like unigrams, bigrams, and custom features to improve model performance.

  • Experimented with different learning schedules and feature modifications to optimize accuracy.

Neural Networks
for Autonomous Driving

View More

Neural Networks for Autonomous Driving

  • Designed machine learning models, including MLP, Transformers, and CNNs, to predict vehicle trajectories for autonomous driving tasks.

  • Utilized lane boundary and waypoint data to train models for trajectory prediction, with metrics focusing on longitudinal and lateral error.

  • Implemented a Transformer-based planner leveraging cross-attention and learned query embeddings for improved trajectory predictions.

  • Built a CNN planner to predict lane boundaries directly from images, bypassing segmentation and depth estimation challenges.

Convolutional Networks for
Vision Tasks

View More

Convolutional Networks for Vision Tasks

  • Developed convolutional networks for image classification, segmentation, and detection tasks.

  • Implemented models to classify racing track images and predict road segmentation masks with depth estimation.

  • Built a training pipeline from scratch, including data augmentation, optimization, and model evaluation using metrics like mIoU and MAE.

  • Experimented with advanced techniques like residual connections, skip connections, and multi-task loss functions.

SuperTuxKart Image Classifier

View More

SuperTuxKart Image Classifier

  • Trained neural networks to classify SuperTuxKart images using linear, MLP, and residual models.

  • Built data loaders, loss functions, and models with PyTorch.

  • Used TensorBoard for tracking training and validation metrics.

  • Accelerated training with GPU resources for efficiency.

  • Ensured robust evaluation with train, validation, and test splits.

Testimonials

Cesar Hernandez

Coworker

"I worked with Mitch for a year at JLG as a software engineer, managing the the Shippensburg and McConnellsburg production facility's software. Mitch was awesome to work with, helping me whenever I needed help with SQL or any front-end related framework questions that came up. Mitch is hardworking and never one to shy from a task, always being present at the Shippensburg facility to help with any software related issues that arose during day-to-day operation. It was great working with such a dependable, capable, and technical person during my year at JLG as a software engineer."

Contact Information

Phone

(262)-484-8551

Email

mitch1.alley@gmail.com

Location

Greater D.C. Area