
"From distilling 70B LLMs on 24 H200 GPUs to scaling teams of 20+ engineers, I build AI systems that ship. Pending LLM patent, 4x inference throughput gains, 15+ production apps, and platforms handling 10,000+ TPS."
10+
Years Experience
20+
Engineers Led
15+
Apps Shipped
1
Pending Patent
Blackbelt in Zen Shotokai Karate
Systems and Methods for Generating an Improved Mixture of Attention Model
QWERKY AI Inc. · 2025
Novel LLM architecture with CUDA implementation

Georgia Institute of Technology
MS Computer Science
Expected May 2028

University of South Carolina
BS Computer Science
Summa Cum Laude
GPA: 3.956 / 4.0
2025 - 2026
2024 - 2025
2022 - 2024
2022 - 2024
2022
2019 - 2022
2015 - 2019
February 2026
Implementing Mamba 1 architecture into Modular's MAX inference framework with custom CPU-only Selective Scan and causal convolution kernels for cross-vendor SSM support.
September 2025
A critical examination of Geoffrey Hinton's "mother AI" proposal, advocating for human-centered AI development that prioritizes sustainability and efficiency over speculative superintelligence concerns.
June 2025
Analysis of Meta's Ring Attention and DeepSeek-V3's Multi-head Latent Attention, exploring how fundamental memory and hardware constraints continue to limit LLM architectures.
May 2025
Exploring hidden sources of unpredictability in LLMs including floating-point arithmetic errors, parallel processing inconsistencies, and hardware variations affecting AI reproducibility.