
"From a pending LLM patent and CUDA kernels to scaling teams of 20+ engineers, I build AI systems that ship โ 4x model throughput gains, 15+ production apps, and platforms handling 10,000+ TPS."
10+
Years Experience
20+
Engineers Led
15+
Apps Shipped
1
Pending Patent
Blackbelt in Zen Shotokai Karate
Systems and Methods for Generating an Improved Mixture of Attention Model
QWERKY AI Inc. ยท 2025
Novel LLM architecture with CUDA implementation
Georgia Institute of Technology
MS Computer Science
2025 - Present
University of South Carolina
BS Computer Science
Summa Cum Laude
2025 - Present
2024 - Present
2022 - 2024
2022 - 2024
2021 - 2022
2015 - 2022
February 2026
Implementing Mamba 1 architecture into Modular's MAX inference framework with custom CPU-only Selective Scan and causal convolution kernels for cross-vendor SSM support.
September 2025
A critical examination of Geoffrey Hinton's "mother AI" proposal, advocating for human-centered AI development that prioritizes sustainability and efficiency over speculative superintelligence concerns.
June 2025
Analysis of Meta's Ring Attention and DeepSeek-V3's Multi-head Latent Attention, exploring how fundamental memory and hardware constraints continue to limit LLM architectures.
May 2025
Exploring hidden sources of unpredictability in LLMs including floating-point arithmetic errors, parallel processing inconsistencies, and hardware variations affecting AI reproducibility.