Building Cutting-Edge AI Research on Consumer Hardware with Mojo
An overview of my ongoing research exploring transformer injectivity, self-improving models, and high-performance AI on an RTX 3090
Discover insights, tutorials, and thoughts on technology, homelab, and development.
An overview of my ongoing research exploring transformer injectivity, self-improving models, and high-performance AI on an RTX 3090
How periodic sparse attention achieves O(n) complexity while maintaining model quality
A novel discovery: hidden state inversion quality predicts model capability, enabling self-improving systems without external feedback
How I achieved 100% token recovery from Mistral-7B hidden states and what it means for AI security
Responsible disclosure of a class of vulnerabilities that allow system prompt extraction from transformer hidden states