Self-Improving Models Without Labels: What I Just Proved and Why It Matters
A 7B model taught itself to generate better security commands using only its own understanding signals. No human labels, no external reward. Here's how and why it matters.
Homelab adventures, infrastructure deep-dives, and lessons learned building enterprise-grade systems on a budget
A 7B model taught itself to generate better security commands using only its own understanding signals. No human labels, no external reward. Here's how and why it matters.
A novel discovery: hidden state inversion quality predicts model capability, enabling self-improving systems without external feedback
A deep dive into Mojo, the MLIR-based language promising 35,000x speedups over Python. What it does well, where it falls short, and who should actually care.
How periodic sparse attention achieves O(n) complexity while maintaining model quality
An overview of my ongoing research exploring transformer injectivity, self-improving models, and high-performance AI on an RTX 3090
My RQ setup works, but Temporal's workflow orchestration promises better handling of complex multi-step jobs. Here's my migration plan and the real tradeoffs.