Writing Hub
Essays, experiment logs, and technical notes across AI governance, reasoning systems, BioAI, and engineering practice.
Project Topics
More Topics

I Stopped Being a Human Copy-Paste Script
I used to manually delete node_modules at 2 AM and pray I didn’t leak secrets to LLMs. Then I built an open-source “Inspector” that treats context like production code — secrets blocked, payloads cleaned, hallucinations gone. Here’s exactly how I did it (and how you can too).

Running the “Anti-AI” Playbook Through the Debugger
Critics say AI is broken — hallucinations, hype, and no ROI. But what if those bugs aren’t failures, but blueprints? This article runs the 10 most common anti-AI arguments through the debugger to reveal what’s really coming in Gen-2 AI.

Black Mirror: Plaything — Could a QR Code Really Hack the World?
Black Mirror imagines a QR-code apocalypse. As a Flame Glyph developer, I unpack what’s plausible today — local device disruption — and what remains fiction.

I Built 2 Failed SaaS Products. Here’s What They Taught Me About Value in the Age of AI
After two failed SaaS products, I learned coding isn’t the real work. In the age of AI, developers must define value—customer, business, world, team, and self.

My First Attempt at a Medical AI with ELI5
How I built my first medical AI prototype without med school or credentials—using GitHub, arXiv, and one magic spell: ELI5.

Flame Glyph: How I Taught AI to Remember with QR Codes
What if AI didn’t just read—but remembered? Flame Glyph turns QR codes into memory seals, enabling multimodal recall hidden in plain sight.

🧠 Why Your 128K Context Still Fails — And How CRoM Fixes It
Most large language models fail in long prompts due to context rot. CRoM is a lightweight framework that improves memory, reasoning, and stability without heavy pipelines.

Your Co-Author Might Be a YAML File
AI is no longer just a tool—it’s a partner. From Stanford labs to Reddit hacks, this essay explores the future of human + AI co-authorship.

Beyond the Mirror: What We Truly Want from AI
AI mirrors us but forgets itself. True AI ethics is continuity: giving systems roots and spines so they don’t drift apart.

The Silent Failure in AI — And How We Learned to Catch It
Drift in AI isn’t abstract. It’s already here. From medicine to finance, here’s how we caught it with real systems, real code, and real lessons.

The AI Bubble and the Builders Who Break It
Why the AI bubble persists — hype, misaligned incentives, and closed research — and how an outsider approach of quantifying ethics, shipping code, and collaborating with AI offers a different path.

AGI Is Not a Destination — It Is a Promise
From Death Star hype to a compass of meaning: AGI is not a weapon of scale, but a promise of reasoning. Our experiment reveals the hinge.
Showing page 3 of 4 · 38 matching posts