Writing Hub
AI governance essays, reasoning systems notes, experiment logs, and technical writing across BioAI and engineering practice.
Project Topics

The Sheepwave Has a New Shape: OpenMythos and the Rise of Architecture Hype
A technical-opinion essay on OpenMythos, Claude Mythos, README-driven AI hype, and why architecture claims need source-level verification before becoming public belief.

OpenMythos v0.5.0 Code Review - Audit Report
OpenMythos collected thousands of GitHub stars and dominated AI discourse for a week. This is what happens when you actually read the code — and why the people who do always arrive too late to matter.

FLAMEHAVEN FileSearch: Why This RAG Engine Feels Different from the Usual Stack
A technical look at FLAMEHAVEN FileSearch: BM25+RRF hybrid retrieval, chunk-addressable indexing, deterministic DSP vectors, and the trade-offs behind a lower-overhead self-hosted RAG engine.

The Next AI Moat May Not Be the Harness Alone: A Mathematically Governed Self-Calibrating Code-Review Layer
As AI harness patterns normalize, differentiation is shifting toward governed self-calibration and implementation fidelity. This piece explores how history-driven, bounded adaptation creates a new layer of defensible AI infrastructure — one that turns local code evolution into a competitive moat.

It Gets Smarter Every Scan: AI-SLOP Detector v3.5.0 and the Self-Calibration Loop
AI-built apps are starting to fail in public. Not every failure is static-analysis territory, but many share the same upstream condition: plausible-looking code passing review without carrying enough real logic. AI-SLOP Detector v3.5.0 adds a self-calibration loop to reduce that gap.

Prompt → RAG → MCP → Agent → Harness, and What?
Why the next layer in AI may be governance infrastructure, not just better agents.

The Harness Is the Product: What the Claude Code Leak Actually Revealed About AI Agent Architecture
The Claude Code leak exposed more than source. It revealed that modern AI agent performance depends heavily on the harness around the model.

95% of AI Businesses Will Die. Here’s How to Not Be One of Them.
What the data, a founder’s confession, and 70 years of tech history tell us about who actually survives.

Prompt, Pray & Push: Why Your AI Agent Keeps Failing You
The one concept that turns expensive spaghetti into great agentic engineering.

The Pull Request Illusion: How AI Is Hollowing Out Software’s Last Line of Defense
GitHub Just Added a Switch to Turn Off Pull Requests. That’s Not a Feature. It’s a Warning.

AI Agents Are Poisoning Your Codebase From the Inside
Explore how AI-generated code can silently degrade software quality through weakened tests, rising code churn, and duplication—and how teams can prevent it with better governance.

LOGOS v1.4.1: Building Multi-Engine AI Reasoning You Can Actually Trust
LOGOS v1.4.1 is a multi-engine AI reasoning orchestrator that enforces consensus, traces failures, and applies governance profiles to reduce drift and make production reasoning more trustworthy.
Showing page 1 of 2 · 20 matching posts