Writing Hub
AI governance essays, reasoning systems notes, experiment logs, and technical writing across BioAI and engineering practice.
Project Topics

Each /slop Is a Calibration Signal — AI-SLOP Detector v3.6.0 and the Claude Code Skill
Every /slop invocation records to a project-scoped history. After 10 re-scanned files, bounded self-calibration adjusts detection weights for your codebase. Here is the mechanism, the data, and what actually shipped in v3.6.0.

AI-SLOP Detector v3.5.0 — Every Claim, Verified Against Source Code
AI-SLOP Detector v3.5.0 made 7 claims on LinkedIn —self-calibration logic, download numbers, defect detection. Here's every claim verified against actual file paths and line numbers. The code speaks for itself.

Can AI Review Physics? Yes — That Is Why We Built SPAR
SPAR is a deterministic framework for claim-aware review: checking whether an output deserves the claim attached to it.

I Built an Ecosystem of 46 AI-Assisted Repos. Then I Realized It Might Be Eating Itself.
An ecosystem of 46 AI-assisted repos can become a closed loop. This article explores structural blind spots, self-validating toolchains, and the need for external validators to create intentional friction.

Why Reasoning Models Die in Production (and the Test Harness I Ship Now)
Project note, essay, or technical log from the Flamehaven writing archive.

Implementing "Refusal-First" RAG: Why We Architected Our AI to Say 'I Don't Know'
Implementing refusal-first RAG means teaching AI to say “I don’t know.” This article explains evidence atomization, Slop Gates, and grounding checks that favor verifiable answers over plausible hallucinations.

LOGOS LawBinder: From Governed Reasoning to Audit-Grade Execution
This article explains how LOGOS v1.4.1 improves production AI reasoning with multi-engine orchestration, complexity-aware governance, and audit-friendly failure tracing.

LOGOS v1.4.1: Building Multi-Engine AI Reasoning You Can Actually Trust
LOGOS v1.4.1 is a multi-engine AI reasoning orchestrator that enforces consensus, traces failures, and applies governance profiles to reduce drift and make production reasoning more trustworthy.

LawBinder v1.3.0: Governance as a Kernel (Not a Guardrail)
LawBinder v1.3.0 shows how AI governance can run like a kernel, using deterministic Rust-based enforcement, replayable audit signatures, and bounded-latency policy checks in the critical path.

HRPO-X v1.0.1: from HRPO paper production-hardened runnable code
Project note, essay, or technical log from the Flamehaven writing archive.

Turning a Research Paper into a Runnable System
Turn a research paper into a runnable system. This article shows how HRPO’s core equations were implemented with bounded policy lag, KL rejection, and execution checks to test real-world fidelity.

🧠 Why Your 128K Context Still Fails — And How CRoM Fixes It
Most large language models fail in long prompts due to context rot. CRoM is a lightweight framework that improves memory, reasoning, and stability without heavy pipelines.
Showing page 1 of 2 · 16 matching posts