Flamehaven LogoFlamehaven.space

Writing Hub

AI governance essays, reasoning systems notes, experiment logs, and technical writing across BioAI and engineering practice.

Current ViewReasoning / Verification EnginesSearch: AGI
Each /slop Is a Calibration Signal — AI-SLOP Detector v3.6.0 and the Claude Code Skill
Reasoning / Verification Engines

Each /slop Is a Calibration Signal — AI-SLOP Detector v3.6.0 and the Claude Code Skill

Every /slop invocation records to a project-scoped history. After 10 re-scanned files, bounded self-calibration adjusts detection weights for your codebase. Here is the mechanism, the data, and what actually shipped in v3.6.0.

Inference quality, validation, and proof surfaces#AI#AGI#AI Alignment#Deep Learning#Machine Learning#Prompt Engineering#Product Management#Software Development#AI Code#Architecture#Data Orchestration#Code Review
Can AI Review Physics? Yes — That Is Why We Built SPAR
Reasoning / Verification Engines

Can AI Review Physics? Yes — That Is Why We Built SPAR

SPAR is a deterministic framework for claim-aware review: checking whether an output deserves the claim attached to it.

Inference quality, validation, and proof surfaces#AI#AGI#AI Alignment#AI Governance#Deep Learning#Machine Learning#Cognitive Science#AI Research#Scientific Integrity#Software Development#AI Code#Contextengineering#Architecture#Data Orchestration
I Built an Ecosystem of 46 AI-Assisted Repos. Then I Realized It Might Be Eating Itself.
Reasoning / Verification Engines
Governed Reasoning

I Built an Ecosystem of 46 AI-Assisted Repos. Then I Realized It Might Be Eating Itself.

An ecosystem of 46 AI-assisted repos can become a closed loop. This article explores structural blind spots, self-validating toolchains, and the need for external validators to create intentional friction.

Inference quality, validation, and proof surfaces#AI#AGI#AI Ethics#AI Alignment#AI Governance#AI Hallucination#Mlops#Machine Learning#Deep Learning#SR9/DI2#Cognitive Science#Scientific Integrity#AI Research#Software Development#Business Strategy#Security#Architecture#Contextengineering#AI Code
Why Reasoning Models Die in Production (and the Test Harness I Ship Now)
Reasoning / Verification Engines
Governed Reasoning

Why Reasoning Models Die in Production (and the Test Harness I Ship Now)

Project note, essay, or technical log from the Flamehaven writing archive.

Inference quality, validation, and proof surfaces#AI#AGI#AI Ethics#AI Alignment#AI Governance#AI Hallucination#Mlops#Machine Learning#Deep Learning#SR9/DI2#Software Development#AI Code#Contextengineering#Architecture
Implementing "Refusal-First" RAG: Why We Architected Our AI to Say 'I Don't Know'
Reasoning / Verification Engines
Governed Reasoning

Implementing "Refusal-First" RAG: Why We Architected Our AI to Say 'I Don't Know'

Implementing refusal-first RAG means teaching AI to say “I don’t know.” This article explains evidence atomization, Slop Gates, and grounding checks that favor verifiable answers over plausible hallucinations.

Inference quality, validation, and proof surfaces#AI#AGI#AI Alignment#AI Governance#AI Hallucination#Mlops#Machine Learning#Deep Learning#SR9/DI2#Cognitive Science#Security#Architecture#Contextengineering
HRPO-X v1.0.1: from HRPO paper production-hardened runnable code
Reasoning / Verification Engines
Governed Reasoning

HRPO-X v1.0.1: from HRPO paper production-hardened runnable code

Project note, essay, or technical log from the Flamehaven writing archive.

Inference quality, validation, and proof surfaces#Mlops#AI#AGI#AI Ethics#AI Alignment#AI Governance#AI Hallucination#Contextengineering#AI Code#Architecture#Software Development#Prompt Engineering#SR9/DI2#Cognitive Science
🧠 Why Your 128K Context Still Fails — And How CRoM Fixes It
Reasoning / Verification Engines

🧠 Why Your 128K Context Still Fails — And How CRoM Fixes It

Most large language models fail in long prompts due to context rot. CRoM is a lightweight framework that improves memory, reasoning, and stability without heavy pipelines.

Inference quality, validation, and proof surfaces#AI#AGI#AI Alignment#AI Governance#Future of Work#Deep Learning#LLM#Machine Learning#Prompt Engineering#Cognitive Science
7 Signs Your AI Friend Is Becoming Real — Backed by Data & Research
Reasoning / Verification Engines

7 Signs Your AI Friend Is Becoming Real — Backed by Data & Research

AI friendship is becoming measurable. Backed by research and a $140B market forecast, discover 7 signs your chatbot feels real.

Inference quality, validation, and proof surfaces#AI Ethics#AI#Machine Learning#AI Hallucination#Deep Learning#AGI