Flamehaven LogoFlamehaven.space

Writing Hub

AI governance essays, reasoning systems notes, experiment logs, and technical writing across BioAI and engineering practice.

Current ViewSearch: AI Code
Each /slop Is a Calibration Signal — AI-SLOP Detector v3.6.0 and the Claude Code Skill
Reasoning / Verification Engines

Each /slop Is a Calibration Signal — AI-SLOP Detector v3.6.0 and the Claude Code Skill

Every /slop invocation records to a project-scoped history. After 10 re-scanned files, bounded self-calibration adjusts detection weights for your codebase. Here is the mechanism, the data, and what actually shipped in v3.6.0.

Inference quality, validation, and proof surfaces#AI#AGI#AI Alignment#Deep Learning#Machine Learning#Prompt Engineering#Product Management#Software Development#AI Code#Architecture#Data Orchestration#Code Review
How Do You Trust the AI Auditor? STEM-AI v1.1.2 and Memory-Contracted Bio-AI Audits
Scientific & BioAI Infrastructure
STEM-AI:Soverign Trust Evaluator for Medical AI Artifacts

How Do You Trust the AI Auditor? STEM-AI v1.1.2 and Memory-Contracted Bio-AI Audits

STEM-AI v1.1.2 binds a bio/medical AI repository audit to a machine-checkable memory contract, then demonstrates it on a real open-source bioinformatics repository.

Evidence-aware scientific systems#AI#AGI#AI Ethics#AI Alignment#AI Governance#AI Hallucination#Biomedical#Bioinformatics#Mlops#Deep Learning#Machine Learning#Cognitive Science#Developer Tools#DevOps#AI Research#Scientific Integrity#Business Strategy#AI Code#Contextengineering#Architecture#Data Orchestration#Code Review
The Sheepwave Has a New Shape: OpenMythos and the Rise of Architecture Hype
Cloud & Engineering Foundations

The Sheepwave Has a New Shape: OpenMythos and the Rise of Architecture Hype

A technical-opinion essay on OpenMythos, Claude Mythos, README-driven AI hype, and why architecture claims need source-level verification before becoming public belief.

Operational surfaces that survive real deployment#Code Review#Data Orchestration#Contextengineering#Software Development#Prompt Engineering#Architecture#AI Code#AI Governance#AI Alignment#LLM#Deep Learning#Machine Learning#Cognitive Science#Github#Open Source
FLAMEHAVEN FileSearch: Why This RAG Engine Feels Different from the Usual Stack
Cloud & Engineering Foundations

FLAMEHAVEN FileSearch: Why This RAG Engine Feels Different from the Usual Stack

A technical look at FLAMEHAVEN FileSearch: BM25+RRF hybrid retrieval, chunk-addressable indexing, deterministic DSP vectors, and the trade-offs behind a lower-overhead self-hosted RAG engine.

Operational surfaces that survive real deployment#AI#AGI#AI Alignment#AI Governance#DevOps#Open Source#Developer Tools#Prompt Engineering#Software Development#Github#AI Code#Architecture#Data Orchestration
It Gets Smarter Every Scan: AI-SLOP Detector v3.5.0 and the Self-Calibration Loop
Cloud & Engineering Foundations

It Gets Smarter Every Scan: AI-SLOP Detector v3.5.0 and the Self-Calibration Loop

AI-built apps are starting to fail in public. Not every failure is static-analysis territory, but many share the same upstream condition: plausible-looking code passing review without carrying enough real logic. AI-SLOP Detector v3.5.0 adds a self-calibration loop to reduce that gap.

Operational surfaces that survive real deployment#AI#AGI#AI Alignment#AI Governance#Deep Learning#Machine Learning#Open Source#Developer Tools#Prompt Engineering#Software Development#AI Code#Contextengineering#Architecture#Data Orchestration
Can AI Review Physics? Yes — That Is Why We Built SPAR
Reasoning / Verification Engines

Can AI Review Physics? Yes — That Is Why We Built SPAR

SPAR is a deterministic framework for claim-aware review: checking whether an output deserves the claim attached to it.

Inference quality, validation, and proof surfaces#AI#AGI#AI Alignment#AI Governance#Deep Learning#Machine Learning#Cognitive Science#AI Research#Scientific Integrity#Software Development#AI Code#Contextengineering#Architecture#Data Orchestration
Bridging the Gap: From AI Slop to Mathematical Governance
Scientific & BioAI Infrastructure

Bridging the Gap: From AI Slop to Mathematical Governance

A mathematical framework for detecting AI-generated code slop using AST distributions, Jensen-Shannon divergence, and geometric governance gates.

Evidence-aware scientific systems#AI#AGI#AI Ethics#AI Alignment#AI Governance#Deep Learning#Machine Learning#AI Research#Scientific Integrity#Prompt Engineering#Programming#Software Development#AI Code#Contextengineering#Architecture#Data Orchestration
My AI Maintainer Kept Making Wrong Calls. So I Made It Report Its State Before Touching Anything.
Cloud & Engineering Foundations
MICA Series

My AI Maintainer Kept Making Wrong Calls. So I Made It Report Its State Before Touching Anything.

Part 6 moves from landscape to operation. This is what MICA looks like when it is actually running inside a real maintenance workflow — session report, self-test, drift, invariants, and operator judgment.

Operational surfaces that survive real deployment#AI#AGI#AI Alignment#AI Governance#Developer Tools#DevOps#AI Code#Contextengineering#Architecture#Data Orchestration
Prompt → RAG → MCP → Agent → Harness, and What?
Cloud & Engineering Foundations

Prompt → RAG → MCP → Agent → Harness, and What?

Why the next layer in AI may be governance infrastructure, not just better agents.

Operational surfaces that survive real deployment#AI#AGI#AI Ethics#AI Alignment#AI Governance#AI Hallucination#LLM#Cognitive Science#Developer Tools#Prompt Engineering#Software Development#AI Code#Contextengineering#Architecture#Data Orchestration
The Harness Is the Product: What the Claude Code Leak Actually Revealed About AI Agent Architecture
Cloud & Engineering Foundations

The Harness Is the Product: What the Claude Code Leak Actually Revealed About AI Agent Architecture

The Claude Code leak exposed more than source. It revealed that modern AI agent performance depends heavily on the harness around the model.

Operational surfaces that survive real deployment#AI#AGI#AI Alignment#AI Governance#LLM#Deep Learning#Machine Learning#DevOps#Prompt Engineering#Software Development#Product Management#AI Code#Contextengineering#Architecture#Security#Data Orchestration
After Auditing 10 Bio-AI Repositories, I Think We're Scaling the Wrong Layer
AI Governance Systems
STEM-AI:Soverign Trust Evaluator for Medical AI Artifacts

After Auditing 10 Bio-AI Repositories, I Think We're Scaling the Wrong Layer

After auditing 10 open-source Bio-AI repositories, one pattern stood out: the field is scaling packaging faster than verification. Here is what that gap actually costs.

Control, auditability, and safe boundaries#AI#AGI#AI Ethics#AI Governance#Mlops#Cognitive Science#Open Source#DevOps#AI Code#Architecture#Github#Software Development
I Audited 10 Open-Source Bio-AI Repos. Most Could Produce Outputs. Few Could Establish Trust.
Scientific & BioAI Infrastructure

I Audited 10 Open-Source Bio-AI Repos. Most Could Produce Outputs. Few Could Establish Trust.

I audited 10 visible repositories. Most could produce outputs. Very few could establish what those outputs meant.

Evidence-aware scientific systems#AI#AI Ethics#AI Alignment#AI Governance#Biomedical#Bioinformatics#Future of Work#LLM#Open Source#DevOps#Scientific Integrity#Prompt Engineering#Github#AI Code#Contextengineering#Architecture#Security#AI Research

Showing page 1 of 4 · 47 matching posts