Flamehaven LogoFlamehaven.space
back to writing
I Audited 10 Open-Source Bio-AI Repos. Most Could Produce Outputs. Few Could Establish Trust.

I Audited 10 Open-Source Bio-AI Repos. Most Could Produce Outputs. Few Could Establish Trust.

I audited 10 visible repositories. Most could produce outputs. Very few could establish what those outputs meant.


notion image

Not which tools were “good.” Whether any of them could prove what they were doing.

Drug development is slow for a reason.
A single new medicine can take roughly ten years and more than a billion dollars to reach a patient. That is not because biology is under-optimized. It is because biology does not forgive guesses.
Every step has to be verified. Every claim has to be traced. When something goes wrong at the wrong stage, people get hurt.
The promise of AI in drug discovery is that it can compress those timelines. Find molecular candidates faster. Surface patterns in genomic data that humans would miss. Run analyses in hours that would otherwise take months.
That promise is real.
But over the past year, something else has been happening alongside it. A growing number of open-source repositories have appeared on GitHub presenting AI systems for drug discovery, genomic analysis, and clinical-adjacent biological research. They look sophisticated. They run without errors. They produce plausible-sounding output.
The question I wanted answered was not whether they looked impressive.
It was whether they could establish what their outputs actually meant.
So I audited ten high-visibility repositories and adjacent systems in the open Bio-AI ecosystem. What I found was not that most of them were obviously wrong. It was that most of them had no structural way to know whether they were right or wrong, and no reliable way to tell the user either.

The Function That Changed the Way I See This Space

notion image
Here is a piece of code from a repository called OpenClaw-Medical-Skills, presented as part of a drug-discovery workflow:
If you are a chemist, you already know why this is a problem.
If you are not, here is the short version.
`SMILES` is a strict string notation for molecular structures. A valid SMILES string is not just text. It has grammar: bond types, atom charges, ring closures, branching rules. This function does not generate molecules. It appends the letter `”C”` or `”F”` to the input string.
The docstring admits that this is a mock. But that disclosure appears in a code comment, not in the user-facing skill surface. The researcher or AI agent reading the corresponding skill document sees a workflow that claims to generate candidate molecules.
notion image
The function runs, returns a list, and produces no obvious error. By the time the pipeline says “candidate molecules generated,” the fact that the underlying operation was string manipulation is already hidden.
This is not just a bug.
It is a trust-surface failure.
And once I saw it clearly, I started seeing the same pattern in different forms across the ecosystem.

What I Audited

notion image
I evaluated ten visible repositories and adjacent systems in the Bio-AI ecosystem using a two-layer process:
  • a technical repository audit focused on structure, execution paths, and file-level findings
  • a trust-scoring framework called STEM-AI v1.0.4, which measures documentation integrity, governance posture, and biological accountability
notion image
The full score table looked like this:
notion image
  • 8 of 10 landed in T0.
  • One reached T1.
  • One reached T2.
  • None reached T3 or T4.
That is the headline result.
But the scores matter less than the reason behind them.

What I Was Actually Measuring

notion image
This audit was not trying to answer the question most people ask first.
It was not asking: which repositories are scientifically correct?
It was asking a prior question:
Does this repository provide enough structural honesty and verification architecture that a researcher or institution could even begin to establish trust before downstream use?
That means checking things like:
  • Does the README honestly describe what the repository can and cannot do?
  • Does the code fail closed when required assets or evidence are missing
  • Are there domain tests, not just syntax checks?
  • Can you trace an output back to a concrete execution state?
  • Are mock or placeholder routines presented as if they were functional?
That is a narrower question than scientific truth, but in practice it comes first. If a system cannot establish what it is doing, it cannot be trusted downstream no matter how impressive its claims are.

Four Patterns Behind the Failures

notion image
After reviewing the repositories, the same four patterns kept appearing.

1. Clinical-Adjacent Scope Without Clinical-Adjacent Accountability

Many of these repositories touch domains that are not harmless playgrounds: drug candidates, genomic interpretation, clinical reports, single-cell analysis, molecular design.
But the code and documentation often behave as if these are neutral productivity tasks. They rarely acknowledge regulatory context, risk boundaries, or the practical consequences of error.

2. CI Validates Form, Not Science

Several repositories have CI/CD. That sounds reassuring until you look at what the CI actually does.
In most cases it checks formatting, section order, or whether a script runs without crashing. It does not check whether the biological output is valid, physically plausible, or scientifically appropriate.
Passing CI in these systems often means only that the software surface is syntactically consistent.

3. Mock-as-Functional

notion image
The OpenClaw-Medical-Skills example is the clearest case, but not the only one.
Again and again, I found architectures where:
  • the function name sounds real
  • the pipeline looks professional
  • the output format is plausible
  • but the implementation is a stub, a placeholder, or an undeclared simplification
That is not the same as ordinary prototype code. Prototype code is usually disclosed as provisional. Mock-as-Functional code is dangerous because it can survive into user-facing workflows while still looking operational.

4. Strong Architecture Undermined by Unsafe Defaults

notion image
Some of the repositories are not poorly engineered in the shallow sense. They have real structure. They have actual systems architecture.
But the defaults betray them.
  • BioAgents is a good example. It has a serious multi-agent platform design, but its rate limiting can disappear in the default mode.
  • BioClaw has container isolation, but one of its important mounts is writable in a way that weakens containment.
  • Biomni has timeout wrappers, but still allows unsandboxed execution through generated subprocesses.
The result is a recurring pattern: the security story exists, but the runtime defaults hollow it out.

The One Repository That Broke the Pattern

notion image
The outlier was ClawBio.
It was not perfect, but it was the only repository in the sample that treated trust as something that had to be built into the runtime, not just implied by the project description.
Two things stood out immediately.
First, it validates biological inputs before analysis. Its scRNA-seq input logic does not simply accept a file because it has the right extension. It checks whether the data appears already processed, whether values are negative, whether integer assumptions have been violated, and whether the analysis should halt before proceeding.
Second, it generates reproducible audit identifiers by hashing the input files. That is not glamorous, but it is exactly the kind of thing missing almost everywhere else in the ecosystem: a deterministic link between a specific input state and a specific output state.
ClawBio reached `T2`, the only repository in the sample to do so. That does not make it deployment-ready. But it does show that open-source Bio-AI does not have to remain trapped at the level of plausible-looking but weakly governed execution.
notion image

What This Means

notion image
The big lesson from this audit is not that the ecosystem lacks cleverness.
It is that the ecosystem is much better at producing the appearance of capability than at producing systems that can bound, trace, and justify that capability before downstream use.
That is why I do not think the core bottleneck in Bio-AI is model quality alone.
The bottleneck is governance.
More specifically:
  • truth-surface separation
  • fail-closed runtime behavior
  • domain regression testing
  • provenance discipline
  • explicit scope boundaries
  • human-in-command reviewability
Without these, even an impressive repository remains an unstable object: too sophisticated to dismiss casually, too weakly governed to trust.

The Minimum Standard I Would Use

notion image
If I were evaluating these systems for institutional use, I would not begin with benchmark claims.
I would begin with six questions:
  1. What can this repository actually do locally?
  1. What depends on hidden assets, external APIs, or hosted outputs?
  1. Does it stop when critical evidence is missing?
  1. Does it have domain-specific tests?
  1. Can outputs be traced to concrete inputs and execution paths?
  1. Is the trust surface honest?
In my view, the minimum threshold for a supervised pilot should be `T3`. None of the repositories in this audit reached it.
That is not a reason to stop building.
It is a reason to be honest about where the field actually is.
These repositories are often valuable as research artifacts, engineering accelerators, or methodology surfaces. But most are not yet trustworthy systems in the stronger sense that medicine, biology, or automated discovery will eventually require.

Final Point

notion image
The AI-accelerated future of biology is probably real.
But if that future is going to matter outside demos, preprints, and impressive GitHub surfaces, then the field will have to solve something more basic than capability.
It will have to solve how to distinguish real outputs from plausible outputs before the cost of being wrong becomes unbearable.
That, more than any benchmark, is the line the field still has to cross.

Audit snapshot date: March 20, 2026. STEM-AI v1.0.4. This article reflects a time-bounded audit of public repository surfaces, workflow reconstruction, and selective file-level review. It is not a regulatory determination or legal judgment.

Share

Related Reading