DCFN - Research

Changelog

Currently v0.5.1. ← Back to home

DCFN-Research — Changelog

What's changed in the engine's user-visible output, in reverse chronological order. Pre-1.0 versioning convention:


v0.5.1 — 2026-05-03

Decision Memo prompt: orphan-year citations no longer allowed


v0.5.0 — 2026-05-01

First prescriptive artifact — R&D Intelligence Decision Memo

Up to v0.4.x, every Research output was descriptive: Article, Tech Report, Bridge Digest, Syntari Record, JSON. All four external instances (Perplexity, Grok, Gemini Deep Research ×2) independently identified the same gap on review — these are excellent analysis packages, but they are not artifacts a customer can FUND, FILE, or ACT ON immediately. DCFN-Patents earned its category by shipping deliverables (provisional drafts, continuation memos) that decision-makers could take directly into a meeting. Research had nothing of equivalent shape.

v0.5.0 ships the first one: an R&D Intelligence Decision Memo, a 2-3 page persona-targeted prescriptive deliverable produced from any successful engine run. Generated alongside the Article and Tech Report — additive, not a replacement. Audience for this prototype is narrow on purpose: pharma R&D directors, biotech corp-dev leads, deep-tech VC desks, and corporate R&D portfolio heads who allocate $1M-$50M on R&D bets. Same operational pattern as Patents' attorney audience (high-LTV, willing to pay $4-20K/seat for a tool that materially affects large decisions). Other personas (Replication, Funder, Licensee) are deferred — one persona ships first to validate willingness-to-pay before forking the synthesis pipeline.

Verified on real session

Generated against data/reports/run_c66198828e46/ (Education & EdTech, 687 articles). Memo correctly cited specific entropy nodes by title + year + severity, named SVW convergence events (svw_002, svw_022, svw_023) with paper titles and scores, surfaced apriori rules with confidence numbers, and honestly flagged the absence of contradiction-resolution data + lack of bridge nodes as engine-input limitations rather than burying them.

Not in this release


v0.4.0 — 2026-05-02

Cloud Run + GPU + GCS sessions migration — foundation cuts

Mirrors DCFN-Patents v0.8.0 → v0.9.0 → v0.6.0 playbook validated 2026-05-01. Lays the groundwork for Render → Cloud Run cutover without changing any user-facing behavior on the live Render service. Phase 3 of the migration plan documented at Needs Review/Pricing + Architecture Decisions/RESEARCH_MIGRATION_PLAN_2026-05-02.md.

Not in this release


v0.3.12 — 2026-05-02

Customer-output URL: dcfn-research.onrender.comresearch.livingedenframeworks.com


v0.3.11 — 2026-05-01

Site footer cleanup (cross-build standard)


v0.3.10 — 2026-04-30

Citation walk: hybrid two-pass for multi-source corpora

v0.3.5's bidirectional citation walk shipped assuming all reference IDs were Semantic Scholar paperIds (40-char hex). In practice the merged corpus pulls from 4-6 sources and references[] is mixed-format: OpenAlex Work IDs, PubMed UIDs, arXiv IDs. v0.3.8's S2-only filter prevented the resulting 400-Bad-Request crash but at the cost of ~zero expansion on multi-source corpora (one local test: 749 non-S2 IDs filtered, 0 added).

v0.3.10 closes the gap with a hybrid two-pass design:

Graceful degradation: if S2 batch returns 429 (free-tier rate limit, common) or any other error, the IDs that came in via DOI translation get re-attempted via per-source fetch. So the entire walk doesn't depend on S2 cooperating.

Metadata richness: expand_via_citation_walk return dict now includes ids_resolved_via_s2_native, ids_resolved_via_doi_translation, ids_resolved_via_per_source, ids_unresolved, per_source_breakdown. Operator (and Z reading the report) can see exactly where each neighbor came from and which sources contributed.

Empirical validation (local Single-Cell corpus, 100 OpenAlex neighbors, S2 rate-limited): 25 articles added in 8.6s via Pass 2 OpenAlex fallback alone. With production S2 API key cooperating, Pass 1 + Pass 2 combined would land substantially more.

New module id_translation.py centralizes paper-ID source recognition + DOI prefix formatting so downstream code doesn't sprinkle prefix-matching logic.

Coverage note: papers without abstracts are filtered by the standard _to_article_record schema requirement. Tier 1+ engine variant will handle abstract-less papers via structural-metadata-only ingest path (separate ship).


v0.3.9 — 2026-04-30

Pipeline timing instrumentation in autonomous-scheduler path (Charter §16)

main.py's user-driven path already had per-stage timing via stage_timings. The autonomous-scheduler path (scheduler.py:_run_autonomous_pipeline) was missing it — only total elapsed was logged. Added per-stage capture for: qeb_encoding, concept_graph, cte_traversal, apriori, svw, hypothesis_generation, calibration, bridge_detection_and_rerank. Surfaces as a single [PIPELINE_TIMING] log line per run + persisted to the report's stage_timings field for downstream tooling.

Triggered by Charter §16 codification (Patents L1 ran 50 min and we had no per-step data to answer "should we upgrade Render tier?"). This closes the gap on the Research autonomous path so the same question is answerable empirically there too.

Note: the multi-source citation walk hybrid (DOI translation + per-source fanout) flagged in v0.3.8 is now tracked as v0.3.10 (next minor).


v0.3.8 — 2026-04-30

Critical fix: every autonomous run since v0.3.7 was crashing

Local Research validation surfaced two bugs in code I shipped earlier today.

Coverage scope

S2 ID filter is a safety net, not the production architecture. Multi-source corpora (OpenAlex, PubMed dominated) require cross-source ID resolution. Hybrid two-pass implementation lands in v0.3.10 (DOI translation + per-source fanout).

Follow-on work tracked separately for v0.3.9: translate non-S2 IDs to S2's prefix syntax (DOI:10.x, PMID:NNN) before batching, OR fan out per-source (OpenAlex API for openalex: IDs, PubMed E-utilities for pmid: IDs). Not a v0.3.8 ship — needs design.


v0.3.7 — 2026-04-30

Hypothesis-target granularity + trajectory anti-drift

Two coupled fixes for Perplexity's 2026-04-30 broad-vocabulary findings.


v0.3.6 — 2026-04-30

Vocabulary-bleed suppression in convergence detection


v0.3.5 — 2026-04-30

Engine depth: bidirectional citation-walk corpus expansion (Charter §12 Pattern B)


v0.3.4 — 2026-04-30

Quality fix


v0.3.3 — 2026-04-30

Quality fix


v0.3.2 — 2026-04-30

Patent attribution accuracy

Footer's "Built on" line was undercounting: said "6 U.S. Patents Pending" and named CTE + QECO. Actual total since the Tesseract Composition supplemental landed (2026-04-20) is 8, and the engine rides more than two substrate patents. Updated to:

Same correction applied to the Firebase brand site's DCFN-Research card.


v0.3.1 — 2026-04-30

Quality fix


v0.3.0 — 2026-04-30

Discovery-driven autonomous runs

The Research engine's autonomous-run path now drives from a discovery-agent-fed queue instead of cycling fixed domains. A discovery agent identifies new research topics worth running by querying Semantic Scholar (with PubMed fallback) for substantive recent activity in curated seed areas, derives a topic configuration from the top results, and proposes it for human review. After a 7-day cooldown without rejection, the proposal auto-promotes into the live run queue, where the engine executes the full pipeline against it once or twice before going dormant.

Why this matters: it converts the autonomous path from "run the same three domains every day" (which produces noise) into "surface new research territory worth exploring" (which produces signal). Each run feeds the Bridge Inbox + LEF Ai Upstream telemetry channels — autonomous runs are the substrate's input.

Engine output

Three-layer report architecture


v0.2.0 — 2026-04-08

Engine output

Sources


v0.1.0 — 2026-03-15

Initial deployment. Single-page intake → multi-source ingest → concept graph construction with typed edges → Cognitive Traversal Engine (5 operations: backward / forward / branch cataloging / entropy / golden token) → SVW convergence detection → Apriori pattern mining → Article + Technical Report generation. Free 5 runs / month / browser; $15 unlock for Layer 2 + Layer 3 deeper traversal.