Roadmap (Definition #1)

Technical Roadmap: Understanding WBE Through Measurement, Reconstruction, and Implementation

Break 'what counts as progress' into a question tree with reading order and minimum arrival conditions

Public Page Updated: 2026-04-03 Provisional draft (updated continuously; summary-layer astrocyte PET route sync on 2026-04-03)

How to use this page

Read this first to avoid getting lost

This page is a map of WBE questions arranged under measurement, reconstruction, implementation, verification, and social deployment. Its purpose is to keep the reader from losing track of which problems are foundational and which remain unresolved.

  • P0-P2 fix what counts as success before measurement or implementation work is discussed.
  • The index allows readers to jump directly to a specific question family.
  • Stronger claims are deliberately placed later so earlier levels are not skipped.
  • V8 now splits invasive language BCIs into throughput / expressivity, transfer-assisted initialization, fixed-decoder durability slices, and adaptive rescue, so one speech-BCI result is not silently promoted to all four.
  • R0 now separates structural scaffold, fast executable state, and maintenance-state families, so `connectome + state` is not treated as one knob.
  • R3 / R5 separate latent-state and maintenance-state questions by evidence tier and timescale, so same-day fit and multiday maintenance do not collapse into one success.
  • R3 / R5 now force maintenance-relevant claims to name whether phospho-signaling, local proteostasis, and cargo-routing were measured, externally calibrated, model-estimated, or omitted, so a generic latent-state label is not treated as maintenance-complete evidence.
  • M2 separates hardware latency from biological timing-state, so a fast device loop is not confused with timing-complete reconstruction.
  • For phase-targeted closed loops, M2 / I1 now separate oscillation estimability, causal-versus-post-hoc targeting benchmark, circular targeting precision, downstream effect, and phase stability, so phase error alone is not treated as sufficient evidence.
  • M1 / M5 keep neural state distinct from vascular transfer state, so a BOLD amplitude difference is not silently promoted to a neural difference.
  • M1 / M3 now treat human in vivo evidence as a split ladder, so proxy class, operational maturity, calibrator role, and family-internal quantity / target / transport differences are kept separate instead of being collapsed into one observability score.
  • M1 / M3 now also keep human barrier-side evidence split inside the blood-CSF-barrier lane itself, so choroid-plexus perfusion, blood-to-CSF transport, water cycling, and apparent BCSFB exchange are not silently merged into BBB water-exchange or tracer-specific BBB transport.
  • M1 / M3 now also separate human local clinical-unit allocation from noninvasive perturbation-conditioned excitability routes, so `human excitability evidence` is not treated as one row.
  • M1 / M3 now also split human astrocyte-related PET into SMBT-1 MAO-B target-validation / disease-context / quantification / biodistribution routes, SL25.1188 MAO-B quantification / severity-conditioned routes, and I2BS routes, so `astrocyte PET` is not treated as one interchangeable human row.
  • M3 now also keeps cohort / physiological regime compatibility and row robustness / availability separate from those three per-row axes, so a proxy-rich bundle is not silently treated as one matched human baseline or one stable stack.
  • M4 / I6 now treat the body / environment boundary as an explicit scientific audit item across fast loop routes and slow internal-milieu routes, so a fast local loop is not silently promoted to solved embodiment.
  • M7 / V4 now separate state annotation, trait-like backbone, biological drift, interface / decoder drift, and recalibration burden, so same-day success and adaptive rescue do not collapse into one longitudinal claim.
  • Wearable OPM-MEG is treated as movement-tolerant macro measurement only when shielding, field control, sensor calibration, and anatomy choice are disclosed.
  • R6 treats personalization not as a pure performance trick, but as a verification problem that must separate target signal from subject fingerprint and setup shortcuts.
  • R1 / R2 now separate focal posterior-support, sparse debiased inference, extended-source extent reconstruction, forward-model uncertainty, and validation board, so a newer inverse family is not silently treated as one generic truth upgrade.
  • R4 now treats DCM / effective-connectivity outputs as route-card claims that must disclose model space, observed-subsystem closure / latent-confound audit, node-definition policy, sampling / transformation sensitivity, observation assumptions, validation, reliability, and abstention.
  • R7 now separates observability, structural identifiability, practical identifiability, ambiguity class, and minimum-sufficiency stop rules, so richer proxies or more modalities are not silently promoted to uniqueness.
Best for
Readers who want the big picture, need a learning order, or want claim strength organized by level
Reading time
20-30 min (5 min for the index only)
Accuracy note
This page organizes the research landscape. It does not mean that the listed items are already solved.

Relatively clear at this stage

What we know now

  • Splitting the problem into P/M/R/I/V/D makes it easier to see which questions are foundational and which sit higher up.
  • The dependency structure that prevents strong claims from skipping earlier layers is fairly clear.
  • This page should be read as a dependency map, not as a checklist of solved items.
  • Invasive language BCI progress is multi-route: throughput, transfer initialization, fixed-decoder durability, and adaptive rescue need separate labels.
  • A restoration target is not one dial; scaffold, fast state, and maintenance-state layers place different ceilings on what can be claimed.
  • Even if connectome or local-activity evidence improves, latent-state and maintenance-state audits are still separate requirements.
  • A prediction-useful latent variable is not automatically a maintenance-complete state estimate; phospho-signaling, local proteostasis, and cargo-routing can still remain omitted families.
  • For hemodynamic modalities, neural interpretation and vascular transfer / CVR audit are separate requirements.
  • Closed-loop device timing and biological conduction timing are different audits; passing one does not auto-pass the other.
  • For phase-targeted loops, a low mean phase error still does not prove that the target oscillation was estimable, that the causal estimator matched a post-hoc benchmark, that the physiological or behavioral effect followed, or that one preferred phase stayed stable across time.
  • A disclosed body / environment boundary, including slow internal-milieu disclosure, is a separate audit from low latency or online accuracy.
  • Human-side advances reduce different error terms and remain differently deployment-limited; proxy-rich evidence is not yet same-subject, same-session, externally calibrated state identification.
  • Proxy class, operational maturity, and calibrator role are different questions; a real human route may still calibrate only one bounded hidden-state family.
  • A human route family name is not yet one row; 1H-MRSI similarity, high-resolution 1H-MRSI metabolite-distribution mapping, 31P metabolite / pH balance, 31P MT exchange-flux, 31P NAD-content mapping, localized functional 31P NAD-dynamics, deuterium metabolite-mapping / absolute-quantification routes, deuterium kinetic-rate imaging, tract-scale transmission-speed estimation, quantity-defined myelin MRI / tissue-health-sensitive ratios, BBB water-exchange, tracer-specific BBB PET transport, blood-CSF barrier / choroid-plexus perfusion / blood-to-CSF transport / water-cycling / apparent-exchange routes, SMBT-1 MAO-B target-validation / disease-context / quantification / biodistribution routes, SL25.1188 MAO-B quantification / severity-conditioned routes, and I2BS routes remain separate routes with different ceilings.
  • Human clinical single-unit allocation, sleep-homeostasis / plasticity recalibration, and state-gated perturbation are also separate human routes with different direct observables and different ceilings.
  • A named astrocyte-related PET paper is not yet one route family; MAO-B target validation, MAO-B disease-context contrast, MAO-B quantification / biodistribution burden, SL25.1188 covariate ceilings, and I2BS routes remain separate questions.
  • Even when human proxy rows look complementary, cohort / physiological regime mismatch, route-local repeatability, cross-centre transfer, and complete-case availability still remain separate gates.
  • Longitudinal success is not one axis; within-day state fluctuation, trait-like backbone, biological drift, interface / decoder drift, and recalibration burden answer different questions.
  • OPM-MEG motion tolerance still depends on magnetic shielding, active nulling, calibration, and source-model disclosure.
  • The fact that personalization helps performance is not the same as showing a population-level neural signal; subject-fingerprint and acquisition-distribution audits are still required.
  • For ESI, source regime / target object, uncertainty object, forward-model uncertainty route, and validation board remain separate requirements; a stronger inverse family is not one universal answer type.
  • Richer observability and better fit do not by themselves collapse the compatible solution set; ambiguity class, candidate-space, bridge / composition validity, experiment-design leverage, and minimum-sufficiency stop rules remain separate.

Still unresolved beyond this point

What we still do not know

  • It is still unsettled which measurement granularity will ultimately be sufficient for WBE.
  • V5 identity questions and D-series social deployment conditions remain unresolved.
  • The shortest practical route may change as data and benchmark infrastructure improve.
  • How specialized human routes can be fused into same-subject, same-session, externally calibrated evidence without recreating latent-state ambiguity remains unresolved.
  • It also remains unsettled which backbone object should be the default target for WBE-relevant longitudinal claims: latent dynamics, representational geometry, or another cross-modal invariant.

Learn the basics

Check the basics in the wiki

Summary

This page is the map of mind-uploading research. Instead of jumping straight to conclusions, it organizes in sequence what to measure, what to reconstruct, how to run it, and how to verify it. In practice, it visualizes where the field currently stands across measurement -> reconstruction -> implementation -> verification -> social deployment.

For New Readers

This page is dense. It is usually easier to read after WBE 101 and EEG 101, with the Glossary nearby if needed.

If The Index Notation Feels Unclear

If you want a beginner explanation of the P/M/R/I/V/D notation and their dependency structure, start with Wiki: How To Read The Roadmap. The main text is closer to a map than a dictionary, so a supporting guide helps.

If You Want To Know This Page's Role Among The Theory Pages

This page is less a theory essay than a map of dependencies among questions. The entry point for claim levels is WBE 101, the long-form discussion of theory and limits is the research note, and design principles live in the framework section inside Perspective. For a one-page view of the theory pages as a whole, see Wiki: Theory Page Reading Guide.

If You Want Only The V5 Identity Entry Point First

Identity questions span V5, V7, M7, R5, and related nodes in this roadmap. If you want to sort out in advance how to think about memory, values, learning, branching, and longitudinal continuity, start with Wiki: Identity Evaluation and Continuity Tests.

If You Get Stuck On Longitudinal Evaluation In M7 / V4

State on a given day, relatively stable trait, and drift caused by learning or updates are not the same thing. If you want to sort out the difference between short-term fluctuation and long-term continuity first, see Wiki: State, Trait, and Drift.

If You Get Stuck On Operations In I5 / I8 / V7

If the differences among safe updates, branch handling, stop rules, and kill switches are unclear, start with Wiki: Update, Branching, and Stop Rules. The roadmap points to the issues; the wiki supplies the operational basics.

If You Get Stuck On Closed Loops In I1 / L3

Even when something is called "real-time," the difficulty depends on how latency, jitter, end-to-end return, and safety stops are measured. For a beginner entry point to that distinction, see Wiki: Closed Loops, Latency, Jitter, and Safety Stops.

If You Get Stuck On "Isn't A Connectome Enough?"

In the R-series, the branching point is not only the wiring diagram but also how cell types, intrinsic excitability/homeostatic set points, synaptic efficiency, delay and myelination, neuromodulation, and glia are treated. If you want to see from primary literature which claims fail when those state variables are missing, see Wiki: Why a Connectome Is Not Enough and Wiki: Homeostatic Plasticity and Maintenance State.

If You Want The Current Human In Vivo Measurement Ladder First

This roadmap now distinguishes not only which variable class a human route constrains but also how specialized or model-heavy that route still is. The site's current rule is stricter than a simple three-axis ladder: per-row proxy class / operational maturity / calibrator role still has to be combined with bundle-level checks on cohort / physiological regime compatibility and row robustness / availability disclosure. For the front-door ladder, start with WBE 101: human observability ladder; for stack-specific ceilings and stop rules, see Wiki: observability and claim ceiling by measurement stack and Verification: Observability Budget. If the argument also bridges live measurement to later fixation / ex vivo follow-up or cross-day reacquisition, add the Verification: State-Continuity Bridge Card and Wiki: State-Continuity Bridge instead of reading same-subject or same-brain as same-state by default.

If You Get Stuck On Terms Like Missing, Unfinished, Or Partially Solved

This page uses short status labels for what is still lacking in each node. If you want to clarify what partially solved, exploratory, not yet standardized, and unfinished each mean, start with Wiki: Reading Partial Solutions, Exploratory Stages, and Unfinished Areas.

If You Want To Organize Return Paths By Claim Ladder Level

Because Roadmap covers all of L0-L5, the next public page depends on which level you want to inspect now. For a one-page route map by claim level, see Wiki: Reading Routes By L0-L5 Level.

If You Want To Choose The Next Single Page After This One

If you want a one-page view of where to go next after Roadmap, measurement, reconstruction, implementation, verification, or social deployment, see Wiki: Five Routes After the Roadmap.

Three Things To Fix First

  • This page's role: not to declare conclusions, but to fix where each question lives.
  • What this page can do: organize current research problems with order and dependencies attached.
  • What this page cannot do alone: it cannot prove that the final questions of identity or consciousness are already settled.
Symbol What It Groups Together Read It First When...
P Definition of progress. It decides what counts as success. You want to align on the target before entering technical discussion.
M Measurement. It covers what can be observed, and at what precision. You want to start from the differences among EEG, fMRI, invasive measurement, and related modalities.
R Reconstruction. It covers what can be estimated from observations. You are interested in inverse problems, latent state, causal models, or uncertainty.
I Implementation. It covers how models run and where they break. You want to inspect closed loops, latency, computational load, or safety design.
V Verification. It covers what must be measured for progress and what counts as failure. You want benchmarks, causal tests, identity questions, or replication criteria.
D Deployment. It covers rights, consent, fairness, and security. You want to read beyond the technical layer into operations and institutions.

How to Use

This page is not a book but a map of the full research landscape. First read through the question tree, then deepen each node according to the learning order. Only two operational rules are essential here: (1) keep the source links and (2) explicitly mark what remains provisional or uncertain.

2026-03 Literature audit: Three walls of measurement, identification, and intervention

In this re-audit, we focused on updating M3 (spatial granularity), M6 (intervention design), R7 (identifiability), and V2 (causal testing) in this roadmap. The reason is simple: the most common misinterpretations in technical discussions surrounding WBE are to read increase in observables as state completeness, to read increase in prediction performance as obtaining a unique solution, and to read introducing one intervention as the completion of causal verification. The following three are lower limits that are relatively consistently supported by primary literature as of March 2026.

Wall What the primary literature now supports Revision policy on this page
Measurement granularity wall EEG / MEG provide global current-field measurements, while fMRI provides hemodynamic proxies, so what we obtain non-invasively in humans is basically macroscopic observation. By contrast, Dorkenwald et al. (2024) and MICrONS Consortium et al. (2025) advance structure-function correspondence in model organisms and local cortex, not state-complete measurement of the human whole brain. In M3, ``what is directly visible and what remains latent'' is fixed for each granularity, and structural scaffolding and state completeness are written separately.
Identifiability Wall Although progress has been made in direct validation of HD-EEG/ESI, Seeber et al. (2019), Unnwongse et al. (2023), and Hao et al. (2025) demonstrate conditional detectability and auditability of localized errors, not general unique recovery of internal states. The conclusion changes depending on the source depth, conductivity assumption, head model, and stimulation conditions. R7 defines predictability, localization, and identity as separate levels, and requires family comparison and abstention conditions.
Intervention validation wall TMS-EEG, phase-locked stimulation, adaptive DBS, and BCI closed loops are all strong evidence routes, but as Hernandez-Pavon et al. (2023), Zrenner et al. (2018), and Wilson et al. (2010) show, comparisons fail unless stimulation site, intensity, masking, artifact windows, and latency / jitter measurements are fixed. Intervention evidence is therefore tiered rather than binary. In M6 and V2, we stage passive observation, held-out perturbation, online loop, local causal intervention, and long-term closed loop, and make request logs explicit.
Points of criticism clarified in this audit

Therefore, this roadmap will not allow the three short-circuits of ``It's enough because it's multimodal'', ``It's unique because it's highly accurate'', and ``It's causal because it's stimulated''. The stronger the claim is, the closer to the WBE, the earlier the measurement granularity, candidate model space, intervention design, and abstention in case of failure should be made public.

If You Want The Quantitative Reason Non-invasive Routes Still Stop Early

The updated measurement-stack wiki now makes explicit why EEG / MEG remain conditional for deep sources even after intracranial validation, and why hemodynamic amplitudes remain limited by vascular-transfer and autonomic-state confounds. For the direct-validation numbers and the resulting ceiling, see Wiki: EEG / MEG visibility, inverse, and validation wall and Wiki: hemodynamic transfer wall.

Index (keywords → QA)

You can jump directly to the corresponding QA from the term. The QA-ID is in parentheses. The same anchor is attached to the main text side.

How to use

First, use the "Quick Index" to move to the target term, then use the "QA Index by Category" to traverse surrounding nodes.

Quick index

QA index by category (all items)

Meta (definition of progress)
Measurement
Reconstruction
Implementation
Verification
Deployment (social implementation)

Definition of forward movement (fixed at the beginning)

In order to claim "progress", it is first necessary to decide what can be reproduced to win (=evaluation axis) andwhat should happen to lose (=falsification conditions). Here, we will compress the underlying questions into three questions.

P0 What is the operational definition of "mind upload"? (What do you want to reproduce?) META

Question:How does this site treat “mind uploading / WBE”? (Is it only the behavior that can be seen from the outside? / Is it the internal causal structure? / Does it include the identity and consciousness?)

Branch (example):(A) Reproduction of behavior/function (equivalent to black box) / (B) Reproduction of neural dynamics (emphasis on content) / (C) Includes continuity of individuality and subjectivity (strong assertion)

Falsifiability:If we talk about results without vague definitions, evaluation becomes an “afterthought” and comparison of progress becomes impossible

Next: Use the "Verification" node on this page to determine the measurable criteria and loss conditions for (A), (B), and (C).

P1 Where to place the claim level (claim staircase)? META

Question:At what level should research results be considered "progress"?

  • L0:Data collection and reproducible analysis (standardization, quality control, retestability)
  • L1:Decoding/encoding (neural → action/stimulus or vice versa prediction)
  • L2:Generative model (can extrapolate neural/behavioral behavior even under unlearned conditions and predict interventions)
  • L3:Closed-loop implementation (interacts with the environment in real time and operates stably)
  • L4:Assertion of identity (assessing continuity of memory, values, and learning using pre-registered criteria)
  • L5:Social implementation (rights, safety, and governance are established at the same time as technology)

Falsifiability:Preventing talking about “achieving L1” as “achieving L4” (switching scope)

Next: Match the "required data", "required model", and "evaluation suite" for each level on this page

P2 What are the minimum deliverables? (Data/Code/Evaluation/Audit) META

Question:What must remain in order to make “progress” reproducible?

  • Data:Raw data + metadata (BIDS etc.) + anonymization/consent log
  • Code:Fixed version analysis pipeline (including environment and dependencies)
  • Evaluation:Pre-registered indicators/tests/baseline comparison
  • Audit:Model differences, reproduction execution logs, failure examples (negative results)

Falsifiability:Only the “how to present” the results takes precedence, making the claim unverifiable

Next: First, lock down the L0-L2 deliverable template for Mind-Upload, including data structure and evaluation scripts.

Tree of questions (detailed version: consolidated on one page)

Below, we will break down the technical questions into "measurement → reconstruction → implementation → verification → social implementation" and summarize them all on one page. Each item is collapsible (click to expand).

The shortest course for those who are lost

  • People who want to know the whole picture first:P0M0R1V1 If you read them in this order, you will understand in one go what to measure, restore, and verify.
  • For those who want to start with EEG:Reading in the order of M0M2M9V1 will make it easier to follow from measurement conditions to evaluation.
  • For those who want to start with the topic of identity:P1V5V7D1 If you read in this order, you will see the conditions necessary for a strong argument and the institutional barriers.
1. Measurement

Measurement: What resolution do you need?

Central question: "What part of the brain, and at what spatiotemporal resolution, should be taken to arrive at a unique object to be reconstructed?"

M0 Assumptions for measurement: In vivo / invasive / destructive scan? MEASUREMENT

Question:Based on safety and scale constraints, fix "what to aim for" on which measurement route.

Branch (example):(A) Train the model with non-invasive (EEG/MEG/fMRI) / (B) Strengthen causal inference with invasive (ECoG/deep/animal) / (C) Obtain structure with destructive/ultra-high resolution (future vision)

Conditions for disproving: Even if only the measurement is “sophisticated” without the purpose (restored object) being defined, identification will become impossible

Next: Determine R0 (restore target) and V0 (verification criteria) on this page first

M1 What is the difference between observed variables: electricity (EEG)/blood flow (fMRI)/spikes (invasive)? EEGFMRI

Question: When the observation changes, the recoverable target also changes, whether structure, state, or learning rule. What can be identified from which observations?

Issues: Time-series speed, spatial resolution, ease of causal intervention, whole-brain coverage, and cost.

Observation route What is directly observed I can say this relatively strongly Latent state that remains as it is
EEG / MEG It is a mixed field of globally synchronized currents. You can follow ms-scale state transitions, global dynamics for each frequency band, and closed-loop time constraints. The uniqueness of deep sources, cell type, neuromodification, glial status, and current synaptic efficiency are not directly determined.
fMRI Hemodynamic proxy. Region-scale recruitment, relatively slow network state, and wide coverage within the same individual are easy to see. ms-scale timing, excitatory/inhibitory separation, fast intervention responses, and the separation of neural change from vascular transfer state / CVR are not directly visible.
whole-brain spatial transcriptomics / cell atlas Ex vivo molecular labels and spatial arrangement. You can enhance the cell-type taxonomy, distribution by region, and molecular map on the structural atlas. Current firing rules, synaptic efficiency, neuromodulatory/glial state, and sleep-history are not directly determined.
Patch-seq / morpho-electric-transcriptomics Compatible with single cell transcriptome, morphology, and electrophysiology. It bridges the cell-type label and morpho-electric phenotype, and strengthens the parameter prior of the local circuit. Whole-brain coverage, same-brain circuit context, current network state, and longitudinal plastic history remain.
ECoG / SEEG Local field potential of the indwelling site and intervention response. In the covered region, we proceed to local dynamics, stimulus response, and relatively strong causal verification. Unindwelled areas, whole-brain coverage, and long-term state completeness remain.
connectomics + same-brain function Correspondence between structural scaffold and functional response in the same brain. Wiring rules for local circuits, cell-type dependent connections, and local correspondence between structure and function can be strengthened. Current synaptic efficiency, intrinsic excitability/homeostatic set point, neuromodulatory fields, glial/metabolic state, and longitudinal plastic history are still separate variables.
local transmitter sensor / axon imaging Local transmitter-linked fluorescence or cholinergic / neuromodulatory axon activity in a named region. Calibrates what mixed arousal proxies do and do not mean, and constrains region-specific chemical timing under an explicit local observable. Whole-brain transmitter field, unsampled receptor families, and same-subject human maintenance-state remain unresolved.
local astrocyte imaging / ensemble readout Astrocyte Ca2+ dynamics, network-state response, or labeled ensemble activity in a local preparation. Strengthens slow-state omission audits and local transmitter-to-astrocyte coupling claims without pretending that glia are background noise. Human whole-brain astrocyte-state, content-general controller identity, and cross-day maintenance completeness remain unresolved.
human receptor / transporter atlas + occupancy PET + displacement / release PET Regional receptor / transporter density priors, administered-drug target engagement, or challenge-linked release proxies. Constrains chemoarchitectural prior, selected receptor-target engagement, or bounded endogenous release proxy for named receptor families. Instantaneous whole-brain transmitter field, unsampled receptor families, and cell-specific downstream effect remain unresolved.
human astrocyte-related PET: MAO-B target-validation route Tracer-family-specific MAO-B binding in healthy humans under named blocker or kinetic-model assumptions. Establishes that a named MAO-B tracer family has human brain entry, selective binding, and a bounded astrocyte-related route role. Disease contrast, tracer-family transfer, covariate ceiling, local controller identity, and same-subject whole-brain maintenance-state remain unresolved.
human astrocyte-related PET: MAO-B disease-context route Named-tracer MAO-B contrast across a declared cohort, disease stage, and reference-model choice. Constrains a bounded disease- and cohort-conditioned astrocyte-related support-state contrast for that specific tracer family. Route-general quantification, cross-cohort generalization, smoking / severity ceiling, local controller identity, and same-subject whole-brain maintenance-state remain unresolved.
human astrocyte-related PET: MAO-B quantification / covariate route Arterial or simplified MAO-B quantification behavior plus cohort- or smoking-conditioned binding shift for a named tracer family. Constrains scan-window / reference-region / kinetic-model choice and covariate ceilings instead of silently promoting the family to one generic meter. Other tracer families, local controller identity, and same-subject whole-brain maintenance-state remain unresolved.
human astrocyte-related PET: I2BS route Tracer-specific I2BS binding under named pharmacologic and kinetic-model assumptions. Constrains a different target-defined astrocyte-related proxy family rather than a drop-in replacement for MAO-B. Interchangeability with MAO-B routes, disease-general transfer, local controller identity, and same-subject whole-brain maintenance-state remain unresolved.
human clearance-support proxy CSF mobility or biomarker-efflux proxy under named device and model assumptions. Constrains macro support-state burden and blocks omission of human clearance routes from human arguments. Local immune-controller identity, responsible synapse, and same-subject whole-brain maintenance-state remain unresolved.
2026-03-25 addendum: transmitter, astrocyte, receptor-PET, and clearance routes are not one stack

The remaining weakness at M1 was that one row still let local transmitter sensors, astrocyte-state readouts, human receptor / transporter PET routes, and human astrocyte / clearance proxies sound closer than they are. The primary literature does not support that compression. Reimer et al. (2016) showed that pupil fluctuations track both adrenergic and cholinergic activity rather than one transmitter, Neyhart et al. (2024) showed that local cortical acetylcholine depends on cholinergic axon activity and local clearance kinetics, Hansen et al. (2022) built a regional receptor / transporter atlas, Wong et al. (2013) and Koepp et al. (1998) constrain occupancy and challenge-limited release, Cahill et al. (2024) showed minute-scale cortical astrocyte-network encoding, Villemagne et al. (2022) established a first-in-human SMBT-1 MAO-B route, Villemagne et al. (2022) then measured reactive astrogliosis across the AD spectrum within that tracer family, Hiraoka et al. (2025) showed that SMBT-1 quantification still depends on named scan-window / reference-region choices relative to kinetic modeling, Mesfin et al. (2026) added a separate whole-body biodistribution burden for the same tracer family, Matsuoka et al. (2026) showed that 11C-SL25.1188 in AD depends on its own simplified quantification route, Tyacke et al. (2018) plus Livingston et al. (2022) showed that a human I2BS route is a different target class with region- and stage-dependent behavior, Best et al. (2026) showed that MAO-B binding can shift with cohort severity and daily cigarette use, Jaisa-Aad et al. (2024) showed in postmortem tissue that MAO-B is promising but still bounded as a reactive-astrogliosis biomarker, and Hirschler et al. (2025) plus Dagum et al. (2026) constrain human CSF mobility and sleep-linked biomarker efflux. These routes do not share the same direct observable, target class, tracer family, quantity type, or safe claim ceiling. On this roadmap, they therefore stay separated into local chemical calibration, local astrocyte-state readout, human neuromodulatory PET prior / engagement, human MAO-B target-validation / disease-context / quantification routes, human I2BS astrocyte-related PET, and human support-state proxy rows.

2026-04-01 addendum: astrocyte PET family labels are not one roadmap row

The remaining weakness after the earlier split was internal to the astrocyte-related PET family itself. Villemagne et al. (2022) answered the first-in-human target-validation question for SMBT-1, Villemagne et al. (2022) answered an AD-spectrum disease-context question, Hiraoka et al. (2025) answered a tracer-family quantification question, Mesfin et al. (2026) answered a whole-body biodistribution and operational-burden question, Matsuoka et al. (2026) answered a simplified SL25.1188 quantification question in AD, Tyacke et al. (2018) answered a first-in-human I2BS target question, and Livingston et al. (2022) plus Best et al. (2026) showed that cognitively impaired and AUD cohorts add stage, severity, and smoking ceilings rather than one route-free astrocyte scalar. In parallel, Jaisa-Aad et al. (2024) showed in postmortem tissue that MAO-B is a promising but bounded biomarker of reactive astrogliosis rather than a generic astrocyte-state meter. Therefore this roadmap now forces readers to keep target class, tracer family, route role, quantification choice, and cohort / covariate ceiling separate before promoting any human astrocyte-related PET result.

2026-03 Supplement: multimodal is not a synonym for "I could see everything"

Mikulan et al. (2020), Seeber et al. (2019), Unnwongse et al. (2023), and Hao et al. (2025) moved forward what can be audited by combining HD-EEG with intracranial ground truth. By contrast, Yao et al. (2023)'s whole-brain spatial atlas, the Patch-seq bridges of Gouwens et al. (2021) and Gamlin et al. (2025), the connectomics of Dorkenwald et al. (2024) and MICrONS Consortium et al. (2025), local transmitter and astrocyte readouts such as Neyhart et al. (2024) and Cahill et al. (2024), and human proxy routes such as Hansen et al. (2022), Villemagne et al. (2022, SMBT-1 target validation / AD-spectrum), Hiraoka et al. (2025), Mesfin et al. (2026), Matsuoka et al. (2026), Tyacke et al. (2018), Livingston et al. (2022), Hirschler et al. (2025), and Dagum et al. (2026) each strengthen different variables while leaving different latent states unresolved. Therefore, this page distinguishes clearly between increasing the number of observations and sufficiently identifying the state variables required for WBE. For the stack-by-stack layout, see Wiki: observability and claim ceiling for each measurement stack.

2026-03-18 addendum: hemodynamic proxy also has a vascular transfer ceiling

For hemodynamic modalities, the remaining uncertainty is not only on the neural side. Murphy et al. (2011) showed that inter-subject CBF / CBV differences contribute to BOLD reactivity, Williams et al. (2023) showed that task BOLD magnitude corresponds strongly to CVR across multiple cortical regions, Wu et al. (2023) showed that baseline CBF partly explains age-related components of BOLD responses, and Epp et al. (2025) showed that significant task BOLD changes can oppose oxygen-metabolism changes across a large fraction of cortex. Therefore, this roadmap does not allow a raw hemodynamic amplitude difference to be promoted directly to a neural difference without a vascular-state / CVR calibration route or explicit abstention.

Next required:Same task/same individual, multimodal simultaneous measurement (possible range) + positioning (M5)

M2 Lower limit on temporal resolution: which time scales should be “saved”? RESOLUTION

Question:What is the necessary temporal resolution in order not to destroy phenomena related to cognition, learning, and consciousness?

  • Candidates:ms (spikes/phase), 10-100ms (coupling dynamics), seconds (state transitions), minutes to days (learning/plasticity)
  • Conditions for disproving:The time resolution is too coarse, making it impossible to reproduce the intervention response or closed-loop control.
  • Next: Measure the acceptable delay in the closed-loop task (I1) and back-calculate the required sampling
There is no single closed-loop time requirement

In the re-audit in March 2026, we decided that closed-loop timing requirements should be handled separately for each loop class. Slow neurofeedback, ERP/command BCI, phase-locked stimulation, and adaptive DBS have different dominant time scales. Wilson et al. (2010) showed that it is necessary to actually measure the latency/jitter of the entire system using hardware, and Belinskaia et al. (2020) reported that an additional 250 / 500 ms delay in alpha neurofeedback worsens learning. On the other hand, in the phase-targeting systems of Mansouri et al. (2018) and Zrenner et al. (2018), the delay should be evaluated as the phase error with respect to the target frequency, rather than the ms value itself.

However, the deeper point is that biological timing is a separate audit from device latency. Seidl et al. (2015) showed that node and internode geometry can be tuned to adjust action-potential timing, Dutta et al. (2018) showed that perinodal astrocytes can reversibly alter conduction velocity, Cohen et al. (2020) showed that saltatory conduction depends on a periaxonal nanocircuit, and Dubey et al. (2022) showed that myelination loss in PV axons disrupts fast inhibition and gamma rhythms. A loop can therefore be hardware-fast while still leaving a timing-state latent on the tissue side.

The phase-targeting case now needs one more split on this page. Zrenner et al. (2020) showed that phase estimability itself worsens when oscillatory amplitude and SNR are low, Gordon et al. (2021) improved prefrontal theta targeting by adding amplitude and stability constraints together with a post-hoc benchmark, Kim et al. (2023) showed across 11 public datasets and 484 participants that prediction accuracy mostly tracks power and SNR, Vigué-Guix et al. (2022) achieved reliable trial-to-trial alpha phase locking without a consistent reaction-time benefit, and Hougland et al. (2025) showed within-session fluctuation and low test-retest reliability of the optimal mu phase. Therefore, this roadmap no longer treats phase-targeted control as one timing number. It separates oscillation estimability, causal-versus-post-hoc benchmark, circular targeting precision, downstream effect, and phase stability / adaptation policy.

  • Do not fix the common gate to 1 ms:Measure the end-to-end median, P95/P99, and worst-case for each loop type.
  • Audit device and tissue separately:Hardware latency/jitter does not prove that biological timing-state has been recovered.
  • For phase-targeting, log the estimability gate:Target band and spatial filter, spectral peak or power/SNR threshold, no-stim rate, and any phase-reset rejection rule.
  • For phase-targeting, keep four extra logs beyond mean phase error:Causal-versus-post-hoc benchmark, circular spread, off-target or random-phase comparator, and fixed-versus-adaptive phase policy.
  • Separate the synchronization system by path: Separately record where LSL/TTL/photodiode/loopback was measured.

Therefore, in this roadmap, we first fix ``which loop class is handled and which timing failure mode is audited'' rather than ``judging a completely closed loop using a single threshold.'' For information on organizing entrances, see Wiki: Closed-loop, delay, jitter, and safety stops.

M3 Lower limit of spatial resolution: Which granularity (region/column/neuron/synapse) should we aim for? RESOLUTION

Question: At what spatial granularity do you assume that the information necessary for “the person's mental functions” resides?

Branch (example):(A) Inter-area connections + states are sufficient / (B) Local circuits are necessary / (C) Synaptic connections are necessary

Primary literature strongly supports: Human EEG / MEG / fMRI obtained non-invasively are basically macro proxies. EEG / MEG capture globally synchronized current fields, while fMRI captures hemodynamics, so they do not directly observe the state of each neuron or synapse. Conversely, Dorkenwald et al. (2024) and MICrONS Consortium et al. (2025) strengthen the link between high-density connectomics and same-brain function in model organisms and local cortex, but they do not justify claims of “state-complete measurement of the whole human brain.”

  • Upper limit of non-invasive human measurement: Tracking of region-to-network level state transitions, relatively global timing constraints, and condition differences.
  • Upper bound on the connectomics front:This is an audit of structural scaffolds, sequential same-brain local scaffolds, stimulus-conditioned conditional predictors, and cell-type dependent connections, not a complete observation of current synaptic efficiency and neuromodulation fields.
  • Remaining latent states:Synaptic weight, receptor state, delay / timing-state, intrinsic excitability / homeostatic set point, neuromodulatory field, glial / metabolic state, and plastic history remain as separate variables.
2026-03-27 addendum: three per-row axes are still not enough without quantity / target / transport splits, cohort, and robustness gates

The remaining weakness in this roadmap was that M3 still let very different human-side advances sound like one generic increase in observability and still let family names stand in for one route. The primary literature does not support that compression. Shapson-Coe et al. (2024) reconstructed a cubic millimeter of surgically resected human temporal cortex at nanoscale resolution, which is a local ex vivo structural scaffold. Johansen et al. (2024) built an in vivo SV2A atlas in healthy humans (17F/16M) calibrated with postmortem autoradiography, which is a regional synaptic-density proxy. Hansen et al. (2022) aggregated receptor / transporter PET into a regional chemoarchitectural prior, while Wong et al. (2013) constrains selected occupancy / target engagement under an administered drug and Koepp et al. (1998), Lippert et al. (2019), plus Erritzoe et al. (2020) constrain challenge-linked dopamine / serotonin release proxies rather than a free readout of the whole transmitter field. Lucchetti et al. (2025) derived a within-subject metabolic similarity matrix from five 1H-MRSI metabolites in 51 healthy participants with replication in 13, whereas Ren et al. (2015) measured ATP synthesis, phosphorus metabolites, and intra-/extracellular pH in 12 resting human brains with 31P-MRS, Ren et al. (2017) measured PCr→γ-ATP and Pi→γ-ATP exchange flux in six subjects using three band-inversion modules plus a five-pool Bloch-McConnell-Solomon model, Guo et al. (2024) mapped whole-brain intracellular NAD content at 7 T, and Kaiser et al. (2026) measured task-evoked NAD+ dynamics in a functionally localized occipital voxel. The deuterium routes split again inside that family: Karkouri et al. (2026) implemented a 7 T absolute-quantification route for HDO / Glc / Glx / Lac maps under a dedicated quantification pipeline, whereas Li et al. (2025) used 7 T dynamic DMRSI, blood-input measurements, and an explicit kinetic model to image metabolic-rate terms. Even that is still not route-free: Ahmadian et al. (2025) showed that human-brain DMI signal depends materially on the administered [6,6'-2H2]glucose dose, and Bøgh et al. (2024) showed that repeatability at 3 T depends on a named acquisition and time-point regime, with the best repeatability at 120 min in that protocol. van Blooijs et al. (2023) show that living-human timing-support already splits at the route-family level, because tract-scale transmission-speed estimation is not the same quantity as myelin-sensitive contrast. Hagiwara et al. (2018), Baadsvik et al. (2024), and Galbusera et al. (2025) show that human myelin MRI is already a quantity-defined family rather than one interchangeable meter, while Colaes et al. (2026) show that T1w/FLAIR may remain a broader tissue-health-sensitive ratio rather than a myelin-specific readout. Morgan et al. (2024) and Padrela et al. (2025) measure BBB water exchange with ASL and expose method- and model-dependent burdens, while Chung et al. (2025) measures tracer-specific BBB permeability-surface-area products with high-temporal-resolution dynamic PET across three radiotracers. Villemagne et al. (2022) established a first-in-human SMBT-1 MAO-B route with pharmacological blockade, Villemagne et al. (2022) then measured reactive astrogliosis across the AD continuum within that tracer family, Hiraoka et al. (2025) showed that even SMBT-1 depends on scan-window / reference-region versus kinetic-model choices, and Tyacke et al. (2018), Livingston et al. (2022), Jaisa-Aad et al. (2024), plus Best et al. (2026) show that target class, region / stage pattern, tissue expression, and cohort covariates still matter. Meanwhile, Hirschler et al. (2025) plus Dagum et al. (2026) advanced human-side CSF mobility and sleep-linked clearance routes, but still through specialized MRI or an investigational device plus multicompartment model. Therefore, this roadmap now asks readers to label not only (1) proxy class, (2) operational maturity / acquisition burden, and (3) safe calibrator role, but also the family-internal quantity / target / transport definition before promoting any human in vivo claim.

One more split is necessary inside the human maintenance-state family. Tallman et al. (2025) add a local human clinical single-unit allocation route in epilepsy patients, but with firing explicitly treated as only an indirect excitability index and with no whole-brain coverage. Huber et al. (2013), Kuhn et al. (2016), and Fehér et al. (2026) instead define sleep-homeostasis / plasticity-recalibration routes through TMS-EEG or PAS outcomes, while Zrenner et al. (2018) plus Khatri et al. (2025) define state-gated perturbation routes whose direct observables remain bounded plasticity-assay or corticospinal responses. Therefore, this roadmap no longer lets human excitability evidence stand in for one row: local clinical-unit allocation and noninvasive perturbation-conditioned routes now stay separate.

Human route Proxy class Operational maturity / burden Safe calibrator role on this page Why the ceiling remains
Local human ultrastructure
Shapson-Coe et al. (2024)
Local ex vivo structural scaffold Destructive surgical-fragment route; not repeatable living-human acquisition Structural scaffold only Ex vivo fragment from one surgical sample, not living whole-brain current state
SV2A PET atlas
Johansen et al. (2024)
Regional synaptic-density proxy Healthy-cohort atlas; tracer and quantification dependent Bounded synaptic-density prior Tracer-defined atlas; not momentary synaptic efficacy or cell-specific plasticity state
Receptor / transporter atlas
Hansen et al. (2022)
Regional chemoarchitectural prior Aggregated PET atlas across many healthy individuals; receptor-family limited Bounded neuromodulatory-family prior Does not read out current transmitter release, occupancy, or whole-brain neuromodulatory state
Occupancy PET
Wong et al. (2013)
Ligand- and dose-limited target-engagement proxy Administered-drug, receptor-family, and bounded scan-window design Selected receptor-target engagement Does not recover endogenous transmitter release, the general transmitter field, unsampled receptor families, or cell-specific downstream effect
Displacement / release-sensitive PET
Koepp et al. (1998); Lippert et al. (2019); Erritzoe et al. (2020)
Challenge-linked endogenous release proxy Task-, drug-, ligand-, receptor-family-, kinetic-model-, and window-specific design Bounded dopamine / serotonin release proxy under a named challenge Does not recover a task-general transmitter field, unsampled receptor families, or cell-specific downstream effect outside the measured window
1H-MRSI metabolic connectome
Lucchetti et al. (2025)
Parcel-level biochemical similarity scaffold 51 healthy participants with 13-person replication; QC and parceling dependent Bounded biochemical scaffold Five-metabolite similarity matrix, not high-resolution metabolite distribution, not energetic balance, and not kinetic rate imaging
Human 31P-MRS energetic balance
Ren et al. (2015)
Macro 31P energetic-balance proxy 7 T resting-brain spectroscopy in 12 subjects; phosphorus-specific acquisition and fitting burden Bounded energetic-balance proxy ATP synthesis / phosphorus metabolite / pH balance, not branch-local mitochondrial controller state
Human 31P MT exchange-flux route
Ren et al. (2017)
Model-conditioned macro 31P MT exchange-flux proxy 7 T band-inversion transfer route in 6 subjects with three ~8 min modules and five-pool modeling burden Bounded 31P exchange-flux proxy PCr→γ-ATP and Pi→γ-ATP exchange-flux under a named MT model, not resting energetic balance, NAD mapping, or branch-local mitochondrial controller state
Human 31P NAD-content mapping
Guo et al. (2024)
Macro 31P NAD-content map proxy 7 T whole-brain route with subspace-based denoising, low-concentration spectral fitting, and long acquisition burden Bounded whole-brain NAD-content map proxy Not task-evoked local NAD dynamics, not branch-local mitochondrial controller state, and not whole-brain moment-to-moment redox control
Human 31P functional NAD dynamics
Kaiser et al. (2026)
Localized functional 31P NAD-dynamics proxy 7 T task fMRS in 25 volunteers with prior fMRI localization and one occipital voxel Bounded task-locked local NAD-dynamics proxy Not whole-brain NAD-content mapping, not task-general energetic-controller identity, and not branch-local mitochondrial controller state
Deuterium metabolite-mapping / absolute-quantification route
Karkouri et al. (2026)
Macro deuterium metabolite-mapping / absolute-quantification proxy 7 T specialized route with absolute-quantification pipeline, dedicated hardware, limited cohorts, and route-specific dose / timing burden Bounded deuterium metabolite-mapping proxy Not kinetic-rate imaging, not branch-local ATP reserve, and not cell-specific bioenergetic controller state
Deuterium kinetic-rate imaging
Li et al. (2025)
Model-conditioned macro deuterium kinetic-rate proxy 7 T dynamic DMRSI with blood-input, explicit kinetic model, very small cohort, and route-specific dose / timing burden Bounded deuterium kinetic-rate proxy Not one general biochemical meter, not branch-local ATP reserve, and not cell-specific bioenergetic controller state
Human tract-scale transmission-speed estimation
van Blooijs et al. (2023)
Tract-scale timing-support proxy Development-sensitive transfer-time / tract-geometry human route rather than myelin-specific quantification or per-axon measurement Bounded tract-scale timing-support proxy Not a myelin-specific quantity, not node / internode microgeometry, and not a per-axon conduction controller readout
Quantity-defined myelin MRI / tissue-health-sensitive ratio family
Hagiwara et al. (2018); Baadsvik et al. (2024); Galbusera et al. (2025); Colaes et al. (2026)
Quantity-defined macro myelin proxy family plus a tissue-health-sensitive ratio route Metric-dependent validity; bilayer mapping remains proof-of-principle, remyelination calibration remains special-case, and T1w/FLAIR may remain broader than myelin Bounded myelin / timing-support or broader tissue-health family, depending on route MTsat, SyMRI, T1w/T2w, qT1, bilayer, and T1w/FLAIR routes do not share one interchangeable claim ceiling, one common quantity, do not collapse into tract-speed estimation, and do not provide a per-axon timing readout
Human clinical single-unit allocation route
Tallman et al. (2025)
Local clinical-unit allocation-related proxy Implanted-human hippocampal single-unit route in epilepsy patients; high-locality and pathology-conditioned Bounded local human allocation-related calibrator Not a whole-brain excitability meter, not direct AIS / channel-state readout, and not a route-free separation of pre-existing versus learning-induced excitability
Human sleep-homeostasis / plasticity route
Huber et al. (2013); Kuhn et al. (2016); Fehér et al. (2026)
Perturbation-conditioned maintenance proxy Noninvasive human TMS-EEG / PAS route with sleep-history manipulation burden Bounded excitability / plasticity-support proxy Does not identify which cell class, synapse, AIS / channel change, or recovery controller produced the change
Human state-gated perturbation route
Zrenner et al. (2018); Khatri et al. (2025)
State-gated perturbation proxy Closed-loop stimulation route with state-estimator and stimulation-timing burden Bounded state-gated excitability / corticospinal proxy Does not measure excitability mechanism itself, AIS geometry, channel distribution, or long-horizon recovery control
Human BBB water-exchange MRI
Morgan et al. (2024); Padrela et al. (2025)
Macro BBB water-exchange proxy ASL method- and model-dependent route with healthy-adult / lifespan burden Bounded BBB water-exchange proxy Water-exchange estimates are not tracer-specific transport and do not read out pericyte / endothelial controller identity
Tracer-specific BBB PET transport
Chung et al. (2025)
Tracer-specific BBB transport proxy High-temporal-resolution dynamic PET plus kinetic modeling across three radiotracers Bounded tracer-specific BBB transport proxy Named tracer PS estimate, not one generic BBB permeability scalar or local controller readout
Human blood-CSF barrier / choroid-plexus perfusion / transport family
Zhao et al. (2020); Sun et al. (2024); Petitclerc et al. (2021, 2026); Anderson et al. (2022); Wu et al. (2026)
Choroid-plexus perfusion, blood-to-CSF water transport, choroid-plexus water cycling, apparent BCSFB exchange, or joint BBB-versus-BCSFB ASL exchange ASL / ultra-long-TE ASL / DCE-MRI / REXI route family with route-specific model burden, validation ceiling, and small-cohort dependence outside the HCP-Aging perfusion analysis Bounded BCSFB / choroid-plexus support-state proxy family Perfusion is not blood-to-CSF transport, water cycling is not apparent exchange, and none of these rows identify a generic BBB scalar, a generic clearance truth, or choroid-plexus epithelial controller identity
Human MAO-B PET: SMBT-1 first-in-human target validation
Villemagne et al. (2022)
Tracer-family-specific MAO-B target-validation proxy Dynamic human PET with selegiline block and reversible-binding readout Bounded MAO-B target-validation route Shows that SMBT-1 can read a human MAO-B route, not that MAO-B disease contrast, quantification transfer, or generic astrocyte-state are solved
Human MAO-B PET: SMBT-1 AD-spectrum contrast
Villemagne et al. (2022)
Pathology-context MAO-B contrast proxy AD-continuum cohort route with subcortical-white-matter reference and cross-sectional cognitive / amyloid / tau comparison Bounded AD-spectrum astrocyte-related support-state proxy Not a route-free MAO-B meter, not generalizable to all diseases, and not a local controller readout
Human MAO-B PET: SMBT-1 quantification / biodistribution burden
Hiraoka et al. (2025); Mesfin et al. (2026)
Tracer-family quantification and biodistribution route Quantification depends on named scan-window / reference choice or arterial modeling, while whole-body biodistribution adds organ-distribution and late-gallbladder burden Bounded SMBT-1 acquisition / quantification route Not interchangeable with disease contrast, not one generic MAO-B number, and not proof that other tracer families or cohorts share the same ceiling
Human MAO-B PET: SL25.1188 disease / severity route
Matsuoka et al. (2026); Best et al. (2026)
Tracer-family-specific MAO-B disease / covariate proxy AD quantification without arterial sampling and AUD severity / smoking-conditioned binding route under named tracer and model assumptions Bounded SL25.1188 disease- and covariate-conditioned proxy Not a generic MAO-B readout across diseases, not covariate-free, and not interchangeable with SMBT-1 results
Human I2BS astrocyte-related PET
Tyacke et al. (2018); Livingston et al. (2022)
I2BS target-defined astrocyte-related proxy First-in-human idazoxan-blocked PET plus cognitively impaired cohort route with region- and stage-dependent behavior Bounded I2BS astrocyte-related proxy Different target class from MAO-B, not drop-in comparable with MAO-B tracer families, and not a generic astrocyte-state meter
Human clearance-support route families
Fultz et al. (2019); Kim, Huang, & Liu (2025); Lim et al. (2025); Yoo et al. (2025); Eide et al. (2023); Hirschler et al. (2025); Dagum et al. (2026)
Macro support-state proxy family Sleep-linked CSF oscillation, parenchyma-CSF water exchange, respiration-conditioned net flow, exercise-conditioned contrast influx / meningeal-lymphatic flow, intrathecal tracer / CSF-to-blood clearance, specialized CSF-mobility MRI, or device- and model-heavy overnight biomarker-efflux inference Bounded clearance-support proxy family These are different clearance-transport rows with different observables, intervention regimes, crossed boundaries, and model burdens, not a route-free glymphatic meter or direct readout of cell-specific immune control, responsible synapse, or local maintenance controller

However, the per-row axes are still not enough when several human rows are combined in one argument. Lucchetti et al. (2025) derive the main 1H-MRSI route from adolescents with a separate 15-35-year replication cohort, Morgan et al. (2024) and Padrela et al. (2025) keep BBB water-exchange claims inside ASL method choices and healthy-adult cohorts, Hirschler et al. (2025) report a 7 T healthy-younger-adult CSF-mobility route, Dagum et al. (2026) study healthy older adults in a randomized sleep-deprivation crossover, Villemagne et al. (2022) remains a healthy-volunteer first-in-human SMBT-1 target-validation route, Villemagne et al. (2022) remains a pathology-context MAO-B astrocyte-related PET route across the AD continuum, Hiraoka et al. (2025) shows that the SMBT-1 route still depends on scan-window / reference-region versus kinetic-model choice, Mesfin et al. (2026) keeps whole-body biodistribution as a separate operational burden, Matsuoka et al. (2026) adds an arterial-free SL25.1188 AD quantification route, Best et al. (2026) shows that MAO-B binding can shift with AUD severity and daily cigarette use, Livingston et al. (2022) shows that the I2BS route changes with region and impairment stage, and Galbusera et al. (2025) constrains remyelination sensitivity in postmortem multiple-sclerosis cortex rather than a living-human longitudinal baseline. Those are not one matched human baseline by default. In parallel, Holiga et al. (2018) showed that MRI repeatability ranges from poor to excellent depending on the metric, Wirsich et al. (2021) showed that cross-setup reproducibility has to be demonstrated rather than assumed, Amiri et al. (2023) showed that same-sample multimodal comparison in CONNECT-ME depended on n = 48 patients with all features and improved sensitivity despite a markedly decreased sample size, and Manasova et al. (2026) showed that the main French dataset itself was highly unbalanced across rows (EEG-LG 290, dMRI 151, aMRI 101, FDG-PET 53, fMRI-RS 44), with missing values substituted by -1 while pairwise disagreements rose in MCS and improved patients. Therefore, on this page, bundle promotion now also requires two bundle-level gates: (4) cohort / physiological regime compatibility and (5) row robustness / availability disclosure after any within-family quantity / target / transport split has already been fixed.

  • Cohort / physiological regime compatibility: name age band, health or pathology regime, and why transfer across cohorts is justified for the claimed latent variable.
  • Row robustness / availability disclosure: name route-local repeatability, cross-centre or cross-scanner transfer window if claimed, the row-overlap geometry, and whether missingness tracks site, severity, tolerance, contraindication, protocol, or only a declared missing-row policy.

If several rows are combined in one argument, this roadmap now asks for same-subject / same-session / same-perturbation acquisition, cohort / regime compatibility, row-level repeatability or route-local reproducibility, availability / complete-case disclosure, external calibration, and a Human Proxy Composition Card before talking about state-complete human measurement. If the bridge is sequential across live and ex vivo or across separated state regimes, it also asks for a State-Continuity Bridge Card plus the longer background rule in Wiki: State-Continuity Bridge. Otherwise the result stays at proxy-rich but ceiling-limited human evidence, because proxy class, operational maturity, calibrator role, cohort compatibility, and robustness still remain mismatched. The front-door table is in WBE 101: human observability ladder, the longer critique is in Wiki: Human Proxy Composition and Route Maturity, and the submission-side stop rule is Verification: Observability Budget.

2026-03-20 addendum: destructive ultrastructure is not a shortcut around the observability wall

It would still be too strong to read nanoscale, petascale, or same-brain EM as if the route had already preserved native state and solved scaling by itself. Lu et al. (2023) showed that fixation route materially changes extracellular-space preservation and that even high-pressure freezing remains thickness-limited, Shapson-Coe et al. (2024) showed that human nanoscale reconstruction is still a rapidly preserved local surgical fragment, and MICrONS Consortium et al. (2025) showed that same-brain function plus EM is a sequential local pipeline rather than simultaneous whole-state capture. Therefore, this roadmap now separates preservation route, registration scope, and throughput / proofreading burden from resolution itself. On this page, destructive ultrastructure without those disclosures stays at structural scaffold or local ex vivo scaffold, not state-complete measurement.

Temporary decision rule:If the observation path stays in the macro proxy, this roadmap first keeps the assertion between L1 and weak L2. In order to raise the claim to local circuits and synaptic granularity, we additionally require either structure/function correspondence in the same brain, external ground truth, or intervention response.

Next, we need:For each assumption (A), (B), and (C), evaluate observability (M1), identifiability (R7), and computability (I3) in bundles, and disclose what to abstain at which granularity. For background, see Wiki: Why wiring diagrams are not enough and Wiki: Homeostatic plasticity and maintenance state.

M4 Whole brain coverage: How much of the “whole” do we need to measure? COVERAGE

Question:How do we handle “boundaries (what is included in the subject)” with limited measurements?

2026-03-19 supplement:Subject boundary is not a philosophical garnish. Musall et al. (2019) and Stringer et al. (2019) showed that ongoing behavior is embedded in large fractions of cortical activity, Saleem et al. (2013) and Ravassard et al. (2013) showed that locomotion, optic flow, vestibular, and other sensory cues reshape cortical and hippocampal codes, Zelano et al. (2016) and Raut et al. (2025) showed that respiration and arousal couple brain dynamics to organism-wide physiology, and Flesher et al. (2021) showed that reintroducing tactile feedback improves a local bidirectional BCI. A slower boundary remains as well: de Quervain et al. (1998) and Oei et al. (2007) showed that glucocorticoid state changes retrieval and hippocampal / prefrontal retrieval activity, while McCauley et al. (2020), Barone et al. (2023), and Birnie et al. (2023) showed that circadian timing and corticosteroid rhythm alter hippocampal plasticity machinery. Therefore, M4 now treats body / environment boundary disclosure as a measurement requirement rather than as a later interpretation issue, and that disclosure includes both fast loop routes and slow internal-milieu routes.

Decision rule:If a submission does not name which sensory, motor, interoceptive, and slow internal-milieu routes were preserved, substituted, matched, perturbed, or omitted, the result stays at local controller, task-specific subsystem loop, or brain-side-only proxy and does not advance to subject-complete L3 language.

Next: Operatively place the "boundary of subject" in verification (V0) and publish a boundary card listing retained / substituted channels, slow internal-milieu routes, surrogate body / environment mapping, loop-removal tests, and residual omitted loops.

M5 Multimodal integration: how to guarantee alignment (MRI/EEG/fMRI)? FUSIONMRI

Question:Different measurements have different coordinate systems, delays, and noise structures. Are integration errors destroying the “signal you want to learn”?

Conditions for disproving: Reconstruction (R2) becomes unstable due to alignment error and reproducibility (P2) decreases

2026-03 supplement: For EEG + fMRI / fNIRS, a shared clock is still not enough. Murphy et al. (2011) and Williams et al. (2023) show that vascular reactivity alters the meaning of task BOLD amplitude, Wu et al. (2023) shows that baseline CBF can partially explain age-related BOLD differences, and Yucel et al. (2015) show that fNIRS needs short-separation regression to suppress superficial autonomic contamination. Therefore, M5 now treats vascular-state / CVR calibration or abstention as part of multimodal alignment for hemodynamic stacks.

2026-03-19 supplement: OPM-MEG adds movement tolerance, but not a free pass on measurement conditions. Boto et al. (2018) and Seymour et al. (2021) showed motion-tolerant OP-MEG recordings, Holmes et al. (2023) and Holmes et al. (2023) extended this to ambulatory movement and two-person hyperscanning by coupling field nulling to optical tracking, Holmes et al. (2025) showed that lighter shielded rooms become plausible only with tSSS plus active compensation, and Rhodes et al. (2025) showed that pseudo-MRI can help when MRI is difficult but still leaves individual MRI as the gold standard. Therefore, on this roadmap, OPM-MEG is read as movement-tolerant macro electrophysiology under disclosed magnetic-field control, sensor calibration, and anatomy route, not as shield-free naturalistic brain readout.

Next required:Does the same data match in multiple pipelines (Analysis differential audit)

M6 Intervention/Stimulation: What is the “experimental design” that increases the possibility of identification? CAUSAL

Question: When identification is not possible through passive observation alone, what kind of intervention (stimulus / task / environmental change) can increase identifiability?

Primary literature strongly supports: Intervention is not a magic trick that automatically reveals causality. Identifiability increases only in designs that specify where, when, and how much stimulation was applied, how the artifact window was defined, and how latency / jitter were measured.

  • State-targeted stimulation: As in Zrenner et al. (2018), conditioning the stimulation time on brain state makes response differences easier to detect even for the same nominal stimulus.
  • Interventions with ground-truth:Intracranial stimulation, simultaneous SEEG/ECoG, phantoms, and animal experiments can calibrate estimation errors on the scalp side with external standards.
  • Closed-loop intervention:With BCI and adaptive stimulation, response differences cannot be interpreted as causal differences unless we measure not only mean latency but also jitter and worst-case path, as shown by Wilson et al. (2010).

What does not count as an intervention:Conditional differences in passive tasks, post-hoc cross-modal correlations, and offline analyzes without stimulus logs alone do not count as strong causal evidence on this page.

Externally dependent tasks:TMS/DBS/invasive interventions on real subjects require IRB, equipment, and clinical cooperation. The first thing to do in this repository is to prepare the intervention grammar, synchronization log format, stop conditions in case of failure, and comparison rules.

Next, we need: On the model (R4) side, perform a design back calculation that says ``This intervention will improve the discrimination between families,'' and fix the submission pack that includes where TTL / LSL / photodiode / loopback was measured.

M7 Longitudinal: Are “personal characteristics” stable against intra-day/day-to-day fluctuations? LONGITUDINAL

Question:Even in the same individual, brain states vary across minutes, hours, days, and months. The operational question here is no longer simply “is the person stable?” but which part of the result reflects state fluctuation, trait-like backbone, biological drift, interface / decoder drift, or adaptive rescue.

2026-03-22 addendum:The older wording here was too weak because it still allowed same-day success, cross-day persistence, and algorithmically rescued performance to sound like one continuous longitudinal achievement. The primary literature does not support that compression. Musall et al. (2019) showed that trial-to-trial cortical dynamics are strongly shaped by richly varied movements, Benisty et al. (2024) showed that spontaneous behavior rapidly changes functional-connectivity structure, and Egger et al. (2024) showed 10-hour EEG changes that materially alter decoding and motivate adaptive decoders. By contrast, Gallego et al. (2020) and Noda et al. (2025) show that a population-level backbone can remain more stable than individual units. Finally, Karpowicz et al. (2025), Wilson et al. (2025), and Wairagkar et al. (2025) show why stabilized operation and fixed-decoder durability must be reported separately. Therefore, M7 now treats longitudinal evidence as a multi-field audit, not one score.

Audit field What must be named What this stops you from overreading
state annotation Arousal, spontaneous behavior, movement, task mode, time-of-day, sleep pressure, medication, or stimulation state. A same-day or cross-day difference cannot be read as trait instability if the state itself was left unlogged.
trait-like backbone The named backbone object that remained stable: for example latent dynamics or representational geometry. Do not treat single-unit or single-channel volatility as if it automatically disproved a stable population backbone.
biological drift Plasticity, remapping, unit turnover, recovery, or learning-related change in the living system. Do not collapse organism-side change into interface or preprocessing failure.
interface / decoder drift Electrode shift, channel loss, impedance change, feature-distribution shift, or decoder mismatch. Do not say the biological representation collapsed when the unstable object was the interface or decoder.
fixed decoder interval / recalibration burden How long the decoder was held unchanged, how often rescue or recalibration occurred, what data it used, and what happened when it failed. Do not rephrase adaptive rescue as if the original decoder had remained valid on its own.

Decision rule:If a result does not report state annotation, fixed decoder interval, and recalibration burden separately, this roadmap stops at state-conditioned performance or participant-specific stabilized operation. It does not promote the result to trait stability, durable fixed-decoder use, or long-horizon deployability.

Next:Bundle Verification: Temporal Validity Card with Wiki: State, Trait, and Drift. If the evidence also bridges across acquisition regimes or days as if it were one latent state, add the State-Continuity Bridge Card rather than relying on same-subject wording alone.

M8 Quality control: How to “quantify” and handle artifacts/missing/noise? QC

Question:Define “good data” in advance and make exclusion, complementation, and weighting transparent. In addition, how do we quantify quality deterioration caused by hardware, such as impedance/CMRR, noise floor, HMD interference, etc.?

  • Impedance/CMRR: Detects imbalance and warns before measurement.
  • Noise floor/SNR: Visualize the effective SNR for each frequency band.
  • VR Interference: Profile HMD electromagnetic and mechanical noise.
M9 Standardization: How can BIDS/metadata be used in a format that allows others to retest? BIDS

Question:Even if the data can be shared, it is not possible to retest if the metadata is weak. What should be converted into metadata at least?

Supplement (Issue #46): If VR/motion measurement is included, specify the coordinate system, time synchronization, and device specifications according to Motion-BIDS.

Next: Leave the protocol (task/equipment/preprocessing) machine readable (P2)

2. Reconstruction

Reconfigure: What to restore? (Circuit/state/learning rule)

Central question: "To what extent is structure (coupling), dynamic state (activity), and learning rules (plasticity) necessary for the individual's mental functions?"

R0 What is the minimum set to restore? (Structure/state/learning rule) RECONSTRUCTION

Question:How far do I have to restore to win in verification (V0)?

2026-03-19 literature audit:The older wording here was too weak because it made structure, state, and plasticity sound like three clean knobs. The primary literature does not support that simplification. Gouwens et al. (2021) showed that transcriptomic type still leaves morpho-electric spread within a cell class. Frey & Morris (1997), Govindarajan et al. (2011), Yang et al. (2014), and Pandey et al. (2021) show that long-term stabilization depends on synapse- and branch-specific tagging, clustered spine stabilization, and proteostatic maintenance rather than on graph structure alone. Hengen et al. (2016), Latchoumane et al. (2017), and Terceros et al. (2026) show that multi-day retention also depends on sleep / homeostatic / transcriptional programs. Gibson et al. (2014), Dutta et al. (2018), Cohen et al. (2020), Suzuki et al. (2011), and Pizzorusso et al. (2002) show that myelin / perinodal state, glial metabolic support, and extracellular-matrix state can change timing, plasticity, and memory expression without rewiring. Therefore, this roadmap no longer treats structure + state as one dial.

Revised branching:(A) structural scaffold only / (B) structural scaffold + fast executable state / (C) structural scaffold + fast executable state + maintenance / plasticity state

Layer inside the restoration target What has to be named explicitly What fails if it is omitted Claim ceiling if still omitted
Structural scaffold Cell identity, gross morphology, connection graph, major axonal routing constraints Local routing, lesion/perturbation path constraints, and circuit-level intervention matching have no substrate to run on Structural prior only, not an executable person model
Fast executable state Momentary activity state, effective synaptic efficacy, intrinsic excitability / operating point, and current neuromodulatory regime (R10) The same scaffold can express different immediate trajectories and intervention responses Structure-conditioned simulator or same-session proxy, not a demonstrated state match
Maintenance / eligibility state Synaptic tags and capture conditions, local proteostatic support, sleep/replay coupling, transcriptional stabilization, myelin / perinodal state, glial metabolic support, ECM / PNN state Learning persistence, day-to-day retention, timing stability, and recovery after perturbation become underdetermined Short-window replay only; no claim here about stable multiday learning or continuity
Restoration-target card required

Every R0 claim on this site must publish a small card that states (1) which state families were directly measured, (2) which were only externally calibrated or model-estimated, (3) which were omitted, (4) which timescale each family is supposed to cover (milliseconds / seconds / hours / days / weeks), and (5) what claim ceiling and abstention boundary follow from those omissions. Without that card, same-session fit and multiday continuity collapse into one vague success label, which this roadmap no longer allows.

Failure condition:If the chosen restoration target cannot reproduce intervention response at the declared timescale, or if the conclusion changes once omitted maintenance-state families are reintroduced, the target fails. A model that fits same-session data while hiding multiday maintenance burdens is not accepted here as a general restore target.

Next:Evaluate identifiability (R7) and computability (I3) simultaneously, but only after publishing the restoration-target card with omitted state families, audited timescales, and claim ceiling. For supporting background, use Wiki: Why wiring diagrams are not enough and Wiki: Homeostatic plasticity and maintenance state.

R1 Inverse problem: What can be uniquely inferred from observation (M1)? INVERSEBAYES

Question: An inverse problem such as EEG source estimation is “ill-posed,” meaning the solution is not uniquely determined. Therefore, it is more important to audit how stable the conclusion remains when geometry, conductivity, electrode coordinates, and prior assumptions change than to ask which solver was chosen.

Policy: In this project, no specific solver is fixed as the standard solution. Acceptance requires (1) an explicit uncertainty expression such as a posterior distribution, confidence interval, or bootstrap / ensemble range, (2) sensitivity analysis for individual MRI, FEM / BEM, conductivity assumptions, and electrode geometry, (3) external validation using simulation, phantoms, simultaneous invasive recording, or intracranial stimulation, and now also (4) a named source regime / target object, uncertainty object, and validation board / operating regime. Bayesian or empirical-Bayes approaches are good candidates, but auditability matters more than the method name.

2026-03-31 addendum:The remaining weakness in this roadmap was to let inverse-family progress sound closer than it is to one shared truth scale. The current primary literature does not support that shortcut. Luria et al. (2024) sample posterior support for focal-source hypotheses, Tong et al. (2025) target sparse spatial-temporal source imaging with debiased estimation and inference, and Feng et al. (2025) target extended-source extent with empirical-Bayesian uncertainty maps. Upstream uncertainty remains separate: Vorwerk et al. (2024) showed that tissue-conductivity uncertainty materially shifts EEG source analysis, and Vorwerk et al. (2026) showed that conductivity estimation can reduce source-localization uncertainty in many presurgical epilepsy scenarios without erasing all depth / location exceptions. Therefore, this roadmap no longer reads new inverse family as one generic upgrade. It asks what object is being estimated, which uncertainty object is returned, and which physics and validation regime remain attached.

R1 inverse-audit field What this roadmap now asks for Why it matters
Source regime / target object Name whether the paper is estimating focal support, a sparse spatial-temporal source set, extended-source extent, or another declared object. Different inverse families are not automatically estimating the same scientific target.
Uncertainty object Name whether the method returns posterior support, debiased intervals / tests, extent-overlap maps, or only location spread. Uncertainty is not one scalar if the returned object changes with the inverse family.
Forward-model uncertainty route State whether conductivity, head-model, and electrode-geometry uncertainty were propagated, fixed, or calibrated. A sharper posterior does not erase uncertainty that still sits upstream in the physics.
Validation board / operating regime Name whether the evidence comes from simulation, phantom, DBS / stimulation ground truth, simultaneous invasive recording, or another board. Success on one board does not automatically transfer to all source classes and clinical / experimental regimes.

Next, we need: In R2 and R7, isolate how much source depth, cranial conductivity, and candidate model space dominate the error, and disclose abstention conditions rather than forcibly collapsing overlapping solutions into one.

R2 Source estimation: “Which representation” of brain activity do you want from EEG? EEGESIBAYES

Question: EEG source estimation is easy to obtain an expression of "area x time", but neurons/synapses cannot be directly seen. Select the expression according to the restoration target (R0).

Issue:In the past, "high-density EEG (128 channels)" was recommended, but recent research has shown that ill-posedness cannot be solved by the number of channels alone. In particular, point estimation methods such as dSPM ignore uncertainty.

Stricter policy:Estimation with uncertainty such as the Block-Champagne system, high-density EEG, and FEM/BEM forward model based on individual MRI are effective means to improve estimation conditions. However, these do not guarantee the accuracy ofdeep brain activity. The hiring condition is not that it can be said to be able to see deeply, but rather that it is able to audit which error sources have been reduced and to what extent. Therefore, submissions must include (1) posterior distributions or confidence intervals, (2) sensitivity analysis of conductivity, electrode coordinates, and head model, (3) validation against external criteria such as simulations/phantoms/concurrent invasive recordings/intracranial stimulation, and (4) conditions for withholding claims regarding deep sources.

2026-03-31 addendum:The older wording still made better source imaging sound too much like one shared representation goal. The current primary literature does not support that shortcut. Pascarella et al. (2023) compare focal-source validation routes, Unnwongse et al. (2023) show depth- and skull-conductivity-dependent localization error across 3,619 stimulation locations in 11 patients, and Hao et al. (2025) show simultaneous HD-EEG / SEEG concordance in a drug-resistant-epilepsy operating regime. Those boards do not yet define one universal scoreboard for focal, sparse, extended, and spontaneous sources. Therefore, R2 now asks the restoration target to say whether it wants focal support, sparse spatial-temporal source activity, extended-source extent, or another declared object before methods are compared.

R2 representation family What it can return relatively directly What this roadmap still refuses to infer automatically
Probabilistic focal-source family Posterior support and alternative configurations for a small-number focal-source hypothesis. It is not promoted automatically to sparse spatial-temporal inference or extended-source extent recovery.
Sparse spatial-temporal family Bias-corrected amplitudes, orientations, depths, and inference for sparse source sets over time. It is not promoted automatically to focal-board optimum across all priors or to extended-source validation.
Extended-source family Estimated location-plus-extent with uncertainty for spatially extended activation. It is not promoted automatically to focal-source best case or to a universal benchmark win across source regimes.

Next:Firstly, fix source imaging verification specifications and benchmarks with external standards without making it unclear whether or not they are publicly implemented. The implementation will then be made public in a form that allows for reproduction, including posterior distribution, sensitivity analysis, and failure examples.

R3 State estimation: How to define latent state and how to verify it? STATE

Question:Can we separate observation noise from the true state and show that the state is “effective for prediction”?

2026-03-25 addendum:The remaining weakness in this roadmap was that latent state still sounded too close to one compressed hidden variable. The rest of this site no longer supports that shortcut. Lee et al. (2003), Havekes et al. (2016), Vierra et al. (2023), and Altas et al. (2024) show that phospho-signaling / second-messenger routing can change plasticity expression without collapsing to transcript or bulk protein abundance. Frey & Morris (1997), Shires et al. (2012), Govindarajan et al. (2011), Pandey et al. (2021), and Chang et al. (2024) show that local proteostasis / tag-capture route is another state family with its own integrative unit and turnover window. Correia et al. (2008), Swarnkar et al. (2021), Aiken & Holzbaur (2024), and de Queiroz et al. (2025) show that cargo-transport / cytoskeletal trafficking state is again a different route: what reaches the relevant spine, branch, bouton, or axon terminal can change plasticity and memory while the graph stays fixed. Therefore, on this page, a latent-state model is no longer accepted as a single unqualified object. It must say which family it is trying to recover, which family remained omitted, and what timescale the claimed state is supposed to govern.

Latent-state family What it constrains relatively directly What still stays latent if omitted
Fast executable state Immediate activity trajectory, momentary effective synaptic efficacy, and same-session intervention response. Why the same scaffold later stabilizes, forgets, or recovers differently across hours to days.
Phospho-signaling / second-messenger route Compartment-specific gating, phosphosite occupancy logic, kinase/phosphatase balance, and local signaling nanodomains that modulate plasticity expression. A model can fit next-step activity while still missing the controller that decides which plastic change is permitted, blocked, or redirected.
Local proteostasis / tag-capture route Which tagged synapse or branch captures plasticity-related proteins, and which late changes survive local turnover. Late stabilization, reconsolidation, and persistence remain underdetermined even if fast-state prediction improves.
Cargo-transport / cytoskeletal trafficking route Which receptors, endosomes, RNA cargoes, and presynaptic components actually reach the required compartment in time. Compartment-specific maintenance and structural plasticity remain latent even when transcript abundance, ATP support, or weight estimates look adequate.
State-family card required for maintenance-relevant R3 claims

If an R3 claim is used to support persistence, reconsolidation, remote memory, recovery after perturbation, or same-state language across days or acquisition regimes, the submission must publish at least (1) family name, (2) direct observable or perturbation route, (3) dominant timescale, (4) integrative unit (synapse / branch / neuron / region / whole-brain proxy), (5) whether that family was directly measured, externally calibrated, model-estimated, or omitted, and (6) the abstention boundary that follows. If regimes or days are bridged, attach the State-Continuity Bridge Card; if persistence or forgetting is part of the claim, attach the Maintenance-State Error Budget. Without that bundle, this roadmap treats the result as a prediction-useful latent variable or same-session proxy model, not as a maintenance-complete state estimate.

False condition:The latent state only serves to compress data, fails on held-out perturbations, or changes interpretation once omitted maintenance families such as phospho-signaling, local proteostasis, or cargo-routing are reintroduced.

Next:Compare connectome-only, fast-state, and maintenance-augmented models on the same held-out perturbation and longitudinal benchmarks, then publish the omitted-family list and the resulting claim ceiling.

R4 Causation: What models can predict response to intervention? CAUSALMULTI-SCALE

Question:Rather than just applying correlations, can we say ``what should change'' in response to changes in stimuli or conditions? What chain of evidence should we use to make the leap from decoding (correlation) to emulation (causation)?

Policy:This project treats Active Inference, DCM, state-space models, SCM, and mechanistic circuit models as competitive candidate sets. DCM is useful for comparing candidate generation models, and SCM is useful for describing interventions and counterfactuals, but neither is an automatic causal detector on its own. If only observational data is used, equivalence classes remain, so we require clarification of candidate model space, family-level comparison, model recovery, and held-out perturbation prediction.

2026-03-19 addendum:The weakness of the older wording was that candidate model space remained a slogan rather than an operational gate. Penny et al. (2004) already fixed that DCM inference is relative to the compared models, Rosa et al. (2012) showed that very large model spaces can be searched efficiently from a full model, Frässle et al. (2021) scaled directed-connectivity estimation to whole-brain human fMRI, and Wu et al. (2024) accelerated regression-style DCM further. But that progress is progress in tractability, not an automatic solution to identifiability. Even whole-brain or faster effective-connectivity estimates still depend on the chosen node set, priors, hemodynamic assumptions, omitted alternatives, and validation design.

Effective-connectivity route card required:Any R4 claim using DCM or a related effective-connectivity estimator must publish (1) compared node set plus omitted competitors, (2) observed-subsystem closure / latent-confound audit, (3) node-definition / parcellation policy, (4) neural-mass / HRF / prior assumptions plus sampling / transformation sensitivity, (5) family comparison and model recovery, (6) held-out perturbation or external validation, (7) test-retest window and condition dependence, and (8) abstention boundary. If those are missing, this roadmap stops at model-conditioned causal hypothesis and does not promote the result to discovered causal wiring or WBE-relevant causal structure.

Improvement measures (positioning of Multi-scale): Multi-scale integration is a promising expansion route, but it should not be the default route. Even when passing EEG-derived macroconstraints to a circuit model, acceptance or rejection is not determined by the "theory name" but by prediction improvement for stimuli, lesions, and task perturbations, OOD generalization, and transparency of abstention conditions.

Stricter implementation (Issue #52): The uncertainty obtained in the inverse problem (R1/R2) is explicitly propagated to the subsequent model. However, what we are fixing here is not the precision implementation of Active Inference itself, but rather theupstream uncertainty does not disappear downstream. The implementation differences between candidate theories are compared on the same bench.

Next: Combine the Intervention Premise Evaluation Task (V2), Model Space Declaration, Family Comparison, and Perturbation Bench with External Validation into one deliverable pack.

R5 Plasticity: What is the boundary between “introducing” and “not including” learning rules? PLASTICITY

Question:If we are dealing with long-term identity (V5), we have to include learning (renewal). However, with updates, verification becomes difficult.

2026-03-25 addendum:The remaining weakness here was that plasticity still read too much like a yes/no switch: either learning is introduced or it is not. The primary literature does not support that simplification. Frey & Morris (1997) and Shires et al. (2012) show tag / capture eligibility, Govindarajan et al. (2011) shows branch-level integration, Fonseca et al. (2006), Pandey et al. (2021), Chang et al. (2024), and Parker et al. (2025) show distinct synthesis-degradation / autophagy / proteasome-capacity routes, and Lee et al. (2003), Havekes et al. (2016), Vierra et al. (2023), and Rodriguez et al. (2025) show that phospho-state can be a gate rather than a stable weight itself. In parallel, Correia et al. (2008), Aiken & Holzbaur (2024), and de Queiroz et al. (2025) show that delivery routes and local RNA localization constrain what plasticity can actually be expressed. Therefore, R5 is no longer treated here as a yes/no choice. It is a claim-timescale audit.

Route What it permits relatively safely Minimum deliverables
(A) fixed model / no plasticity claim Same-session fit, bounded closed-loop control, or a narrow intervention result without persistence claims. Fixed decoder horizon, same-session hold-out, abstention on forgetting / reconsolidation / recovery, and an explicit statement that no maintenance-complete claim is being made.
(B) fixed fast-state model + maintenance disclosure Cross-day or cross-regime comparison without claiming that the controller of persistence has been identified. Temporal Validity Card, state annotation, Maintenance-State Error Budget, fixed-decoder versus rescue-mode separation, and the relevant family route cards when phospho, proteostasis, cargo, or other maintenance controllers are cited.
(C) explicit plasticity / maintenance model Named persistence, reconsolidation, recovery, or remote-memory claims with a declared controller family. Named plasticity rule, family-level ablations, held-out perturbation prediction, fixed-decoder durability plus recalibration burden, State-Continuity Bridge Card when acquisition regimes are bridged, and a failure / fallback policy if the controller assumptions break.
Site rule for plasticity claims

On this site, the words plasticity, adaptation, consolidation, recovery, and long-term stabilization are not treated as one achievement. If the claim reaches beyond same-session fit, attach the Temporal Validity Card. If it bridges across acquisition regimes or days as if it were one state sample, attach the State-Continuity Bridge Card. If the claim depends on maintenance families, attach the Maintenance-State Error Budget plus the relevant family route cards for phospho-signaling, local proteostasis, cargo-routing, or other cited controllers. Without that bundle, R5 stops at an adaptive or learning-capable model, not a verified continuity-preserving plasticity route.

Next required:When learning is included, report fixed-decoder durability, rescue mode, recalibration burden, family-level ablations, and held-out perturbation response on the same benchmark; then disclose which timescale of plastic claim was actually passed.

R6 Personalization: General model + individual parameters? Or is it completely individual? PERSONALIZATION

Question:Personalization is unavoidable when discussing identity. On the other hand, too much personalization leads to overfitting and broken reproducibility.

2026-03-18 addendum:The reason this section was deepened is that if personalization is read only as a trick for improving performance, the distinction between the target neural variable and a subject / session fingerprint collapses. Chaibub Neto et al. (2019) showed that diagnosis learning can absorb subject characteristics when repeated measures are not participant-disjoint, Wang et al. (2020) and Di et al. (2021) showed time-robust person identification from resting-state EEG alone, and Gibson et al. (2022) summarized strong subject-driven EEG variation. The same boundary also extends to acquisition distribution: Hu et al. (2018) showed that reference montage and electrode setup alter scalp potentials, Melnik et al. (2017) showed that system / subject / session all influence EEG, and Xu et al. (2020) showed that cross-dataset variability is driven in part by amplifier, cap, sampling rate, and filtering differences. The real boundary in R6 is therefore how much individual difference a given claim is allowed to use and how population signal is separated from fingerprint and setup routes.

Route What can be read relatively safely at this route Minimum deliverables
(A) shared model + subject-disjoint evaluation Whether a person-independent signal remains and how far population-level generalization can be established. Leave-one-subject-out / leave-one-session-out, raw-recording ancestry, subject/session/site/device/reference-system/electrode-layout metadata-only baselines, a fingerprint classifier, harmonization logs, and an explicit guarantee that windows cut from the same raw recording do not cross train/test.
(B) shared encoder + light subject adaptation How much calibration is needed to land a shared representation on the target subject. Frozen vs trainable parts, adaptation budget, when target-subject data is introduced, pre/post-adaptation scores, fixed decoder horizon, and cross-day degradation plus recalibration burden.
(C) fully subject-specific decoder / controller How far a participant-conditioned communication / control route works for that individual. Training time, cooperation burden, within-subject hold-out, cross-day hold-out, silence / abstention behavior, recalibration burden, and an explicit ceiling that does not claim other-person generalization.
Site rule for this section

On this site, any result that calls itself a diagnosis model, state biomarker, or general decoder must report at least (1) split independence unit, (2) raw-recording / window ancestry, (3) subject / session / site / device / reference-system / electrode-layout disjointness, (4) metadata-only or fingerprint baselines, (5) the harmonization log for channel map / reference / sample rate / filters, (6) when target-subject or target-site data was used, and (7) the adaptation budget. Without that, the result is treated first as a subject-aware classifier, participant-conditioned decoder, or fingerprint/setup-unresolved biomarker, not as a population marker or a subject-independent mechanism.

Branch example: Route (A) asks whether any population signal is present, route (B) asks how much transfer plus calibration burden is needed, and route (C) asks whether an individual route is operationally usable. Therefore, the fact that personalization scores higher is neither a success nor a failure by itself; it only becomes interpretable after you state which claim family you want to support and how far fingerprint and setup routes were controlled.

Next required:Bundle the Verification: Specificity & Shortcut Card, the evaluation-family and split design on Datasets, and the Temporal Validity Card so same-subject success, cross-subject generalization, and cross-day maintenance are managed as different deliverables.

R7 Identifiability: Does estimation approach the “unique solution”? What about uncertainty? IDENTIFIABILITY

Question:When there are many different models that explain the same observation, which assumptions can be used to narrow it down? Where does the uncertainty come from: sensor geometry, conductivity, preprocessing, candidate model ensembles, and intervention design?

Recondition:If the conclusion is significantly reversed due to small preconditioning differences, conductivity assumptions, or changes in the candidate model space, or cannot be recovered using held-out perturbations or external criteria, we do not consider the identification to be successful.

Next, we need:In addition to prior distribution/regularization, we publish all of the compared and excluded families, sensitivity analysis of electrode geometry, head model, and conductivity, simulation/phantom/invasive calibration, and abstention conditions when models overlap, and only adopt robust conclusions. The submission-side rule for this node is Verification: Identifiability Card.

Separate predictability, localization, and identity

This page does not treat improved held-out prediction, reduced localization error, and internal-state recovery that approaches a unique solution as claims of equal strength. The first two are important advances, but the third requires additional conditions before it can be claimed publicly.

  • Predictability: The ability to conditionally predict the next observation. It is useful for model compression and proxy improvement, but it does not by itself guarantee uniqueness of internal state.
  • Localization: Reduced positional error of a source under specified conditions. This matters for direct validation, but other uncertainties remain in deep, low-amplitude, and unobserved regions.
  • Identifiability: Conclusions remain stable even when candidate families and sensitivity analyses are included, and they can be recovered with held-out perturbations and external criteria.
2026-03 Supplement: Correlation, prediction, and identification are not treated with the same strength

The EEG literature with direct validation is rather revealing of its limitations when read carefully. Seeber et al. (2019) showed that the subcortical signal was conditionally detectable in 256ch scalp EEG, but did not claim general unique reconstruction. Unnwongse et al. (2023) showed in direct verification using intracranial stimulation that localization error depends on the conductivity assumption and source depth, and Hao et al. (2025) reported that source power and source depth greatly influenced the error in a patient cohort of simultaneous HD-EEG/SEEG. Therefore, even if a cross-modal correlation or a held-out prediction appears, it is primarily an improvement in predictability / localization, and does not necessarily lead to a unique solution to the internal state.

2026-03-26 supplement: richer observables are not the same as a collapsed solution set

The remaining weak point here was that more measured variables could still sound too close to unique recovery. Villaverde (2019) reviewed that observability and structural identifiability are different questions, and Villaverde et al. (2019) showed that unknown inputs, states, and parameters often have to be analysed jointly. In neuroscience, Prinz et al. (2004) showed that similar circuit activity can arise from disparate parameters, Rasero et al. (2024) showed that similar human activation patterns can still hide different macroscopic network states, and Beiran & Litwin-Kumar (2025) showed that connectome-constrained dynamics remain degenerate until additional recordings are supplied. A second weakness remained after that: Vafaii et al. (2024) showed both common and divergent organization across simultaneous Ca2+ and BOLD, Chen et al. (2025) showed tightly coupled global progression plus two distinct network patterns in simultaneous EEG-PET-MRI, and Lu et al. (2023), MICrONS Consortium et al. (2025), and Egger et al. (2024) show why same-brain or same-subject sequential pipelines still remain exposed to preservation change and state drift. Therefore, this roadmap now treats R7 as a five-step audit: (1) observability budget, (2) structural identifiability in the declared model space, (3) practical identifiability under the actual dataset and perturbation design, (4) bridge / composition validity when multimodal or sequential evidence is invoked, and (5) external validation plus abstention when candidate sets still overlap.

Ambiguity class How it fools the reader What would actually raise the claim ceiling What this roadmap still refuses to count
Symmetry / reparameterization ambiguity Several parameter or state transformations still produce the same declared outputs, so a good fit looks unique when it is not. Add a symmetry-breaking observable, known input / initial condition, or an explicit reformulation that removes the hidden transformation. More optimizer restarts, longer fitting, or more repeats of the same uninformative protocol.
Regime-restricted degeneracy Different internal states or parameter sets agree inside one narrow task, state, or operating regime and only diverge outside it. Use state transitions, persistent excitation, targeted recordings, or perturbation contrasts that make the surviving alternatives separate. More passive samples from the same narrow regime when the alternatives remain co-linear.
Model-family / omitted-mechanism ambiguity An approximate family looks well identified until a complementary condition reveals systematic residuals or missing mechanisms. Run a discrepancy stress test, compare richer / alternative model families, and require held-out falsification instead of post hoc explanation. Narrower intervals inside one misspecified family or one headline fit improvement.
Task-variable interaction / representation ambiguity A reduced representation predicts correlations while missing the interaction structure that actually drives behavior. Preserve or perturb the task-variable interaction and compare intervention-side predictions of the competing representations. Correlation-only dimensionality reduction, generic multimodal fusion, or interpretability language without perturbation.
Evidence pattern What it can improve What it still does not close Required next audit on this page
Same-session multimodal bundle Can tighten observability and expose complementary errors Does not by itself remove shared-driver risk, quantity mismatch, or latent-state multiplicity Fusion Card plus the R7 candidate-space audit
Connectome-constrained predictor Can restrict one family of candidate dynamics Does not by itself remove dynamical degeneracy or omitted-state families Observability Budget (Connectome-Constrained Model Card) plus held-out recordings / perturbations
Same-brain or same-subject sequential pipeline Can connect specimen identity across stages Does not by itself guarantee same-state continuity across fixation, delay, or regime shift State-Continuity Bridge Card plus temporal-validity disclosure
2026-03-31 supplement: informative protocols must target ambiguity, stress omitted mechanisms, and declare minimum sufficiency

The remaining weakness after splitting observability from identifiability was that a protocol could still be described with generic words such as multimodal, naturalistic, or closed-loop without saying why that design should actually separate the survivors. The primary literature does not support that shortcut. Diop & Fliess (1991) made persistent trajectories explicit, and Raue et al. (2010) showed that identifiability analysis can guide new experiments rather than only criticize old ones. Chis et al. (2016) then showed that sloppiness is not identifiability and that design should optimize explicit identifiability criteria rather than proxy uncertainty scores. White et al. (2016) showed that complementary experiments can expose omitted mechanisms rather than solve them. In current neuroscience, Beiran & Litwin-Kumar (2025) showed that a small targeted recording set can remove degeneracy in connectome-constrained networks, and Langdon & Engel (2025) showed that preserving causal interactions among task variables can recover behaviorally relevant computation that correlation-only reductions miss. Gevertz & Kareva (2024) then showed that identifiability analysis can derive a minimally sufficient schedule, and Liu et al. (2025) showed that active learning can reduce the observations needed to reach practical identifiability. Therefore, this roadmap now asks not only whether a protocol is richer, but which ambiguity it targeted, which identifiability objective chose it, whether it exposed omitted-mechanism error, and what minimum-sufficiency stop rule would have been enough.

Design-side audit field What R7 now asks for What this stops you from overreading
Named ambiguity State which surviving states, parameters, controllers, or model families remain confusable after the current fit, and which one the next protocol step is trying to separate. We added another task or modality is not read as targeted ambiguity reduction.
Identifiability objective Name whether the next condition was chosen to optimize practical identifiability, model discrimination, posterior contraction, profile-likelihood width, or another declared criterion. A richer protocol is not read as informative merely because a generic optimization score improved.
Orthogonal regime / targeted recording leverage Disclose the stimulus range, state transitions, perturbation timing, or targeted recording subset that should rotate the surviving candidates apart rather than merely resample one narrow regime. More samples from the same operating slice are not read as collapsed degeneracy.
Model-discrepancy stress Report whether the new condition exposed systematic residuals, failed fits, or omitted mechanisms, and what hierarchy of richer or alternative models was checked afterwards. Tighter nominal uncertainty is not read as trustworthy when the design only forced hidden model error into view.
Minimum-sufficiency stop rule Predefine the smallest condition / time-point / recording set required to raise the claim, the uncertainty or overlap threshold used to stop, and how much candidate overlap remained once that minimum design was reached. Open-ended data accumulation is not read as solved identifiability if the stopping logic was never declared.

Operational link: When this design-side disclosure is missing, this roadmap stops at predictive fit, localized source hypothesis, or model-conditioned mechanism. The submission-side operational version remains Verification: experiment-design leverage.

</details>
R8 Compression: Can the “verification standard” be maintained no matter what information is discarded? COMPRESSION

Question: It would be ideal to “save everything” for WBE, but that is not realistic. Can we identify the minimum representation that still satisfies the evaluation target (V0)?

Next: Vary the compression ratio and measure where the performance of the evaluation suite breaks down

R9 Audit: What is the mechanism for "keeping" model differences and failure examples? AUDIT

Question:Because this is an area where the conclusion changes every time it is updated, it is necessary to record differences, failures, and negative results.

Next: Connect model/data/evaluation versions so that they can be reproduced by re-running (P2)

R10 Neuromodulation: How to incorporate mood, alertness, and volume transmission? NEUROMODULATION

Question:Synaptic connections alone cannot reproduce state-dependent regulation of dopamine, serotonin, acetylcholine, and related volume-transmission systems. The weak point here was that this node still let pupil / HRV, local transmitter imaging, receptor maps, occupancy PET target engagement, and displacement / release-sensitive PET sound more interchangeable than they are. That was too weak. Reimer et al. (2016) showed that pupil fluctuations track both adrenergic and cholinergic activity, Neyhart et al. (2024) showed that local cortical ACh depends on axon activity and local clearance, Hansen et al. (2022) and Goulas et al. (2021) showed that receptor maps are structured regional priors, Wong et al. (2013) showed ligand- and dose-limited D2 target engagement under administered lurasidone, and Koepp et al. (1998), Lippert et al. (2019), plus Erritzoe et al. (2020) showed challenge- and time-window-limited dopamine or serotonin release proxies rather than a free readout of the whole transmitter field.

Operational rule:On this site, neuromodulation is not treated as a single mood scalar. It is a ladder with at least five non-equivalent rungs. (1) mixed arousal proxy such as pupil / HRV / locomotion is a coarse covariate only. (2) local axon or transmitter sensor calibrates a measured region's chemical signal but not the whole brain. (3) receptor / transporter atlas is a regional chemoarchitectural prior, not the current state. (4) occupancy PET is ligand-, receptor-family-, dose-, and window-limited target engagement under an administered drug. (5) displacement / release-sensitive PET is a challenge-, ligand-, receptor-family-, and time-window-limited endogenous release proxy. None of these rungs alone is accepted here as the ground truth of the current whole-brain neuromodulatory state.

Next:When an R10 claim is made, publish which rung was used, plus receptor family, ligand / challenge, dose when applicable, temporal window, spatial scope, and abstention boundary. Then compare the same baseline with and without that rung under the same held-out state transition. For the stack-level ceiling, see Wiki: neuromodulatory proxy ladder; for the broader hidden-state argument, see Wiki: Why wiring diagrams alone are not enough.

3. Implementation

Implementation: On what basis? What about real-time performance?

Central question: "Which computational platform (HPC/cloud/neuromorphic/hybrid) should we use to run the restored model? Can it be put into a closed loop (body/environment)?"

I0 Execution platform: Which one are you aiming for: general-purpose computing/GPU/HPC/neuromorphic? IMPLEMENTATION

Question:Can the implementation platform achieve both “sufficient fidelity” and “realistic cost” for the restoration target (R0)?

Next:Estimate the model granularity (I2) and calculation amount (I3) as a set

I10 Time continuity: Discrete time (RNN) or continuous time (ODE)? NEURAL-ODEDYNAMICS

Question:Biological brains operate in continuous time, whereas standard RNNs/Transformers are discrete time. This gap creates distortion in dynamics.

Suggestion:Adopt Neural ODEs (Neural Ordinary Differential Equations) or CTRNNs as the implementation framework. By modeling as $\frac{dh(t)}{dt} = f(h(t), t, \theta)$, sampling at arbitrary time resolution and memory-efficient learning using the adjoint method are possible.

Next:Integrating the different time scales of EEG (milliseconds) and fMRI (seconds) in the same system of differential equations

I1 Closed loop: How many ms of delay can be tolerated in real time? REALTIME

Question:If we are to evaluate “personhood”, the essence is interaction with the environment (delay/noise). Tolerable delay is task dependent.

Redefinition (2026-03): Instead of a single number, audit acceptable delays for at least (a) state feedback / neurofeedback, (b) ERP / command BCI, (c) phase-locked stimulation, and (d) burst/state-triggered neuromodulation. Phase-targeting now focuses first on oscillation estimability and then on phase error, causal-versus-post-hoc benchmark, circular precision, functional comparator, and fixed-versus-adaptive phase policy. Adaptive DBS focuses on biomarker burst detection delay and stopping rules.

Supplement (Issue #46): We actually measure the end-to-end delay and jitter using CLET, photodiode, loopback, and TTL, and position LSL's time_correction as software offset correction. This is not to say that LSL alone guarantees actuation onset.

Next: On the evaluation suite (V1) side, leave the loop class explicit, median/P95/P99/worst-case latency, number of abstentions/freezes/stops, and performance degradation curve for additional delay. For phase-targeted loops specifically, also leave the target band / spatial filter, power-SNR gate, no-stim rate, causal-versus-post-hoc benchmark, mean phase offset plus circular spread, off-target or random-phase comparator, and fixed-versus-adaptive phase policy. For information on organizing entrances, see Wiki: Closed-loop, delay, jitter, and safety stops.

I2 Model granularity: Where do we fight between spiking/rate/abstraction? GRANULARITY

Question:While the fidelity increases (tentatively) as the granularity increases, identification (R7) and calculation (I3) break down. What is the granularity to win?

Next required:Associate "required measurement (M2/M3)", "required calculation (I3)", and "passing evaluation (V0)" for each granularity

I3 Computational amount: Where are the bottlenecks in memory/power/parallelization? SCALING

Question:Increasing the fidelity of a model will explode computational resources. Where should approximation/compression (R8) be performed?

Next: Measure the boundaries of how the evaluation suite can be compressed without compromising performance

I4 Initialization: How to give the “initial state” of the model? INITIALIZATION

Question:Even if the structure is the same, the behavior will be different if the initial state is different. At what point in measurement (M7) should we “start”?

Disproval condition: The result is unstable due to the arbitrariness of the initialization, and the identity evaluation (V5) does not hold.

Next: Fix the initialization procedure and measure the sensitivity (robustness) to initialization

I5 Learning and Drift: How to design “safe updates” for learning? LEARNING

Question: Allowing learning (R5) involves identity (V5) and safety (D2). How do I monitor updates and where do I stop them?

Next required:Drift indicator (V4) + kill switch/isolation (I8)

I6 Embodiment: How much input/output (sensors/movement/language) should we have? EMBODIMENT

Question:Closed-loop verification (I1/V1) requires an environment and a body (or a substitute thereof). What is the minimum physicality?

Stricter policy:The word “embodiment” is too cheap if it only means that some output device existed. Flesher et al. (2021) showed that tactile feedback improves robotic-arm control, but this is evidence for a local surrogate sensorimotor loop, not that whole-body coupling is solved. Combined with Musall et al. (2019), Stringer et al. (2019), Saleem et al. (2013), Ravassard et al. (2013), and Zelano et al. (2016), this means I6 must disclose which body / environment loops were present rather than talking about embodiment as a yes/no label.

  • Declare retained / substituted channels: vision, touch, proprioception, vestibular cues, respiration / arousal routes, dialogue partner, and task-state cues.
  • Describe the surrogate plant: robotic hand, cursor, speech synthesizer, virtual avatar, or other actuator, together with latency, noise, and saturation.
  • Run removal tests: feedback-off, sensory-substitution-off, altered environment, or loop perturbation conditions that show what really carries the performance.

Next: Fix the environment (VR/game/dialogue) and design an evaluation that can be repeated under the same conditions, then publish a body / environment boundary card plus ablation plan.

I7 Reproducibility: Is there a guarantee that the results will not change due to implementation differences (language/hardware)? REPRO

Question:Even with the “same model”, the behavior changes depending on numerical errors and parallel order. How much decisiveness do you require?

Next: Verify that the error is within tolerance using a fixed random number, fixed environment, and differential test (P2).

I8 Safety: How to create isolation, monitoring, and containment? SAFETY

Question:The closer you get to the strong claims (L4-L5 of P1), the more safety becomes a technical requirement.

Next:Include a triad of isolation environments, monitoring metrics, and outage protocols in your L3 and beyond implementation plan

I9 Thermodynamics: How to audit implementation costs and irreversibility? THERMODYNAMICS

Question:To what extent should thermodynamics be included in the evaluation when comparing WBE implementations? Things you should not mix here are Landauer lower bound, organizational-level energy budget, time irreversibility of coarse-grained neural dynamics, and model-based entropy flow. The question to ask is not ``Can this prove identity?'' but rather how to audit the physical costs and nonequilibrium indicators without overstating them.

2026-03 Literature Audit:Bérut et al. (2012) demonstrated a logically irreversible minimum dissipation lower bound for bit erasure, not wall-power or NESS reproduction conditions for the entire digital brain. What Lynn et al. (2021), de la Fuente et al. (2023), and Nartallo-Kaluarachchi et al. (2025) demonstrate is the usefulness of measuring broken detailed balance and time irreversibility from coarse-grained time series such as fMRI/ECoG/MEG, and is not an acceptance condition for microscopic physical dissipation or identity. Ishihara & Shimazaki (2025) also show that entropy flow estimation is a model-based quantity that requires assumptions such as state-space kinetic ising model, pairwise coupling, and mean-field approximation. A second correction is that the observed trajectory is not automatically thermodynamically closed. Martínez et al. (2019) showed that waiting-time asymmetry can reveal hidden dissipation even when observable current vanishes, Hartich & Godec (2024) showed that this interpretation can fail when coarse-graining and time reversal do not commute, Martínez et al. (2024) replied by limiting the original claim to local-in-time coarse-grainings and, where needed, second-order semi-Markov constructions, and Blom et al. (2024) showed that coarse lumping can hide dissipative cycles and induce memory so that estimates become far too small when the observed trajectory is naively treated as Markov.

Treatment to be fixed on this page

  • Do not make it a required gate:Maintaining EPR > 0, Virtual Dissipation Protocol, biological brain 20W matching, and communication: calculation ratio matching are not set as acceptance conditions at this time.
  • Treat as supplementary logs: Wall-plug power, FLOPs, time irreversibility, and model-based entropy flow are recorded separately.
  • Explicit modality dependence:Since the estimators have different meanings for fMRI / MEG / ECoG / spike train, they are not directly substituted for each other.
  • Name estimator family and null control:Transition-flux lower bounds, time-shifted correlation asymmetry, inversion classifiers, visibility-graph indices, and state-space kinetic Ising entropy flow are not interchangeable, and shuffle / surrogate controls are part of the claim.
  • Audit observed-state closure and memory order:State whether the observed process is treated as Markov, 1st/2nd-order semi-Markov, or another non-Markov process, what hidden states / cycles were collapsed, and whether a low estimate could be caused by hidden dissipation rather than actual near-equilibrium behavior.
  • Dividing the lower bound and the main body:The irreversibility obtained from the observation time series is often a coarse-grained lower bound, not the microscopic physical dissipation itself.

Next: We need an irreversibility benchmark that is reproducible within the same modality and preprocessing chain, plus a log schema that separates wall-plug power, FLOPs, and simulated metabolic overhead. At minimum, publish an irreversibility route card naming signal route, state definition, observed-state closure / memory order, estimator family, null / surrogate control, lower-bound or model-based quantity type, and hardware power isolation. For more information, see Wiki: irreversibility route card.

4. Verification

Verification: What must be met to say that they are the same?

Central question:"How do we define technically verifiable criteria among identity/consciousness/behavior, and how do we set conditions for disproving it?"

V0 Target of verification: Which of identity, consciousness, or behavior should be verified “engineeringly”? VERIFICATION

Question: You cannot verify what cannot be measured. What should count as “measurable conditions for success”?

Branching (example):(A) Equivalence of behaviors and abilities / (B) Equivalence of intervention responses / (C) Continuity of autobiographical memories and values (need to be careful)

Next: Pre-register the evaluation suite in V1 and link it to the claim ladder in P1.

V1 Evaluation suite: What should we measure as “progress”? (Pre-registration) BENCHMARK

Question:Fix tasks, metrics, baselines, statistics, and failure conditions in advance.

  • Behavior:Generalization in unlearned tasks, reaction times/errors, learning curves
  • Neuro:State transitions, spectral features, network metrics, intervention responses
  • Temporary identity:Coherence of autobiographical memory, stability of preferences, consistency of self-model

Next: First, create a “small but retestable” suite for L0-L2.

V2 Causal testing: Can stimuli/interventions produce the “same response”? CAUSAL

Question:Correlation matching can be achieved simply by “simulating” things. A strong verification will be whether there is agreement with the intervention.

Strength of evidence is treated in four stages:This roadmap does not reduce causal testing to a single type of pass/fail. The primary literature now supports staged evidence of varying strength.

  • Gate 1 / held-out perturbation prediction:The ability to offline predict responses when changing task conditions, stimulus timing, and environmental changes.
  • Gate 2 / online human-in-the-loop: A loop that includes real-time input and feedback while maintaining performance and auditing latency / jitter.
  • Gate 3 / local causal intervention: Interventions such as TMS-EEG, intracranial stimulation, and adaptive DBS in which stimulation site, intensity, artifact handling, and safety-stop conditions are disclosed.
  • Gate 4 / long-run closed loop:The ability to track recalibration loads, recovery times, and even failure modes over multiple sessions and long-term operations.

Required logs: Stimulation site, intensity, masking, artifact window, synchronization path, median end-to-end latency / P95 / P99, jitter, dropout, recovery time, and a body / environment boundary card naming retained / substituted sensory, motor, and interoceptive channels, slow internal-milieu routes such as circadian phase or glucocorticoid / metabolic regime, plus the surrogate body / environment contract. If these depart from the TMS-EEG recommendations summarized by Hernandez-Pavon et al. (2023), the result is not treated here as a comparable causal test.

Decision rule:The success of Gate 1 alone does not assert L3 or "counterfactual equivalence." In order to advance to a stronger claim, it is necessary to submit a bundle of evaluation items, external validation, and abstention conditions that connect M6 (intervention design) and R4 (causal model). For detailed operational verification, see Verification Causal Perturbation Suite.

Next: Create evaluation items that connect M6 (intervention design) and R4 (causal model), and reduce at least Gate 1 / Gate 2 to specifications that can be re-executed by a third party.

V3 Generalization: Can you maintain your “identity” even in out-of-distribution (OOD)? OOD

Question:If it only works under the same conditions as training, it is more like "playback" than uploading. Measuring consistency in novel situations.

Next:Define the performance degradation curve under untrained tasks, environmental changes, and noise conditions

V4 Long-term: How to evaluate learning, drift, and forgetting? LONGITUDINAL

Question:Long-term verification is not just whether a score stayed high. It asks whether state fluctuation, trait-like backbone, biological drift, interface / decoder drift, and maintenance-route burden were separated well enough that the time claim becomes falsifiable.

2026-03-22 addendum:The older wording was too close to an abstract question about “acceptable change.” The site's current technical rule is narrower. Wilson et al. (2025) showed that long-term iBCI control can be maintained by unsupervised recalibration, but that usability and fixed-decoder durability are different claims. Wairagkar et al. (2025) showed a strong communication route while also quantifying degradation of a decoder trained on a specific earlier day. Therefore, V4 now treats long-term verification as the conjunction of temporal validity, bridge validity when acquisition regimes are crossed, and maintenance-state disclosure when persistence or forgetting is the claim.

  • Temporal validity: audit fixed decoder interval, state annotation, interface / decoder drift, recalibration burden, and transfer ceiling.
  • Bridge validity: if “same-subject” evidence crosses live-to-fix, cross-regime, or cross-day reacquisition, attach the State-Continuity Bridge Card; same-subject is not same-state by default.
  • Maintenance route: if the claim concerns persistence, forgetting, reconsolidation, remote memory, or recovery, attach the maintenance-state error budget instead of treating temporal hold as sufficient evidence.

Required next:M7 longitudinal audit + Temporal Validity Card + update log (P2). For multiday or cross-regime claims, add the relevant bridge and maintenance-route cards instead of leaving the time burden implicit.

V5 Identity: What is “causal identity” that goes beyond psychological continuity? IDENTITYLEGAL

Question: Similarities in memory and personality alone, that is, psychological continuity, are not enough to distinguish identity from copying or imitation by an LLM-like system. A more rigorous engineering definition is needed.

  • New standard: Causal Identity: Extending Parfit's psychological theory, the condition for identity is that the system's ``future predictive ability (accuracy of active inference)'' is statistically indistinguishable from that of the biological brain.
  • Indicator:Execute theCausal Perturbation Protocol, which is an extension of the Turing Test. We will introduce an index (e.g. Perturbational Complexity Index, PCI) to verify the statistical identity of the biological brain's response to physical perturbations such as TMS and the response to virtual perturbations in emulation, and confirm the correspondence of dynamic causal structures.
  • Conditions for disproving:They have memory, but their adaptation/prediction patterns to new environments deviate from the original (zombie/imitator)
  • Next: Make predictive accuracy a core metric in V8's "imitation-versus-imitation" testing.
V6 Consciousness: To what extent can theory-dependent predictions be made “testable”? CONSCIOUSNESS

Question:Consciousness cannot be directly observed. Therefore, the question is whether the prediction differences produced by theories (IIT / GNWT / FEP, etc.) can be benchmarked while minimizing the task/report confound.

  • Theories are not accepted on a winner-take-all basis: Ferrante et al. (2025) demonstrated the need for preregistered prediction competition rather than the victory of a single theory. Therefore, on this site, IIT / GNWT / FEP will compete as a prediction family, and the theory name itself will not be used as a pass/fail indicator.
  • PCI / PCI-ST is a perturbation benchmark: Casali et al. (2013) and Comolatti et al. (2019) gave a strong benchmark candidate, but not a universal meter. We treat it as an auxiliary measure of L2/L3 only if we disclose the stimulation site, intensity, masking, artifact window, and cohort calibration.
  • Spontaneous complexity / criticality is a calibration track: Maschke et al. (2024) showed an association between spontaneous criticality and PCI, while Casarotto et al. (2024) showed a dissociation between spontaneous features and PCI in a minimally conscious state. Therefore, the resting-state marker is not a PCI replacement, but rather a proxy that requires external calibration.
  • No-report / criterion placement is fixed first: As Cohen et al. (2024) and Fahrenfort et al. (2025) show, if we do not control for post-perceptual processing and criterion placement, there remains a possibility that we are looking at task/report confound rather than theoretical differences.

Note:Theory is treated only as a tool for hypothesis generation, and the success or failure of implementation (WBE) is determined not by "conformity to a specific theory" but by the bundle of V2 (causality), V3 (OOD), and V4 (long-term stability). Even when using geometry or topology, the main pass/fail points are placed in perturbation and generalization.

V7 Copy/branch: How do you handle the “principal” when multiple instances run? BRANCHING

Question:Easy to reproduce from an engineering standpoint. Verification design will also fail if branching is not assumed.

Next: Connect individual ID/version management (P2) with identity evaluation (V5)

V8 LLM/Distinguishing from imitation: How do you judge whether the external appearance is similar but the content is different? DISAMBIGUATION

Question:Conversations and reports are easy to imitate, and the stronger the prior distribution of the language model, the more likely it is. Therefore, what we want to distinguish here is not the abstract "AI-likeness" but the extent to which neural signals exceed language prior and the modality, task, and time scale in which they are established.

Track Relatively strong statements as of 2026-03 Things I can't say yet
Non-invasive semantic / caption decoding Tang and Horikawa used fMRI to advance semantic reconstruction of continuous language and generation of descriptions of video content. The goal here is totransfer part of the semantic expressionto an external sentence. Since dependence on subject cooperation, task conditions, and caption candidate space remains, this alone cannot be said to be a reproduction of self-model or WBE.
Non-invasive word / speech decoding Défossez et al. and d'Ascoli et al. have advanced speech/word decoding from MEG/EEG, showing that MEG and reading tasks are advantageous and data volume strongly dominates performance. Open-ended thought reading and strong cross-day/cross-task generalizations are another matter. It is impossible to make a strong argument while hiding candidate set and device dependencies.
Invasive streaming speech neuroprosthesis Willett, Littlejohn, and Wairagkar made significant advances in speech BCI as a local communication subsystem. Large vocabulary, streaming throughput, and closed-loop own-voice synthesis are important achievements. What is shown here is local progress in communication routes, not whole-brain emulation or identity preservation. Throughput also does not by itself determine transferability, fixed-decoder durability, or rescue burden.
Transfer-assisted speech-decoder initialization Singh et al. showed that distributed intracranial recordings can support a group-derived decoder that outperforms models trained on individual data alone, enabling a transfer-learning route for individual initialization. This is evidence for cross-subject initialization and shared latent structure, not for subject-independent final deployment or for durable fixed-decoder operation after initialization.
Fixed-decoder durability slice Willett et al. reported a bounded no-new-day-training slice, and Wairagkar et al. made it impossible to ignore the distinction between fast voice synthesis and the time horizon over which a decoder remains usable without rescue. This still does not show that fixed decoders remain adequate over long horizons, across changing tasks, or across body / environment regime changes.
Adaptive stabilization / rescue Karpowicz et al. and Wilson et al. showed that latent-dynamics alignment and unsupervised recalibration can stabilize iBCI performance when recording conditions drift. This is a rescue route, not proof that the original decoder stayed stable on its own. It therefore cannot be collapsed into fixed-decoder durability or subject-independent deployment.
2026-03-28 addendum: invasive language BCIs need an operational split

The older front-door wording was still too coarse because it let readers compress several different achievements into one `speech BCI success` bucket. The primary literature does not support that. Willett et al. (2023) strengthened same-session throughput and also exposed a bounded fixed-decoder slice, Littlejohn et al. (2025) strengthened streaming throughput / expressivity, Wairagkar et al. (2025) strengthened instantaneous voice synthesis with silence fallback, Singh et al. (2025) strengthened transfer-assisted initialization, and Karpowicz et al. (2025) plus Wilson et al. (2025) strengthened adaptive stabilization. Those papers do not all answer the same operational question, so this roadmap no longer places them on one row.

Minimum evaluation pack
  • Isolation of neural contribution: Report no brain, time-shuffle, trial-shuffle, LM-only, and no-LM; if decoding is search-based, also publish candidate-set size.
  • Bounds of generalization: Report held-out stories, held-out vocabulary, cross-day, cross-task, and cross-subject separately and do not hide the need for subject cooperation or individual adaptation.
  • Operational route label: State whether the result is same-session throughput, transfer-assisted initialization, fixed-decoder durability, or adaptive rescue after alignment / recalibration.
  • Streaming metrics: Report not only words per minute but also P50/P95/P99 latency, silence / abstention, dropout, recalibration burden, and recovery time.
  • Do not over-promote: Do not elevate decoding success, embedding similarity, or conversational naturalness into claims about emulation, WBE, or identity preservation.

Next: Connect an evaluation pack that bundles LM-only / shuffle / OOD / cross-day / latency / abstention / recalibration to V2 (intervention), I1 (closed loop), and R4 (causal modeling) so that progress on the communication route is not conflated with claims about WBE.

V9 Supplementary exam: Is the design designed to allow a third party to reach the “same conclusion”? REPRO

Question: The shortest path to research is to create a situation in which other people can repeat the test. The more identity is debated, the more important repeatability becomes.

Next:Make data/code/evaluation publicly available to the extent that P2 (deliverables) is met

V10 Model distance: How do we measure the "closeness" between the biological brain and emulation? METRICMATH

Question:Simple Euclidean distance in parameter space often does not reflect differences in the functional behavior of models. How do we quantify the more essential "closeness of behavior"?

Policy: Do not adopt a single mathematical distance as the standard solution. At present, stronger external criteria come from observable behavior such as held-out perturbation responses, OOD degradation curves, calibration errors, and recovery times. Distances such as Fisher-Rao, representation distance, and manifold separation are compared as candidates, but none of them is the default pass / fail indicator.

Next: Compare on the same benchmark how well candidate distances match the external standards in V2 / V3 / V4. Sensitivity analysis should reveal whether the ranking stays stable when preprocessing, sample count, and state differences are changed.

V11 Topology: Is the “structure” of consciousness conserved? TDAGEOMETRY

Question:Scalar values such as PCI (complexity) alone do not tell us whether the "qualitative structure" of consciousness is the same.

Policy:Topological data analysis (TDA) or manifold geometry is a promising track forancillarydescribing structural differences falling in scalar indicators. Yoon et al. (2024) provided a mathematical basis for comparing topologies in cross-populations, but the iScience 2024 mouse V1 circular manifold study also showed that there are sampling obstacles to retrieving truthful topologies. Therefore, topology distance cannot be directly linked to guaranteeing consciousness structure or passing or failing WBE.

Next: Through sampling, embedding dimension, distance function, preprocessing, session drift, and sensitivity analysis with perturbation responsiveness, only phase features that remain stable are reported as auxiliary indicators.

5. Deployment & Governance

Social implementation: What is needed beyond technology?

Central question:"What kind of systems, ethics, and legal foundations other than technology are needed for L5 (social implementation)?"

D0 Ethical Framework: What are the rights of emulated beings? ETHICS

Question:If WBE is successful, is emulation a rights subject? Is stopping “death”? Is reproduction “birth”?

Current status:A framework for protecting neural data is being developed through the EU AI Act, the US MIND Act (proposed in 2025), etc., but it does not address the issues specific to WBE (duplication, suspension, modification).

Next: Develop design guidelines for the ethics review committee in parallel with technical verification (L0-L2)

D1 Consent and withdrawal: How do we design informed consent? CONSENT

Question:Consent to the WBE process is given in advance, but how can we guarantee the right to withdraw during and after the process?

Next: Design a staged consent model (re-consent at each phase) and engineering implementation of termination rights

D2 Access and equity: Who should have access to WBE? EQUITY

Question:If WBE becomes a reality, how will we deal with the disparity problem where only the wealthy have access to "immortality"?

Next need:Position open access design of verification commons as the first step to prevent disparities

D3 Security: How to design for exploitation risks and controls? SECURITY

Question:How to deal with the risks of emulation modification, forced mental manipulation, military use, etc.?

Next: In conjunction with the I8 (containment) design, design technical guarantees for access control, modification auditing, and termination rights.

Definition catalog of open problems (strict version)

Here, we don't use "unresolved" as an ambiguous term; we define what is input, what is output, and under what conditions it passes/fails. Even if there is a wealth of prior research, if the criteria for judgment are ambiguous, the research will not accumulate. On the other hand, if the definition is strict, small advances can be reliably accumulated.

ID Definition of unsolved problems (rigorization) Existing points Unresolved boundary (next point to close)
U0 Operational identity: For intervention set I and time window T, can it be defined as "identical" if the conditional distribution difference between the biological system and the model system is less than a threshold? There is a framework definition in V0/V5. Introduced the distinction between psychological continuity and causal identity. The minimum completeness conditions, threshold settings, and pre-registration rules for intervention set I have not been determined.
U1 Identifiability of the inverse problem: When estimating the potential source x from the observation y, is it possible to quantify "closeness to the unique solution" by the degree of concentration of the posterior distribution? The policy for estimation with uncertainty has been clearly stated in R2. However, public implementation and external validation are incomplete. Overall identification possibility including forward model error (conductivity/shape) has not been verified.
U2 Lower limit of spatiotemporal resolution: Is it possible to define the minimum temporal/spatial resolution for each task without destroying cognition, learning, and intervention responses? The points of discussion have been sorted out in M2/M3. Insufficient measured boundaries (delay tolerance/granularity collapse point) in closed-loop tasks.
U3 Subject Boundary: Can the minimum set of physiological, neurological, and environmental elements included in the subject be fixed with a verifiable operational definition? Hybrid measurement protocol exists, and M4 / I6 now add a public body / environment boundary-card policy. Sensitivity analysis of how much the V5 judgment changes when the boundary changes has not been completed.
U4 Counterfactual Equivalence: Can equivalence be determined based on predicted matching of intervening branches rather than observed matching? Theoretical policy was clarified in R4/V2/V6. Insufficient implementation of counterfactual virtual generator, false intervention baseline, and stopping rules.
U5 Learning rules and personality drift: If learning updates are allowed, to what extent is the expression change considered to be the same? The problem has been raised in R5/V4. The allowable drift threshold and kill switch conditions for each update rule are undefined.
U6 Separation of personalization and generalization: Is it possible to identify individual-specific parameters and task-common parameters and avoid overfitting? Present evaluation policy in R6. There is a lack of standard benches and test plans for intra-individual/inter-individual decomposition.
U7 Multimodal alignment: When integrating EEG/fMRI/behavior/physiology, can coordinates, delays, and noise differences be kept auditable? Required in M5/M9, Motion-BIDS policy (Issue #46). Concordance rate evaluation of multiple integrated pipelines using the same data is not yet developed.
U8 Closed-loop stability: Can the model be controlled without breaking down under delays, noise, and environmental changes? Organize requirements using I1/I6. Identification of safe operation area based on actual measurements of delay and jitter is insufficient.
U9 OOD robustness: Can the sameness index be maintained in unlearning environments, novel tasks, and noise conditions? Clarified the problem in V3. Standardization of the deterioration curve (at what speed it collapses) has not been established.
U10 Thermodynamic consistency: Is it possible to formulate implementation conditions under which information processing satisfies physical irreversibility and dissipation constraints? Theoretical organization has progressed with I9 and Issue #58/#61/#62. The unification of measurable experimental indicators (dissipation rate, causal density, efficiency) has not yet been completed.
U11 Calibration of perturbation/spontaneous/geometry proxies: Under what conditions do the PCI/PCI-ST, spontaneous complexity, criticality, and manifold/topology indicators match and under what conditions do they dissociate? Defined the role difference between benchmark hierarchy and auxiliary analysis in V6 / V11. Lack of external calibration across multicohort, no-report, and perturbation conditions.
U12 Branch identity (multiple instances): How to handle individual ID, responsibility, and evaluation attribution after multiple branches. The necessity of version control was clearly demonstrated in V7. A system design with a fixed branching point and evaluation period has not been started yet.
U13 Imitation separation test: Can behavioral imitation (LLM type) and causal structure preservation be experimentally distinguished? 3-axis benchmark in Issue #10, integration policy in V8. There is a lack of test groups that simultaneously impose intervention, closed loop, and internal state consistency.
U14 Implementation of retestability: Is it possible to fix the operation so that a third party can reach the same conclusion using the same data and procedures? Requirements have been sorted out in P2/V9. The operational flow for publishing the execution environment, differential logs, and failure examples is incomplete.
U15 Social implementation governance: Can rights, consent, and responsibility distribution be designed in synchronization with technical evaluation? Issues have been extracted from D0-D2. The linkage rules (including cancellation criteria) between technical KPI and system KPI are undefined.
Judgment rules

U0-U15 above are not "good ideas" but "falsifiable research questions". Each problem is determined to have been ``solved'' only when the inputs, outputs, evaluation indicators, and failure conditions are all complete.

Additional breakdown of the three most important issues (deeper)

A

U4 causal equivalence

  • How to create a minimally complete set of intervention dictionaries (stimulus, task, pharmacology).
  • How to calibrate the thresholds of counterfactual matching (KL, JS, FIM, etc.).
  • How to suppress overestimation in sham intervention/sham conditions.
  • How to pre-register the Stop Rule in case of failure.
B

U10 Thermodynamic consistency

  • In what time window should we measure the dissipation rate of information processing?
  • How to deal with the difference between physical irreversibility and computational irreversibility.
  • Is it possible to create a common index that can be compared even if the implementation platform changes?
  • How to optimize the trade-off between energy efficiency and causal performance.
C

U12 Branched identity

  • Where to place the branch point definition (fixed initial condition).
  • How to set the evaluation attribution period after branching.
  • How to manage responsibility distribution and audit ID when there are multiple branches.
  • How to deal with the mismatch between legal subjectivity and technical identity.

Previous research map (by priority area)

Prior research necessary to solve unresolved problems is organized by field. The purpose is not to ``list the literature,'' but to clarifywhich evidence works for which problem.

area Representative previous research (excerpt) Contribution to U0-U15
WBE basics/personality Sandberg & Bostrom (2008), Yamakawa et al. (2024), Parfit (1984), Clowes (2021), Weber (2025) U0, U12, U15
Consciousness theory and verification Tononi (2015), Tononi et al. (2016), Albantakis et al. (2023), Doerig et al. (2019), Ferrante et al. (2025), Casali et al. (2013), Comolatti et al. (2019) U0, U4, U11
EEG/MEG inverse problem Michel & Brunet (2019), Wipf & Nagarajan (2009), Cai et al. (2021), Sechet et al. (2025), Vorwerk et al. (2014), Medani et al. (2025) U1, U2, U7
Preprocessing/synchronization/connectivity Wilson et al. (2010), Thompson et al. (2013), Mowla et al. (2017), Belinskaia et al. (2020), Mansouri et al. (2018), Zrenner et al. (2018), Appelhoff & Stenner (2021), Kothe et al. (2025) U1, U7, U8, U14
Decoding Huth et al. (2016), Tang et al. (2023), Défossez et al. (2023), Horikawa (2025), d'Ascoli et al. (2025), Willett et al. (2023), Littlejohn et al. (2025), Wairagkar et al. (2025), Wilson et al. (2025) U4, U8, U9, U13
Causal/active reasoning Friston et al. (2003), Friston (2010), Friston (2017), Parr & Friston (2019), Laukkonen et al. (2025), Correa et al. (2021) U4, U5, U9
Biological constraints Cook et al. (2019), Scheffer et al. (2020), Kasthuri et al. (2015), Santello et al. (2019), Ozcete et al. (2024), Gamlin et al. (2025) U2, U3, U6, U7
Measurement extension (OPM etc.) Boto et al. (2018), Seymour et al. (2021), Holmes et al. (2023), Holmes et al. (2025), Rhodes et al. (2025) U2, U7, U8
Standardization/Reproducibility Gorgolewski et al. (2016), Pernet et al. (2019), EEG-BIDS update, Open benchmark practices U14
Ethics, rights, legal system Yuste et al. (2017), Ienca & Andorno (2017), MIND Act discourse (2025) U12, U15
Thermodynamics/computability Seifert (2012), Kitazono et al. (2018), complexity reduction studies on IIT (2025) U10, U11

Reading order (for practical use)

  • First layer: Solidify U1/U14 (measurement and reproducibility) first. If this is weak, all higher-level discussions will be unstable.
  • Second layer: Verify U4/U9 (causal and generalization) on an intervention basis and establish the difference from imitation.
  • Third layer: Update U10/U11/U12/U15 (physics, identity, system) by connecting it with the empirical results.

Template for scientific core question page (recommended)

When creating a new technical proposal or question page, we recommend copying and using the Markdown template below. It is configured to force a thought flow that fixes R0 (restoration target) and V0 (verification criteria) first.

View template (click to expand)
## Template for scientific core question page (all items included/can be copied and pasted as is)

> Purpose: **"Fix R0 (recovery target) and V0 (verification criteria) first, then discuss M0 (measurement) and I0 (identifiability)" **This is a template that enforces the order.
> *Each section is clearly marked as “unconfirmed if it is unconfirmed” so that you can proceed.

---

# 0. Title

* **Page title**:
* **One-sentence summary (conclusion first)**:

  * "This page identifies the necessary measurement requirements (spatiotemporal resolution, observed variables) in order to be able to restore _____ to the standard of _____."

---

# 1. Central Question (CQ)

* **Central question**:

  * “At what spatio-temporal resolution should _____ be acquired to bring the “object to be restored (R0)” closer to being unique?”
* **Focus of question** (which is the main battlefield):

  * [ ] Selection of observed variables (what to measure)
  * [ ] Lower limit of resolution/coverage (how fine/wide)
  * [ ] Tolerable range for noise, missing measurements, and individual differences
  * [ ] Identification including causality (possibility of intervention)

---

# 2. Fixed terminology/scope (disambiguation elimination)

* **Target species**: Human/Animal (species: ___)/Simulation
* **Status range**: At rest/on task (task: _____)/sleep/pharmacological intervention
* **Time scale**: milliseconds to seconds/minutes to hours/days to years
* **Spatial scale**: cell/column/region/whole brain
* **Meaning of “restoration”** (by definition, not intuition): ____

---

# 3. R0: Restore target (most important/determined first)

## 3.1 What to restore (definition of target)

* **Restore target R0**:

  *Example) Structure (wire connections) / dynamics (state transitions) / functions (input → output) / memory / strategies / self-models including subjective reports, etc.
* **R0 output format** (What should be output in the end to be considered “restored”):

  * Examples) Parameter sets, generative models, state space models, agent policies, simulators, etc.

## 3.2 Definition of identity (equivalence relation) = “success”

* **Equivalence relation (success if what is the same)**:

  * “If ____ matches, R0 is considered the same.”
* **Tolerance** (No matter how different they are, they are treated the same):

  * Indicator: ____, Threshold: ____

## 3.3 Assumptions (assumptions implicitly included in R0)

*Example) The brain is Markovian/observable latent states exist/can be expressed as a learnable function class, etc.
* **List of assumptions**:

  * A1:____
  * A2:____
  * A3:____

---

# 4. V0: Verification criteria (pass/fail test)

## 4.1 What constitutes “recovery” (test design)

* **Verification task**: ____
* **Evaluation data**: Within-individual/inter-individual generalization/conditional extrapolation (counterfactual)
* **Baseline**: ____ (e.g. simple model, existing method)

## 4.2 Indicators (quantitative) and passing lines (thresholds)

* **Main indicators**: ____ (e.g. prediction accuracy, likelihood, causal effect agreement, behavioral distribution distance)
* **Passing line**: ____ (threshold or ratio)
* **Reproducibility requirements**: n = ___, recall rate ___, presence or absence of pre-registration ___

## 4.3 When using intervention/causation (recommended: include if possible)

* **Intervention measures**: stimulation/pharmacology/injury/closed-loop control
* **Concordance condition at the time of intervention**: ____ (definition of “response matches”)

---

# 5. M0: Measurement premise (MEASUREMENT)

## 5.1 Assumptions of measurement route (which one to assume)

* **Assumption**:

  * [ ] in vivo (non-invasive)
  * [ ] in vivo (invasive)
  * [ ] Destructive scan (ex vivo / future vision)
* **Safety/Ethics/Scale constraints**: ____

## 5.2 Observation target (what to measure) candidate list

* Electricity: EEG/MEG/ECoG/single/multipoint
*Blood flow: fMRI/fNIRS
*Optics: Ca imaging, etc.
* Structure: MRI/DTI/EM etc.
* **Observables mainly treated on this page**: ____

## 5.3 Expected spatiotemporal resolution/coverage

* **Temporal resolution**: ____
* **Spatial resolution**: ____
* **Coverage**: local/wide area/whole brain
* **Sampling constraints (duration/number of times)**: ____
* **Noise floor/missing**: ____

---

# 6. P0: Estimated model/representation (specify model assumptions)

* **Latent representation to be estimated**: state space/factor model/generative model/agent policy/connectome, etc.
* **Learning paradigm**: Supervised/Self-supervised/Data assimilation/Simulator combination
* **Required induction bias**: _____ (e.g. low dimensionality, sparsity, symmetry, anatomical constraints)
* **Computational resources/implementation constraints**: ____

---

# 7. I0: Identifiability (the “core” of this page)

## 7.1 Claim (in 1-2 sentences)

* **I0 claim**:

  * "In order to bring R0 close to uniqueness, it is necessary to obtain the observable quantity ____ in time and space ____ (at least ____ is the lower limit)."

## 7.2 Required conditions (lower limit: impossible without this)

* **Requirements (minimum set of observations)**: ____
* **Intuition/Reason**: Unobserved degrees of freedom remain/becomes a many-to-one mapping/become indivisible due to noise, etc.
* **Specific example of “unidentifiable”**: ____ (Another solution that produces the same observation can be constructed, etc.)

## 7.3 Sufficient condition (upper limit: theoretically possible if this is achieved)

* **Sufficient condition**: ____
* **Reality**: Impossible now/Possible in the future/Alternatives

## 7.4 Failure modes (important: enumerated proactively)

* [ ] Many-to-one (different R0 yields the same observation)
* [ ] Model misspec (P0 is wrong)
* [ ] Cause and effect cannot be determined due to lack of intervention
* [ ] Generalization breaks down due to individual differences
* [ ] Unidentifiable due to noise floor/missing measurements
* **Detection method**: ____ (Which indicator should be used to determine “failure”)

## 7.5 Measure of “approaching uniqueness”

* **Uniqueness score** (candidate): concentration of posterior distribution/equivalence class size/discrimination distance, etc.
* **Measures to be adopted**: ____

---

# 8. Strategic branching (A/B/C) *Can be used as an example

> For each branch, be sure to write a set of not only the "difference in M0" but also the **route to reach V0** and **probability of satisfying I0**.

## A: Train the model non-invasively (EEG/MEG/fMRI, etc.)

* **Aim**: ____ (Example: Gain generalization with large-scale, long-term data)
* **M0**: ____ (resolution/coverage/constraints)
* **P0**: ____ (expression/learning)
* **Route to reach V0**: ____ (which test to pass)
* **Bottleneck on I0**: ____ (What tends to make it impossible to identify)
* **“Judgment information” obtained from this branch**: ____ (information that is effective in selecting the next branch)

## B: Invasion (ECoG/deep/animal) enhances causal inference

* **Aim**: ____ (Example: Strengthen causality/identification through intervention)
* **M0**:____
* **P0**:____
* **Route to reach V0**: ____
* **Improvements on I0**: ____ (crushing many-to-one with intervention, etc.)
* **Handling of extrapolation (human generalization)**: ____

## C: Obtain structure with destructive and ultra-high resolution (future vision)

* **Aim**: ____ (Example: Consider upper limit on sufficient condition side)
* **M0**: ____ (obtainable structure information)
* **P0**: ____ (structure → function mapping assumption)
* **Route to reach V0**: ____
* **I0's conclusion**: ____ (Is it sufficient in principle/still insufficient?)

---

# 9. Rebuttal conditions/Stop rule

## 9.1 Total falsification (condition denying the root of the page)

* **Rebuttal conditions**:

  * “Even if only the measurement is advanced without defining R0 (recovery target), the inability to identify will not be resolved.”
  * Specifically: ____ (the same observation can form another R0, it does not pass through V0, etc.)

## 9.2 Stopping rules by branch (withdrawal line)

* Stop of A: ____
* Stop B: ____
* Stop C: ____
* **Pivot condition**: ____ (judgment criteria for A → B, B → C, etc.)

---

# 10. Next action (fix priority on this page)

* **Confirm first**:

  1. R0 (restore target/equivalence relationship)
  2. V0 (pass/fail test)
  3. I0 (lower limit assertion of necessary conditions)
* **Things to do this week**: ____
* **Next section to add**: ____
* **Undecided points (list of questions)**:

  * Q1:____
  * Q2:____

---

## Appendix: “Summary of main points” frame for overview on one page (optional)

* **R0**:____
* **V0**:____
* **M0**:____
* **P0**:____
* **I0 Conclusion (lower limit/upper limit)**: ____
* **Branch to adopt**: A/B/C (Reason: ____)
* **Disproval/Suspension**: ____

---

Learning order (minimum overall picture)

“Order” is important. Even if you are strong in measurement, you will not be able to claim progress if your verification standards are ambiguous. On the other hand, even if the verification is excellent, if the restoration target cannot be defined, the implementation will go astray.

  1. Overall picture:WBE roadmap (overall process/bottleneck)
  2. Measurement:What is lost at what resolution (EEG/fMRI/invasive/connectome)
  3. Reconstruction: Inverse problem → state estimation → generative model → intervention prediction, and proceed to “falsifiable reconstruction”
  4. Implementation:Fix model granularity and enable auditing of impact of execution base differences
  5. Verification:Write the evaluation suite and counter-conditions first, and keep the update history

List of prior research (expanded version)

Below are the main documents directly related to the open questions of U0-U15. For a detailed list of extensions, please read Paper Collection and Perspective References.

A. WBE basics, personality, philosophy

  1. Sandberg, A., & Bostrom, N. (2008). Whole Brain Emulation: A Roadmap.
  2. Yamakawa, H., et al. (2024). Technology roadmap toward the completion of whole-brain architecture.
  3. Parfit, D. (1984). Reasons and Persons.
  4. Whitehead, A. N. (1929). Process and Reality.
  5. Clowes, R. W. (2021). Slow Continuous Mind Uploading.
  6. Weber, C. (2025). The multiplicity objection against uploading optimism.

B. Consciousness theory/verification index

  1. Tononi, G. (2015). Integrated information theory.
  2. Tononi, G., et al. (2016). IIT: from consciousness to its physical substrate.
  3. Albantakis, L., et al. (2023). IIT 4.0.
  4. Doerig, A., et al. (2019). The unfolding argument.
  5. Boly, M., et al. (2017). Front vs back NCC debate.
  6. Tsuchiya, N., et al. (2015). No-report paradigms.
  7. Casali, A. G., et al. (2013). PCI.
  8. Comolatti, R., et al. (2019). PCI-ST.
  9. Ferrante, O., et al. (2025). Adversarial testing of GNWT and IIT.
  10. Cohen, M. A., et al. (2024). Neural signatures of visual awareness independent of postperceptual processing.
  11. Fahrenfort, J. J., et al. (2025). Criterion placement threatens the construct validity of neural measures of conscious contents.
  12. Casarotto, S., et al. (2024). Dissociations between spontaneous electroencephalographic features and perturbational complexity index in the minimally conscious state.
  13. Maschke, C., et al. (2024). Critical dynamics in spontaneous EEG predict anesthetic-induced loss of consciousness and perturbational complexity.
  14. Yoon, B., et al. (2024). Tracking the topology of neural manifolds across populations.
  15. The topological structure of population activity in mouse visual cortex encodes visual stimuli. (2024).
  16. Massimini, M., et al. (2005). Breakdown of effective connectivity during sleep.

C. Measurement/inverse problems/uncertainty estimation

  1. Michel, C. M., & Brunet, D. (2019). EEG source imaging review.
  2. Wipf, D., & Nagarajan, S. (2009). Unified Bayesian framework for MEG/EEG source imaging.
  3. Cai, C., et al. (2021). Robust noise estimation with Champagne. doi:10.1016/j.neuroimage.2020.117411
  4. Sechet, D., Kowalski, M., Mokhtari, S., & Torrésani, B. (2025). Revisiting CHAMPAGNE: Sparse Bayesian Learning as Reweighted Sparse Coding. doi:10.1109/SAMPTA64769.2025.11133512
  5. Feng, Z., Mishne, G., Hashemi, A., et al. (2025). Block-Champagne: Imaging Extended E/MEG Source Activation with Empirical Bayesian Uncertainty Quantification. doi:10.1109/TMI.2025.3642620
  6. Vorwerk, J., et al. (2014). Head volume conductor modeling guideline.
  7. Aydin, U., Vorwerk, J., Küpper, P., et al. (2019). Influence of head tissue conductivity uncertainties on EEG dipole reconstruction. doi:10.3389/fnins.2019.00531
  8. Mikulan, E., Russo, S., Bares, M., et al. (2020). Simultaneous human intracerebral stimulation and HD-EEG, ground-truth for source localization methods. doi:10.1038/s41597-020-0467-x
  9. Seeber, M., Cantonas, L.-M., Hoevels, M., et al. (2019). Subcortical electrophysiological activity is detectable with high-density EEG source imaging. doi:10.1038/s41467-019-08725-w
  10. Unnwongse, K., Achakulvisut, T., Wu, J. Y., et al. (2023). Direct validation of EEG source imaging by intracranial electric stimulation in human patients. doi:10.1093/braincomms/fcad023
  11. Hao, S., Zhao, H., Feng, Z., et al. (2025). HD-EEG source imaging with simultaneous SEEG recording in drug-resistant epilepsy. doi:10.1111/epi.18552
  12. Logothetis, N. K. (2008). What we can do and what we cannot do with fMRI. doi:10.1038/nature06976
  13. Purdon, P. L., et al. (2013). EEG signatures of loss/recovery of consciousness.
  14. Boto, E., Holmes, N., Leggett, J., et al. (2018). Moving magnetoencephalography towards real-world applications with a wearable system. doi:10.1038/nature26147
  15. Seymour, R. A., Alexander, N., Mellor, S., O'Neill, G. C., Tierney, T. M., Barnes, G. R., & Maguire, E. A. (2021). Using OPMs to measure neural activity in standing, mobile participants. doi:10.1016/j.neuroimage.2021.118604
  16. Holmes, N., Rea, M., Hill, R. M., et al. (2023). Enabling ambulatory movement in wearable magnetoencephalography with matrix coil active magnetic shielding. doi:10.1016/j.neuroimage.2023.120157
  17. Holmes, N., Rea, M., Hill, R. M., et al. (2023). Naturalistic hyperscanning with wearable magnetoencephalography. doi:10.3390/s23125454
  18. Holmes, N., Leggett, J., Hill, R. M., et al. (2025). Wearable magnetoencephalography in a lightly shielded environment. doi:10.1109/TBME.2024.3465654
  19. Iivanainen, J., Borna, A., Zetter, R., et al. (2022). Calibration and Localization of Optically Pumped Magnetometers Using Electromagnetic Coils. doi:10.3390/s22083059
  20. Rhodes, N., Rier, L., Boto, E., Hill, R. M., & Brookes, M. J. (2025). Source reconstruction without an MRI using optically pumped magnetometer-based magnetoencephalography. doi:10.1162/IMAG.a.8
  21. Wu, T., Xiao, W., Peng, X., Wu, T., & Guo, H. (2025). Crosstalk reduction in optically pumped magnetometers arrays for biomagnetic measurement. doi:10.1063/5.0273491
  22. Shapson-Coe, A., Januszewski, M., Berger, D. R., et al. (2024). A petavoxel fragment of human cerebral cortex reconstructed at nanoscale resolution. doi:10.1126/science.adk4858
  23. Lu, X., Han, X., Meirovitch, Y., et al. (2023). Preserving extracellular space for high-quality optical and ultrastructural studies of whole mammalian brains. doi:10.1016/j.crmeth.2023.100520
  24. Johansen, A., Beliveau, V., Colliander, E., et al. (2024). An In Vivo High-Resolution Human Brain Atlas of Synaptic Density. doi:10.1523/JNEUROSCI.1750-23.2024
  25. van Blooijs, D., Nunes, A., van den Boom, M. A., et al. (2023). Developmental trajectory of transmission speed in the human brain. doi:10.1038/s41593-023-01272-0
  26. Baadsvik, E. L., Weiger, M., Froidevaux, R., Schildknecht, C. M., Ineichen, B. V., & Pruessmann, K. P. (2024). Myelin bilayer mapping in the human brain in vivo. doi:10.1002/mrm.29998
  27. Lucchetti, F., Céléreau, E., Steullet, P., et al. (2025). Constructing the human brain metabolic connectome with MR spectroscopic imaging reveals cerebral biochemical organization. doi:10.1038/s41467-025-66124-w
  28. Ren, J., Sherry, A. D., & Malloy, C. R. (2015). 31P-MRS of healthy human brain: ATP synthesis, metabolite concentrations, pH, and T1 relaxation times. doi:10.1002/nbm.3384
  29. Guo, R., Yang, S., Wiesner, H. M., Li, Y., Zhao, Y., Liang, Z.-P., Chen, W., & Zhu, X.-H. (2024). Mapping intracellular NAD content in entire human brain using phosphorus-31 MR spectroscopic imaging at 7 Tesla. doi:10.3389/fnins.2024.1389111
  30. Kaiser, A., Vind, F. A., Duarte, J. M. N., Jelescu, I., Lin, Y., Yu, X., Widmaier, M., Wenz, D., & Xin, L. (2026). Ultra-high field 31P functional magnetic resonance spectroscopy reveals NAD+ dynamics in brain energy metabolism during visual stimulation. doi:10.1177/0271678X261415784
  31. Li, X., Zhu, X.-H., Li, Y., et al. (2025). Quantitative mapping of key glucose metabolic rates in the human brain using dynamic deuterium magnetic resonance spectroscopic imaging. doi:10.1093/pnasnexus/pgaf072
  32. Karkouri, J., Novoselova, M., Rodgers, C. T., et al. (2026). Absolute Quantification of Brain Deuterium Metabolic Imaging in Healthy Volunteers and Glioblastoma Patients at 7T. doi:10.1002/mrm.70308
  33. Hagiwara, A., Hori, M., Kamagata, K., et al. (2018). Myelin Measurement: Comparison Between Simultaneous Tissue Relaxometry, Magnetization Transfer Saturation Index, and T1w/T2w Ratio Methods. doi:10.1038/s41598-018-28852-6
  34. Galbusera, R., Weigel, M., Bahn, E., et al. (2025). Quantitative T1 is sensitive to cortical remyelination in multiple sclerosis: A postmortem MRI study. doi:10.1111/bpa.70010
  35. Colaes, R., Radwan, A., Billiet, T., et al. (2026). Evaluating the T1w/FLAIR ratio as a proxy for myelin: Associations with myelin water imaging, diffusion metrics, and cognition. doi:10.1007/s00234-025-03875-9
  36. Morgan, C. A., Thomas, D. L., Shao, X., et al. (2024). Measurement of blood-brain barrier water exchange rate using diffusion-prepared and multi-echo arterial spin labelling: Comparison of quantitative values and age dependence. doi:10.1002/nbm.5256
  37. Padrela, B. E., Slivka, M., Sneve, M. H., et al. (2025). Blood-brain barrier water permeability across the adult lifespan: A multi-echo ASL study. doi:10.1016/j.neurobiolaging.2024.12.012
  38. Chung, K. J., Abdelhafez, Y. G., Spencer, B. A., et al. (2025). Quantitative PET imaging and modeling of molecular blood-brain barrier permeability. doi:10.1038/s41467-025-58356-7
  39. Villemagne, V. L., Harada, R., Dore, V., et al. (2022). First-in-Humans Evaluation of 18F-SMBT-1, a Novel 18F-Labeled Monoamine Oxidase-B PET Tracer for Imaging Reactive Astrogliosis. doi:10.2967/jnumed.121.263254
  40. Villemagne, V. L., Harada, R., Dore, V., et al. (2022). Assessing reactive astrogliosis with 18F-SMBT-1 across the Alzheimer disease spectrum. doi:10.2967/jnumed.121.263255
  41. Hiraoka, K., Mesfin, B., Wu, Y., et al. (2025). Kinetic and quantitative analysis of [18F]SMBT-1 PET imaging for monoamine oxidase B. doi:10.1007/s12149-025-02083-y
  42. Mesfin, B., Ishioka, Y., Ichinose, Y., et al. (2026). Whole-body biodistribution of [18F]SMBT-1: a novel PET tracer for monoamine oxidase B imaging in healthy humans. doi:10.1007/s12149-025-02144-2
  43. Matsuoka, K., Takado, Y., Kimura, Y., et al. (2026). Quantification of monoamine oxidase B expression with 11C-SL25.1188 for imaging reactive astrocytes in patients with Alzheimer's disease. doi:10.1007/s00259-025-07542-2
  44. Tyacke, R. J., Myers, J. F. M., Venkataraman, A., et al. (2018). Evaluation of 11C-BU99008, a PET ligand for the imidazoline2 binding site in human brain. doi:10.2967/jnumed.118.208009
  45. Livingston, N. R., Calsolaro, V., Hinz, R., et al. (2022). Relationship between astrocyte reactivity, using novel 11C-BU99008 PET, and glucose metabolism, grey matter volume and amyloid load in cognitively impaired individuals. doi:10.1038/s41380-021-01429-y
  46. Best, L. M., Truong, J., McCluskey, T., et al. (2026). MAO-B status in alcohol use disorder: a [11C]SL25.1188 PET imaging study of putative astrogliosis. doi:10.1038/s41380-025-03355-9
  47. Jaisa-Aad, M., Muñoz-Castro, C., Healey, M. A., Hyman, B. T., & Serrano-Pozo, A. (2024). Characterization of monoamine oxidase-B (MAO-B) as a biomarker of reactive astrogliosis in Alzheimer’s disease and related dementias. doi:10.1007/s00401-024-02712-2
  48. Hirschler, L., Runderkamp, B. A. R., Decker, A., et al. (2025). Region-specific drivers of CSF mobility measured with MRI in humans. doi:10.1038/s41593-025-02073-3
  49. Dagum, P., Elbert, D. L., Giovangrandi, L., et al. (2026). The glymphatic system clears amyloid beta and tau from brain to plasma in humans. doi:10.1038/s41467-026-68374-8
  50. Tallman, C. W., Siler, S. N., Lee, H., et al. (2025). Neuronal allocation and sparse coding of episodic memories in the human hippocampus. doi:10.1038/s41598-025-21967-7
  51. Huber, R., Mäki, H., Rosanova, M., Casarotto, S., Canali, P., Casali, A. G., Tononi, G., & Massimini, M. (2013). Human cortical excitability increases with time awake. doi:10.1093/cercor/bhs014
  52. Kuhn, M., Wolf, E., Maier, J. G., Mainberger, F., Feige, B., Schmid, H., et al. (2016). Sleep recalibrates homeostatic and associative synaptic plasticity in the human cortex. doi:10.1038/ncomms12455
  53. Fehér, K. D., Henckaerts, P., Hirsch, V., Bucsenez, U., Kuhn, M., Maier, J. G., et al. (2026). A nap can recalibrate homeostatic and associative synaptic plasticity in the human cortex. doi:10.1016/j.neuroimage.2026.121723

D. Preprocessing/synchronization/network analysis

  1. Chang, C.-Y., et al. (2018). Evaluation of Artifact Subspace Reconstruction.
  2. de Cheveigne, A. (2020). ZapLine.
  3. Vinck, M., et al. (2011). Weighted Phase Lag Index (wPLI).
  4. Staniek, M., & Lehnertz, K. (2008). Symbolic Transfer Entropy.
  5. Wilson, J. A., Mellinger, J., Schalk, G., & Williams, J. C. (2010). A procedure for measuring latencies in brain-computer interfaces. doi:10.1109/TBME.2010.2047259
  6. Thompson, D. E., Warschausky, S. A., & Huggins, J. E. (2013). Classifier-based latency estimation: a novel way to estimate and predict BCI accuracy. doi:10.1088/1741-2560/10/1/016006
  7. Mowla, M. R., Huggins, J. E., & Thompson, D. E. (2017). Enhancing P300-BCI performance using latency estimation. doi:10.1080/2326263X.2017.1338010
  8. Belinskaia, A., Smetanin, N., Lebedev, M., & Ossadtchi, A. (2020). Short-delay neurofeedback facilitates training of the parietal alpha rhythm. doi:10.1088/1741-2552/abc8d7
  9. Mansouri, F., Fettes, P., Schulze, L., et al. (2018). A real-time phase-locking system for non-invasive brain stimulation. doi:10.3389/fnins.2018.00877
  10. Zrenner, C., Desideri, D., Belardinelli, P., & Ziemann, U. (2018). Real-time EEG-defined excitability states determine efficacy of TMS-induced plasticity in the human motor cortex. doi:10.1016/j.brs.2017.11.016
  11. Khatri, U. U., Pulliam, K., Manesiya, M., Vieyra Cortez, M., Millán, J. del R., & Hussain, S. J. (2025). Personalized whole-brain activity patterns predict human corticospinal tract activation in real-time. doi:10.1016/j.brs.2024.12.1193
  12. Holt, A. B., Kormann, E., Gulberti, A., et al. (2019). Phase-Dependent Suppression of Beta Oscillations in Parkinson's Disease Patients. doi:10.1523/JNEUROSCI.1913-18.2018
  13. Zrenner, C., Galevska, D., Nieminen, J. O., Baur, D., Stefanou, M. I., & Ziemann, U. (2020). The shaky ground truth of real-time phase estimation. doi:10.1016/j.neuroimage.2020.116761
  14. Gordon, P. C., Dörre, S., Belardinelli, P., Stenroos, M., Zrenner, B., Ziemann, U., & Zrenner, C. (2021). Prefrontal Theta-Phase Synchronized Brain Stimulation With Real-Time EEG-Triggered TMS. doi:10.3389/fnhum.2021.691821
  15. Bruegger, D., & Abegg, M. (2021). Prediction of cortical theta oscillations in humans for phase-locked visual stimulation. doi:10.1016/j.jneumeth.2021.109288
  16. Vigué-Guix, I., Morís Fernández, L., Torralba Cuello, M., Ruzzoli, M., & Soto-Faraco, S. (2022). Can the occipital alpha-phase speed up visual detection through a real-time EEG-based brain-computer interface (BCI)? doi:10.1111/ejn.14931
  17. Kim, B., Erickson, B. A., Fernandez-Nunez, G., Rich, R., Mentzelopoulos, G., Vitale, F., & Medaglia, J. D. (2023). EEG Phase Can Be Predicted with Similar Accuracy across Cognitive States after Accounting for Power and Signal-to-Noise Ratio. doi:10.1523/ENEURO.0050-23.2023
  18. Little, S., Pogosyan, A., Neal, S., et al. (2013). Adaptive deep brain stimulation in advanced Parkinson disease. doi:10.1002/ana.23951
  19. Tinkhauser, G., Pogosyan, A., Little, S., et al. (2017). The modulatory effect of adaptive deep brain stimulation on beta bursts in Parkinson's disease. doi:10.1093/brain/awx010
  20. Hernandez-Pavon, J. C., Metsomaa, J., Mutanen, T. P., et al. (2023). TMS combined with EEG: Recommendations and open issues. doi:10.1016/j.brs.2023.02.009
  21. Hougland, J. R., Kirchhoff, M., Vetter, D. E., Ahola, O., Jooß, A., Humaidan, D., & Ziemann, U. (2025). Fluctuations in the optimal sensorimotor mu-rhythm phase associated with high corticospinal excitability during TMS-EEG. doi:10.1016/j.brs.2025.09.019
  22. Appelhoff, S., & Stenner, T. (2021). In COM we trust: Feasibility of USB-based event marking. doi:10.3758/s13428-021-01571-z
  23. Kothe, C., Shirazi, S. Y., Stenner, T., et al. (2025). The lab streaming layer for synchronized multimodal recording. doi:10.1162/IMAG.a.136

E. Decoding, generative model, imitation separation

  1. Huth, A. G., de Heer, W. A., Griffiths, T. L., Theunissen, F. E., & Gallant, J. L. (2016). Natural speech reveals the semantic maps that tile human cerebral cortex. doi:10.1038/nature17637
  2. Tang, J., LeBel, A., Jain, S., et al. (2023). Semantic reconstruction of continuous language from non-invasive brain recordings. doi:10.1038/s41593-023-01304-9
  3. Défossez, A., Caucheteux, C., Rapin, J., Kabeli, O., & King, J.-R. (2023). Decoding speech perception from non-invasive brain recordings. doi:10.1038/s42256-023-00714-5
  4. Horikawa, T. (2025). Mind captioning: Evolving descriptive text of mental content from human brain activity. doi:10.1126/sciadv.adw1464
  5. d'Ascoli, S., Ferrante, O., et al. (2025). Towards decoding individual words from non-invasive brain recordings. doi:10.1038/s41467-025-65499-0
  6. Willett, F. R., Kunz, E. M., Fan, C., et al. (2023). A high-performance speech neuroprosthesis. doi:10.1038/s41586-023-06377-x
  7. Littlejohn, K. T., Dabagia, M., Ladwig, A., et al. (2025). A streaming brain-to-voice neuroprosthesis to restore naturalistic communication. doi:10.1038/s41593-025-01905-6
  8. Wairagkar, M., Card, N. S., Singer-Clark, T., et al. (2025). An instantaneous voice-synthesis neuroprosthesis. doi:10.1038/s41586-025-09127-3
  9. Singh, A., Wu, E., Ramsey, N. F., et al. (2025). Transfer learning via distributed brain recordings enables reliable speech decoding. doi:10.1038/s41467-025-63825-0
  10. Wilson, G. H., Stein, E. A., Kamdar, F., et al. (2025). Long-term unsupervised recalibration of cursor-based intracortical brain–computer interfaces using a hidden Markov model. doi:10.1038/s41551-025-01536-z

E2. Longitudinal validity, state / trait / drift, and recalibration

  1. Musall, S., Kaufman, M. T., Juavinett, A. L., Gluf, S., & Churchland, A. K. (2019). Single-trial neural dynamics are dominated by richly varied movements. doi:10.1038/s41593-019-0502-4
  2. Benisty, H., Barson, D., Moberly, A. H., et al. (2024). Rapid fluctuations in functional connectivity of cortical networks encode spontaneous behavior. doi:10.1038/s41593-023-01498-y
  3. Egger, A., Bayon, C., d'Almeida, J., et al. (2024). Chrono-EEG dynamics influencing hand gesture decoding: a 10-hour study. doi:10.1038/s41598-024-70609-x
  4. Gallego, J. A., Perich, M. G., Chowdhury, R. H., Solla, S. A., & Miller, L. E. (2020). Long-term stability of cortical population dynamics underlying consistent behavior. doi:10.1038/s41593-019-0555-4
  5. Noda, T., Kienle, E., Eppler, J.-B., et al. (2025). Homeostasis of a representational map in the neocortex. doi:10.1038/s41593-025-01982-7
  6. Karpowicz, B. M., Ali, Y. H., Wimalasena, L. N., et al. (2025). Stabilizing brain-computer interfaces through alignment of latent dynamics. doi:10.1038/s41467-025-59652-y

F. Causal inference/active inference/counterfactual hypothetical

  1. Friston, K. J., Harrison, L., & Penny, W. (2003). Dynamic causal modelling. doi:10.1016/S1053-8119(03)00202-7
  2. Penny, W. D., Stephan, K. E., Mechelli, A., & Friston, K. J. (2004). Comparing dynamic causal models. doi:10.1016/j.neuroimage.2004.03.026
  3. Rosa, M. J., Friston, K., & Penny, W. (2012). Post-hoc selection of dynamic causal models. doi:10.1016/j.jneumeth.2012.04.013
  4. Friston, K. (2010). Free-energy principle.
  5. Friston, K. (2017). Active inference: a process theory.
  6. Parr, T., & Friston, K. J. (2019). Generalised free energy.
  7. Lee, H.-L., Zahneisen, B., Hugger, T., et al. (2017). Tracking dynamic effective connectivity from fMRI using changes induced by anesthesia. doi:10.1016/j.neuroimage.2017.02.012
  8. Frässle, S., Manjaly, Z. M., Do, C. T., Kasper, L., Pruessmann, K. P., & Stephan, K. E. (2021). Whole-brain estimates of directed connectivity for human connectomics. doi:10.1016/j.neuroimage.2020.117491
  9. Wu, H., Hu, X., & Zeng, Y. (2024). A fast dynamic causal modeling regression method for fMRI. doi:10.1016/j.neuroimage.2024.120954
  10. Jafarian, A., Assem, M. K., Kocagoncu, E., et al. (2024). Reliability of dynamic causal modelling of resting-state magnetoencephalography. doi:10.1002/hbm.26782
  11. Hauser, A., & Bühlmann, P. (2012). Characterization and greedy learning of interventional Markov equivalence classes of directed acyclic graphs. JMLR
  12. Vink, J. J., Ramos-Nuñez, A. I., Bellesi, A., et al. (2020). The brain's functional connectome is a poor predictor of the brain's causal activity flow. doi:10.1371/journal.pcbi.1007866
  13. Villaverde, A. F. (2019). Observability and Structural Identifiability of Nonlinear Biological Systems. doi:10.1155/2019/8497093
  14. Villaverde, A. F., Tsiantis, N., & Banga, J. R. (2019). Full observability and estimation of unknown inputs, states and parameters of nonlinear biological models. doi:10.1098/rsif.2019.0043
  15. Massonis, G., & Villaverde, A. F. (2020). Finding and breaking Lie symmetries: implications for structural identifiability and observability in biological modelling. Symmetry, 12(3), 469. doi:10.3390/sym12030469
  16. Diop, S., & Fliess, M. (1991). Nonlinear observability, identifiability, and persistent trajectories. Proceedings of the 30th IEEE Conference on Decision and Control, 714-719. doi:10.1109/CDC.1991.261405
  17. Raue, A., Becker, V., Klingmüller, U., & Timmer, J. (2010). Identifiability and observability analysis for experimental design in nonlinear dynamical models. Chaos, 20(4), 045105. doi:10.1063/1.3528102
  18. Chis, O.-T., Villaverde, A. F., Banga, J. R., & Balsa-Canto, E. (2016). On the relationship between sloppiness and identifiability. Mathematical Biosciences, 282, 147-161. doi:10.1016/j.mbs.2016.10.009
  19. White, A., Tolman, M., Thames, H. D., Withers, H. R., Mason, K. A., & Transtrum, M. K. (2016). The limitations of model-based experimental design and parameter estimation in sloppy systems. PLOS Computational Biology, 12(12), e1005227. doi:10.1371/journal.pcbi.1005227
  20. Gevertz, J. L., & Kareva, I. (2024). Minimally sufficient experimental design using identifiability analysis. npj Systems Biology and Applications, 10, 2. doi:10.1038/s41540-023-00325-1
  21. Liu, X., Wanika, L., Chappell, M. J., & Branke, J. (2025). Efficient data collection for establishing practical identifiability via active learning. Computational and Structural Biotechnology Journal, 27, 4992-5006. doi:10.1016/j.csbj.2025.10.058
  22. Rasero, J., Betzel, R., Sentis, A. I., Kraynak, T. E., Gianaros, P. J., & Verstynen, T. (2024). Similarity in evoked responses does not imply similarity in macroscopic network states. doi:10.1162/netn_a_00354
  23. Langdon, C., & Engel, T. A. (2025). Latent circuit inference from heterogeneous neural responses during cognitive tasks. Nature Neuroscience, 28, 665-675. doi:10.1038/s41593-025-01869-7
  24. Laukkonen, R., Friston, K., & Chandaria, S. (2025). A beautiful loop.
  25. Correa, J. D., Lee, S., & Bareinboim, E. (2021). Nested counterfactual identification.

G. Biological constraints (connectome, neuromodulation, glia)

  1. Cook, S. J., et al. (2019). Whole-animal connectomes of both C. elegans sexes.
  2. Scheffer, L. K., et al. (2020). Adult Drosophila central brain connectome.
  3. Kasthuri, N., et al. (2015). Saturated reconstruction of neocortex volume.
  4. Santello, M., et al. (2019). Astrocyte-neuron interactions.
  5. Ozcete, O. D., et al. (2024). Neuromodulatory volume transmission.
  6. Reimer, J., McGinley, M. J., Liu, Y., et al. (2016). Pupil fluctuations track rapid changes in adrenergic and cholinergic activity in cortex. doi:10.1038/ncomms13289
  7. Hansen, J. Y., Shafiei, G., Markello, R. D., et al. (2022). Mapping neurotransmitter systems to the structural and functional organization of the human neocortex. doi:10.1038/s41593-022-01186-3
  8. Goulas, A., Changeux, J.-P., Wagstyl, K., et al. (2021). The natural axis of transmitter receptor distribution in the human cerebral cortex. doi:10.1073/pnas.2020574118
  9. Koepp, M. J., Gunn, R. N., Lawrence, A. D., et al. (1998). Evidence for striatal dopamine release during a video game. doi:10.1038/30498
  10. Lippert, R. N., Cremer, A. L., Edwin Thanarajah, S., et al. (2019). Time-dependent assessment of stimulus-evoked regional dopamine release. doi:10.1038/s41467-018-08143-4
  11. Neyhart, E., Zhou, N., Munn, B. R., et al. (2024). Cortical acetylcholine dynamics are predicted by cholinergic axon activity and behavioral state. doi:10.1016/j.celrep.2024.114808
  12. Dorkenwald, S., et al. (2024). Neuronal wiring diagram of an adult brain. doi:10.1038/s41586-024-07558-y
  13. MICrONS Consortium, et al. (2025). Functional connectomics spanning multiple areas of mouse visual cortex. doi:10.1038/s41586-025-08790-w
  14. Prinz, A. A., Bucher, D., & Marder, E. (2004). Similar network activity from disparate circuit parameters. doi:10.1038/nn1352
  15. Beiran, M., & Litwin-Kumar, A. (2025). Prediction of neural activity in connectome-constrained recurrent networks. doi:10.1038/s41593-025-02080-4
  16. Gamlin, C. R., et al. (2025). Connectomics of predicted Sst transcriptomic types in mouse visual cortex. doi:10.1038/s41586-025-08805-6
  17. Cahill, M. K., et al. (2024). Network-level encoding of local neurotransmitters in cortical astrocytes. doi:10.1038/s41586-024-07311-5
  18. Lee, H.-K., Barbarosie, M., Kameyama, K., Bear, M. F., & Huganir, R. L. (2003). Regulation of distinct AMPA receptor phosphorylation sites during bidirectional synaptic plasticity. doi:10.1016/S0092-8674(03)00122-3
  19. Havekes, R., Park, A. J., Tolentino, R. E., et al. (2016). Compartmentalized PDE4A5 signaling impairs hippocampal synaptic plasticity and long-term memory. doi:10.1523/JNEUROSCI.0248-16.2016
  20. Vierra, N. C., et al. (2023). Endoplasmic reticulum-plasma membrane junctions couple electrical activity to Ca2+-activated PKA signaling in neurons. doi:10.1038/s41467-023-40930-6
  21. Altas, B., Tuffy, L. P., Patrizi, A., et al. (2024). Region-specific phosphorylation determines Neuroligin-3 localization to excitatory versus inhibitory synapses. doi:10.1016/j.biopsych.2023.12.020
  22. Rodriguez, A. C., Kramár, E. A., Augustynski, A. S., et al. (2025). HDAC3 Serine 424 phospho-mimic and phospho-null mutants bidirectionally modulate long-term memory formation and synaptic plasticity in the adult and aging mouse brain. doi:10.1523/JNEUROSCI.1619-24.2025
  23. Frey, U., & Morris, R. G. M. (1997). Synaptic tagging and long-term potentiation. doi:10.1038/385533a0
  24. Shires, K. L., Da Silva, B. M., Hawthorne, J. P., et al. (2012). Synaptic tagging and capture in the living rat. doi:10.1038/ncomms2250
  25. Govindarajan, A., Israely, I., Huang, S.-Y., & Tonegawa, S. (2011). The dendritic branch is the preferred integrative unit for protein synthesis-dependent LTP. doi:10.1016/j.neuron.2010.12.008
  26. Fonseca, R., Vabulas, R. M., Hartl, F. U., Bonhoeffer, T., & Nagerl, U. V. (2006). A balance of protein synthesis and proteasome-dependent degradation determines the maintenance of LTP. doi:10.1016/j.neuron.2006.08.015
  27. Pandey, K., Yu, X.-W., Steinmetz, A., & Alberini, C. M. (2021). Autophagy coupled to translation is required for long-term memory. doi:10.1080/15548627.2020.1775393
  28. Chang, Y. C., Gao, Y., Lee, J. Y., et al. (2024). Identification of secretory autophagy as a mechanism modulating activity-induced synaptic remodeling. doi:10.1073/pnas.2315958121
  29. Parker, D., Davidson, K., Osmulski, P. A., et al. (2025). Proteasome augmentation mitigates age-related cognitive decline in mice. doi:10.1111/acel.14492
  30. Correia, S. S., Bassani, S., Brown, T. C., et al. (2008). Motor protein-dependent transport of AMPA receptors into spines during long-term potentiation. doi:10.1038/nn2063
  31. Swarnkar, S., Avchalumov, Y., Espadas, I., et al. (2021). Molecular motor protein KIF5C mediates structural plasticity and long-term memory by constraining local translation. doi:10.1016/j.celrep.2021.109369
  32. Aiken, J., & Holzbaur, E. L. F. (2024). Spastin locally amplifies microtubule dynamics to pattern the axon for presynaptic cargo delivery. doi:10.1016/j.cub.2024.03.010
  33. de Queiroz, B. R., et al. (2025). Axonal RNA localization is essential for long-term memory. doi:10.1038/s41467-025-57651-7
  34. Stringer, C., Pachitariu, M., Steinmetz, N., et al. (2019). Spontaneous behaviors drive multidimensional, brainwide activity. doi:10.1126/science.aav7893
  35. Saleem, A. B., Ayaz, A., Jeffery, K. J., Harris, K. D., & Carandini, M. (2013). Integration of visual motion and locomotion in mouse visual cortex. doi:10.1038/nn.3567
  36. Ravassard, P., Kees, A., Willers, B., et al. (2013). Multisensory control of hippocampal spatiotemporal selectivity. doi:10.1126/science.1232655
  37. Zelano, C., Jiang, H., Zhou, G., et al. (2016). Nasal respiration entrains human limbic oscillations and modulates cognitive function. doi:10.1523/JNEUROSCI.2586-16.2016
  38. Raut, R. V., Rosenthal, Z. P., Wang, X., et al. (2025). Arousal as a universal embedding for spatiotemporal brain dynamics. doi:10.1038/s41586-025-09544-4

H. Standardization, reproducibility, ethics, system

  1. Gorgolewski, K. J., et al. (2016). BIDS.
  2. Pernet, C. R., et al. (2019). EEG-BIDS.
  3. Fleming, S. M., et al. (2023). Open letter on IIT interpretability and testability.
  4. Yuste, R., et al. (2017). Four ethical priorities for neurotechnologies and AI.
  5. Ienca, M., & Andorno, R. (2017). New human rights in neuroscience.
  6. Schumer, C., et al. (2025). MIND Act discourse.

I. Thermodynamics, irreversibility, energy budget

  1. Bérut, A., Arakelyan, A., Petrosyan, A., et al. (2012). Experimental verification of Landauer’s principle linking information and thermodynamics. doi:10.1038/nature10872
  2. Attwell, D., & Laughlin, S. B. (2001). An energy budget for signaling in the grey matter of the brain. doi:10.1097/00004647-200110000-00001
  3. Lynn, C. W., Cornblath, E. J., Papadopoulos, L., et al. (2021). Broken detailed balance and entropy production in the human brain. doi:10.1073/pnas.2109889118
  4. de la Fuente, L. A., Perl, Y. S., Zamberlan, F., et al. (2023). Temporal irreversibility of neural dynamics as a signature of consciousness. doi:10.1093/cercor/bhac177
  5. Nartallo-Kaluarachchi, R., Bonetti, L., Fernández-Rubio, G., et al. (2025). Multilevel irreversibility reveals higher-order organization of nonequilibrium interactions in human brain dynamics. doi:10.1073/pnas.2408791122
  6. Ishihara, K., & Shimazaki, H. (2025). State-space kinetic Ising model reveals task-dependent entropy flow in sparsely active nonequilibrium neuronal dynamics. doi:10.1038/s41467-025-66669-w
  7. Martínez, I. A., Bisker, G., Horowitz, J. M., & Parrondo, J. M. R. (2019). Inferring broken detailed balance in the absence of observable currents. doi:10.1038/s41467-019-11051-w
  8. Hartich, D., & Godec, A. (2024). Comment on “Inferring broken detailed balance in the absence of observable currents”. doi:10.1038/s41467-024-52602-0
  9. Martínez, I. A., Bisker, G., Horowitz, J. M., & Parrondo, J. M. R. (2024). Reply to: Comment on “Inferring broken detailed balance in the absence of observable currents”. doi:10.1038/s41467-024-52603-z
  10. Blom, K., Song, K., Vouga, E., Godec, A., & Makarov, D. E. (2024). Milestoning estimators of dissipation in systems observed at a coarse resolution. doi:10.1073/pnas.2318333121