Wiki

Wiki: Straight path from EEG to L0

Freeze benchmark object, temporal regime, and observation logs before touching the model

Mind Uploading Research Project

Public Page Updated: 2026-04-03 Practical route / temporal-validity sync

How to use this page

Read this first to avoid getting lost

This page is a practical route from EEG basics to a reproducible L0 artifact pack. The 2026-03 rebuild fixes one weakness in the old route: choosing a dataset and running a model still sounded too simple because benchmark object, temporal regime, event semantics, and verification gates were not kept separate enough.

  • The route is still short, but it is no longer a generic six-step checklist.
  • Dataset choice is now treated as benchmark design, not only file selection.
  • Subject/session split and same-day/cross-day regime are fixed before modeling, and temporal validity now separates fast labels from slow internal-milieu disclosure.
  • BIDS events, HED semantics, and LSL synchronization are treated as different layers.
  • A score does not become progress until benchmark object, temporal validity, lineage, and stopped claim are all visible.
Best for
People who want to move from EEG basics to reproducible L0 work without collapsing dataset choice, split hygiene, temporal validity, and verification into one step
Reading time
12-18 minutes
Accuracy note
This page does not claim that passing L0 solves identity, WBE, or higher-level state recovery. It is an entry point for reproducible analysis only. The stronger public rules still live on Verification.

Relatively clear at this stage

What we know now

  • L0 is about reproducible input, logs, preprocessing, baselines, and failure disclosure rather than high headline accuracy.
  • A starter dataset is not enough by itself; the independent hold-out unit, metric bundle, and benchmark-governance status also change what the score means.
  • Observation logs, derivative lineage, and temporal-validity fields are separate artifacts.
  • Current literature supports careful separation of subject/session shortcuts, acquisition-distribution shortcuts, clock-domain logs, and cross-day validity.

Still unresolved beyond this point

What we still do not know

  • Which starter route best prepares later L1-L2 work still depends on task family and target variable.
  • The field still lacks one universally accepted default card bundle for all EEG benchmarks.
  • How much temporal-validity disclosure will become standard across EEG benchmarks remains unsettled.

Learn the basics

Check the basics in the wiki

What the wiki is for

The wiki is a learning aid. For the project's official current synthesis, success criteria, and operating rules, always return to the public pages.

What this route now fixes first

The short path from EEG to L0 still exists, but it now starts with a stricter question: what exactly will your score mean? Current primary literature and official standards do not support the shortcut that says "pick a public dataset, run preprocessing, train a model, report accuracy." Before a score matters, this site now fixes benchmark object, temporal regime, event semantics and clock domain, artifact lineage, and the stopped claim.

Why this page needed a 2026-03 rebuild

The old route was too permissive. Pernet et al. (2019), the current BIDS specification, and Pernet et al. (2020) show why raw identity, derivatives, and reporting provenance must be explicit. Hermes et al. (2025) and Kothe et al. (2025) show why event semantics and synchronization still need separate audits. Chaibub Neto et al. (2019), Melnik et al. (2017), Xu et al. (2020), and Di et al. (2021) show why subject and acquisition shortcuts can survive loose evaluation. Egger et al. (2024) show that even within roughly half a day, EEG decoding conditions can shift enough to matter for robustness. The official EEG Challenge (2025) pages then show that benchmark governance itself can change what a leaderboard means. Therefore, this page no longer treats "dataset -> preprocessing -> score" as a sufficient beginner route.

Scope of this page

This page stays on the technical and natural science side. It does not argue about philosophy, law, or identity. It only fixes what must be observable, logged, and audited before an EEG result can count as reproducible L0 work on this site.

Six gates from EEG to L0

Order Page to open What is fixed here What must exist before moving on
1 EEG 101 Fix the measurement ceiling: what EEG directly observes, what remains latent, and what kind of claim it cannot support on its own. A one-line stopped claim such as "this route aims at reproducible macro-state analysis, not source-complete or WBE-complete recovery."
2 Datasets and Baseline / Benchmark / Pre-registration / Model Card Choose a benchmark object, not just a file bundle: task, target, independent hold-out unit, metric bundle, version, extra-data policy, and benchmark-governance status. A short benchmark card naming dataset/version, task, target, split unit, main metric bundle, and whether official rules or postmortems changed the benchmark meaning.
3 Dataset splits and data leakage and State, trait, and drift Freeze the temporal regime: subject, session, and time disjointness; same-session versus cross-day scope; fixed versus recalibrated decoder interval. A split manifest plus a temporal-validity note stating whether the result is same-session, same-day, cross-day, or longer-horizon, and whether the decoder stays fixed.
4 Event synchronization and observation logs Freeze the observation contract: event times, event semantics, label provenance, clock domain, delay/jitter/drift notes, and report-usage flags. An observation log that separates time anchor, semantics, and synchronization layer instead of mixing them into one generic metadata note.
5 Hands-on and L0 minimum artifact pack Produce the first reproducible artifact bundle: raw identity, derivative identity, run identity, QC, baseline output, and failure registry. A rerunnable derivative package with dataset provenance, command or pipeline provenance, environment pin, QC report, baseline output, and at least one named failure mode.
6 Verification Convert the artifact bundle into a bounded claim: L0 ceiling, observability ceiling, shortcut ceiling, and temporal-validity ceiling. A submission-ready stopped claim plus the required companion cards if the result starts to imply target specificity or temporal durability.

Why these gates must stay separate

What older beginner routes tended to compress What current sources actually support How this site now reads the route
"Dataset choice" as only a file download step Saito & Rehmsmeier (2015) show why metric choice changes what a binary score means, and the official EEG Challenge (2025) rules plus final leaderboard show that governance changes can alter benchmark meaning after launch. Choosing data now means choosing the benchmark object: task, target, split/randomization rule, metric bundle, version, and governance status.
"Clean split" as the whole leakage solution Chaibub Neto et al. (2019), Melnik et al. (2017), Xu et al. (2020), and Di et al. (2021) show that subject/session and acquisition-distribution structure can remain highly predictive. The route now fixes both split hygiene and shortcut resistance. A clean split is necessary, but it is not treated as proof that the target neural variable was isolated.
"Same-session score" as temporal generalization Egger et al. (2024) show that EEG decoding conditions can change materially across a day-night window, and this site's Temporal Validity Card plus state-trait-drift rule now separate fixed-decoder interval, fast labels, slow internal-milieu disclosure, and recalibration burden. The route now asks the reader to decide same-session, same-day, cross-day, or longer-horizon scope before training, and to log whether the regime changed through movement / arousal alone or through slower circadian / endocrine-metabolic state as well.
"Events are in BIDS" as if semantics and timing were solved together The current BIDS specification and Hermes et al. (2025) support structured events and machine-readable semantics, while Kothe et al. (2025) makes clear that synchronization middleware does not by itself measure device-side delay. This route now separates time anchor, event semantics, and clock/synchronization audit into distinct observation artifacts.
"Pipeline ran" as if provenance were sufficient Gorgolewski et al. (2016), Pernet et al. (2019), Pernet et al. (2020), and the current BIDS specification separate raw datasets, derivatives, and generated-by provenance. The route now requires raw identity, derivative identity, and run identity to be visible as different objects before L0 is called complete.

Minimum artifact bundle before one score matters

Artifact What it must disclose What goes wrong if it is missing
Benchmark object Dataset/version, task, target, independent hold-out unit, metric bundle, extra-data policy, and benchmark-governance status. A score can be overread as if it applied to a different task, split regime, or official rule set.
Split manifest Which subject/session/time units are disjoint, how folds were frozen, and which grouping variables were respected. The evaluation can drift silently as folds or grouping assumptions change.
Temporal-validity note Same-session versus cross-day scope, fixed versus recalibrated decoder, fast state labels, and any relevant slow internal-milieu disclosure. Same-day success can be silently promoted to longitudinal stability or deployability.
Observation log Event times, semantics, scorer or report provenance, clock domain, delay/jitter/drift notes, and bad-segment annotations. The route cannot distinguish a signal problem from a label or timing problem.
Derivative lineage Source dataset, generated-by pipeline, version or commit, environment pin, command provenance, and output locations. Reanalysis and audit become impossible even if the main score is reproducible once.
Stopped claim What the result supports and what it still does not support, in one or two sentences. L0 can be overread as source localization truth, stable biomarker evidence, or WBE-relevant state capture.
Operational inference used on this site

Not every field above is a single mandatory key in one external standard. The stronger requirement on this site is an operational inference from current standards, primary literature, and challenge practice: if a result is to count as comparable L0 progress, the benchmark object, temporal regime, observation contract, and derivative lineage all need their own artifacts rather than one mixed prose paragraph.

Five accidents this route now tries to stop early

Common accident Why it is scientifically weak Where to return
Choosing a starter dataset without naming the benchmark object The data may be fine, but the score will still be uninterpretable if task, metric bundle, and governance status are missing. Baseline / Benchmark / Pre-registration / Model Card
Using a subject/session split but not naming temporal scope The result can still be same-session or same-day only, even if the split sounds clean. State, trait, and drift
Treating `events.tsv` as if it fully solved label meaning Time anchors, condition semantics, and report-derived labels are different objects and can fail independently. Event synchronization and observation logs
Treating LSL or trigger lines as hardware ground truth Network synchronization does not automatically measure display, audio, amplifier, or device-internal delays. Event synchronization and observation logs
Reporting one score without lineage and failure disclosure The route becomes impossible to audit, extend, or compare even if the run once looked successful. L0 minimum artifact pack and Verification

Where to go next

If the measurement ceiling of EEG is still not clear, return to EEG 101. If the main uncertainty is split hygiene or benchmark provenance, go to Dataset splits and data leakage. If time anchors, label provenance, or synchronization are the problem, go to Event synchronization and observation logs. Once the first artifact bundle exists, route it through Verification and attach the Temporal Validity Card or Specificity & Shortcut Card whenever the claim starts to reach beyond plain reproducible analysis.

References