Skip to content

Thought-to-Text Status

"Mind typing / brain-to-text" is one of the earliest practical goals of BCI: letting the user type with nothing but thought. In 2024–2026 this goal reached milestones of 60+ WPM invasively and 30+ WPM non-invasively, but real-time, consumer-grade accuracy remains distant. This article compares the current state of the art.

1. Evaluation Metrics

Main Metrics

  • WPM (Words Per Minute): text-input speed
  • CPM (Characters Per Minute)
  • WER (Word Error Rate): accuracy
  • Latency: thought → display
  • Calibration time: training before first use

Baseline References

Input method Speed
Proficient keyboard 60–80 WPM
Smartphone 30–40 WPM
Siri voice 100+ WPM (theoretical)
Eye-gaze typing (tracker) 10–15 WPM
Early BrainGate 6 WPM (2006)
Willett 2021 handwriting BCI 90 CPM (~18 WPM)
Willett 2023 speech BCI 62 WPM
Metzger 2023 avatar 78 WPM

2. Invasive Status (2024–2025)

Top Performance

  • Willett 2023: 62 WPM, WER 9.1% (50-word vocabulary)
  • UC Davis Card 2024: 62+ WPM with LLM
  • Metzger 2023: 78 WPM speech + avatar

Limitations

  • Requires Utah Array or ECoG
  • Craniotomy
  • Only used for severe disability
  • Neuralink's 1024 channels → theoretical ceiling of 150+ WPM
  • Higher speeds expected to be reported clinically in 2026–2027
  • Approaching "natural conversation"

3. Non-invasive Status

EEG Brain-to-Text

  • MindBig Data P300 typing: ~5 WPM
  • DeWave 2023: category-level, not word-level
  • EEGPT 2024: general pretraining, weak at fine detail

MEG Brain-to-Text

  • Meta Défossez 2023: recognizes 10 words from MEG signals
  • Progress but far from practical
  • MEG equipment is not portable ($2M+)

fMRI Brain-to-Semantics

  • Tang 2023: semantic level, not word level
  • Akin to "meaning translation"
  • Not "typing"

Non-invasive Performance Bottlenecks

  • Poor signal
  • High noise
  • Consumer grade may reach 2020-era invasive levels in 10 years

4. Invasive vs Non-invasive Comparison

Dimension Invasive (Willett/Neuralink) Non-invasive (Meta/consumer)
Speed 60+ WPM < 10 WPM
Accuracy WER < 10% WER 30–50%
Latency < 500 ms Seconds
Invasiveness High (craniotomy) Zero
Channel count 96–1024 4–128
Users Severe disability Anyone
Usable in 2026 Yes (research) Partially

5. Middle Ground: Minimally Invasive

Stentrode

  • 16 intravascular channels
  • ~10–15 WPM (estimated)
  • Already useful for ALS users
  • See Synchron_Stentrode

Precision Layer 7

  • 1024 surface channels
  • Theoretically close to invasive grade
  • Lower surgical risk
  • May be the intermediate stop on the consumer path

6. LLM Acceleration

Rescoring

See LLM post-processing fusion.

  • GPT-4 semantically corrects BCI output
  • Sometimes boosts effective fluency from 7 WPM to 30 WPM equivalent

Autocomplete

  • Type "I want" → LLM predicts "a coffee"
  • User confirms rather than types
  • Greatly accelerates real communication

Dialog Management

  • LLM manages the whole conversation
  • BCI only needs low-bandwidth confirm/select
  • Consumer-grade BCI becomes viable

7. Consumer-grade Feasibility Analysis

Status Quo (2026)

  • Non-invasive EEG: not accurate enough, not fast enough
  • Stentrode: too invasive
  • Apple AirPods EEG: too early

Problems

  • Consumers won't wear an EEG helmet
  • Won't undergo surgery to type
  • Existing touchscreens and voice are already good enough

Possible Breakthroughs

  • Dry-electrode headgear (integrated with AR glasses)
  • Wrist EMG (Meta CTRL-Labs style)
  • Emotion/intent level rather than word level
  • "Assistive input" rather than "keyboard replacement"

8. Mind Typing in AR / VR

Vision Pro (2024)

  • Eye tracking + gestures + voice
  • No EEG
  • Gaze ≈ approximate mind typing

Future

  • Periocular EEG
  • Gesture EMG
  • Multimodality = each modality weak, combined strong

Meta Orion (2024)

  • Wrist EMG band + AR glasses
  • Gesture-neutral (no large movements needed)
  • Replacement for keyboard input on AR

9. Who Breaks Through to Consumer Grade First?

Candidates

  1. Apple: AirPods EEG + Vision Pro integration
  2. Meta: CTRL-Labs EMG + Orion
  3. Snap: NextMind integrated into Spectacles
  4. Google: no consumer-BCI initiative publicly disclosed as of 2026-04
  5. Samsung: Galaxy ecosystem

Projections

  • 2027–2028: AR glasses + multimodal input, approaching practical assistive input
  • 2030: mind typing as one of AR's primary inputs
  • 2035+: on par with the keyboard

10. Limits: Why "Full Keyboard Replacement" Is Hard

1. Bandwidth Limits

  • Cortical surface signal is not dense enough
  • Whole-word speed is capped

2. User Fatigue

  • Sustained "thinking" is more tiring than fingers
  • Neurofeedback demands focus

3. Precision

  • Natural thought is non-linear
  • Typing demands symbolic thinking

4. Privacy

  • The boundary of "think-to-type" is fuzzy
  • May leak things not intended for transmission

11. Disabled Users vs Healthy Users

Disabled Users

  • Genuine must-have: no other option
  • Tolerate low speeds
  • Early beneficiaries

Healthy Users

  • Superior alternatives (keyboard, voice)
  • Only have an edge in AR scenarios
  • Slower adoption

Strategic divergence: invasive targets medical; non-invasive targets consumer AR.

12. Logic Chain

  1. Mind typing reached 62 WPM invasively and < 10 WPM non-invasively in 2024–2026.
  2. Willett 2023 and Metzger 2023 are invasive milestones; Meta Défossez is a non-invasive probe.
  3. LLM acceleration brings low-bandwidth BCIs close to practical (autocomplete + rescoring).
  4. Consumer-grade isn't practical today, but AR glasses + EMG + eye tracking form a multimodal path.
  5. Apple, Meta, Snap are competing in consumer BCI.
  6. 2027–2030 consumer AR BCI expected to mature.
  7. Disabled vs healthy: invasive medical vs non-invasive consumer is the differentiation.

References

  • Willett et al. (2023). A high-performance speech neuroprosthesis. Nature.
  • Métzger et al. (2023). A high-performance neuroprosthesis for speech decoding and avatar control. Nature.
  • Willett et al. (2021). High-performance brain-to-text communication via handwriting. Nature.
  • Défossez et al. (2023). Decoding speech perception from non-invasive brain recordings. Nat Machine Intelligence.
  • Card et al. (2024). An accurate and rapidly calibrating speech neuroprosthesis. NEJM.

评论 #