The Data Set

A multi-modal dataset of human reasoning — built from 7 years of educational video, enriched with transcripts, EEG signals, and facial analysis to create the richest thinking-data corpus ever assembled.

Karpathy AwardThe Origin StoryData Pipeline
Milestone

Recognized by Andrej Karpathy

The Omega Quest dataset caught the attention of Andrej Karpathy — former Director of AI at Tesla, founding member of OpenAI, and one of the most influential voices in machine learning today. His recognition validated our thesis: that high-quality, multi-modal human reasoning data is the missing ingredient in the next generation of AI systems.

This moment marked a turning point for the project. Karpathy's endorsement brought visibility and confirmed that the direction we'd been pursuing — capturing real human thinking processes across multiple modalities — was exactly what the AI research community needed. The tweet sparked a wave of interest from researchers, builders, and data contributors worldwide.

Origin Story

7 Years of Thinking Out Loud

The Uncertain Systems YouTube channel started over 7 years ago as a personal project — recording the process of working through quantum mechanics, physics, and mathematics problems on camera. Every video captures real, unscripted reasoning: hypotheses formed, mistakes made, intuitions tested. Hundreds of hours of genuine human thinking, preserved on video. This archive is the seed of the Omega Quest dataset.

Exploring intuition around the expectation value formula

Entanglement Simulator — Hypergraph to Wavefunction

Digging into expectation values... deeply

Exploring the position and momentum conversion — Basic flaw identified!

Visit the Uncertain Systems Channel
Data Pipeline

How Video Becomes Training Data

Each video in the archive goes through a multi-stage enrichment pipeline. Raw footage is transformed into a richly annotated, multi-modal data record that captures not just what was said, but how the person was thinking — their cognitive load, emotional state, attention patterns, and reasoning flow.

Example Source Video

“Exploring intuition around the expectation value formula” — 40 minutes of a human working through quantum mechanics, thinking out loud.

Raw Video
Transcription
EEG Mapping
Facial Analysis
Merged Record

Enrichment Layers — Concrete Example

Below is a real sample of what a single 30-second segment (timestamp 12:45–13:15) looks like after processing through all four data layers.

Video Metadata

video_idus_qm_ev_042
titleExploring intuition around the expectation value formula
duration40:27
segment12:45 → 13:15
domainquantum_mechanics
topic_tags["expectation_value", "wavefunction", "probability"]
difficultyadvanced
speakerdaniel_colomer
resolution1920×1080 @ 30fps

Transcript Data

whisper_modellarge-v3
languageen (confidence: 0.98)
text"So now I'm looking at this integral and thinking... if ψ is normalized, the expectation value should just be this inner product. But wait — that assumes the operator is Hermitian, right? Let me check that assumption before I go further because if it's not, then this whole derivation falls apart..."
word_count52
word_timestamps[{"word":"So","start":765.0,"end":765.2}, {"word":"now","start":765.2,"end":765.4}, ...]
reasoning_markers["thinking...", "But wait", "Let me check", "assumes"]
hesitation_count3
self_correctiontrue

EEG / Neural Data

deviceMuse 2 (4-channel)
channels["TP9", "AF7", "AF8", "TP10"]
sample_rate256 Hz
alpha_power (μV²)8.42 → 3.17 (↓ 62%)
beta_power (μV²)2.89 → 7.61 (↑ 163%)
theta_power (μV²)5.14 → 6.82 (↑ 33%)
cognitive_load0.84 / 1.0 (HIGH)
interpretationSharp alpha suppression + beta spike at 12:52 — onset of deep analytical reasoning. Theta increase suggests working memory engagement.
artifact_rejection2 blinks removed (ICA)

Facial Analysis Data

modelMediaPipe Face Mesh v2
landmarks478 points @ 30fps
gaze_directionupper-left → center → paper (12:48–12:55)
brow_furrow0.72 / 1.0 (onset at 12:51)
micro_expressionsurprise (12:52, 340ms) — "But wait"
lip_compression0.61 / 1.0 (12:53–13:02)
head_tilt12° left (sustained, 12:50–13:10)
emotional_arcfocused → surprised → concentrated → resolved
engagement_score0.91 / 1.0

Merged Training Record

All four layers merge into a single training record. This is what gets fed to the model — a holistic snapshot of a human reasoning moment, annotated across every measurable dimension.

{
  "segment_id": "us_qm_ev_042_seg_026",
  "timestamp": {
    "start": 765,
    "end": 795
  },
  "video": {
    "domain": "quantum_mechanics",
    "topic": "expectation_value",
    "difficulty": "advanced"
  },
  "transcript": {
    "text": "So now I'm looking at this integral and thinking... if ψ is normalized, the expectation value should just be this inner product. But wait — that assumes the operator is Hermitian, right? Let me check...",
    "reasoning_markers": [
      "thinking...",
      "But wait",
      "Let me check",
      "assumes"
    ],
    "self_correction": true
  },
  "eeg": {
    "cognitive_load": 0.84,
    "alpha_delta": -0.62,
    "beta_delta": 1.63,
    "theta_delta": 0.33,
    "event": "analytical_onset @ 12:52"
  },
  "facial": {
    "engagement": 0.91,
    "emotional_arc": [
      "focused",
      "surprised",
      "concentrated",
      "resolved"
    ],
    "micro_expression": {
      "type": "surprise",
      "at": "12:52",
      "duration_ms": 340
    },
    "brow_furrow": 0.72
  },
  "labels": {
    "reasoning_type": "verification",
    "cognitive_event": "assumption_check",
    "quality_score": 0.94
  }
}

Beyond Text

Current AI training data is mostly text. Our dataset adds neural signals, facial expressions, and audio prosody — the full picture of human cognition.

Real Reasoning

Not scripted, not edited. Every data point comes from genuine problem-solving moments — mistakes, corrections, and breakthroughs included.

Temporal Alignment

Every modality is time-synced to the millisecond. The model can learn that a beta-spike and a brow-furrow coincide with a verbal self-correction.

Help Build the Dataset

Contribute your thinking data and help build the richest multi-modal reasoning corpus ever assembled.

Learn How to Contribute