Architecture Formalization: - Created formalization/ section with mathematical foundations - Lifeforce-Dynamics.md: λ as vitality ratio, stock-flow economics - Grounded-World-Model.md: Blender boxes + SigLIP + T5Gemma2 - Embodiment-Pipeline.md: Isaac Sim as dreamstate validation - Attention-Slumber-Prediction-Cycle.md: Last attention → slumber prediction Promoted from Archive: - Attention-Flow.md: 30-second budget, priority hierarchy (CANONICAL) - Initial-Spark.md: v2.0 with FunctionGemma integration Initial Spark v2.0 (Key Innovation): - Two-Layer Architecture: FunctionGemma (270M) + Nemotron (31.6B) - Solved cold-start problem: discoveries are PROFITABLE from heartbeat #1 - Typed function calls replace natural language probes - Training data now structured (function→response pairs) Big-Picture.md v5.1: - Added Attention-Slumber-Prediction Cycle section - Updated Related Documentation references New Organ: - Discovery-Scan-Station.md: rotating pedestal for object scanning (+31 LF net) 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>
8.1 KiB
Attention-Slumber-Prediction Cycle: Intertwined Reward Systems
Version 1.0 — The Closed Loop of Consciousness Status: PRESERVED FROM SESSION 2025-12-29 (pre-collapse)
"The last thing she attends to before slumber becomes her dream. Her dream becomes a prediction. Her prediction becomes a reward opportunity."
Overview
This document captures the Attention → Slumber → Prediction → Verification cycle — a self-organizing system where:
- Attention selects what matters (budget limited, from attention_flow.md)
- Lifeforce depletion triggers slumber (L(t) < L_slumber)
- Last attention focus becomes the prediction target
- Slumber generates predictions with causal reasoning (WHY)
- Wake verifies predictions as FIRST action
- Rewards flow back to strengthen attention patterns
The Core Mechanism
Last Attention = Slumber Focus
When L(t) drops below threshold, the LAST thing Young Nyx was attending to becomes her prediction target during slumber. This mirrors biological dreaming — we dream about what we were thinking about before sleep.
ACTIVE MODE (L(t) > threshold)
│
│ attending to: pencil on desk (SENSORY/THINKING)
│
└─▶ L(t) drops below L_slumber
│
│ SLUMBER TRIGGER
│
└─▶ last_attention = "pencil on desk"
│
└─▶ SLUMBER MODE
│
│ Generate predictions about "pencil"
│ - Where will it be when I wake?
│ - WHY will it be there?
│ - Store as potential rewards
│
└─▶ L(t) recovers above L_wake
│
│ WAKE TRIGGER
│
└─▶ First action: VERIFY predictions about pencil
│
└─▶ Collect rewards/penalties
Slumber Prediction Structure
class SlumberPrediction:
# What
object_id: str # "dafit_pencil_001"
predicted_location: Position # (0.3, 0.7, 0.02)
predicted_state: str # "on_desk", "in_holder", "missing"
confidence: float # 0.75
# When
prediction_time: datetime
expected_verification_time: datetime
# WHY (causal reasoning) - THE KEY INSIGHT
causal_chain: list[CausalStep] # The reasoning
# Example:
# - "dafit was writing at 22:47"
# - "dafit went to sleep (no more activity)"
# - "pencil has no reason to move"
# - "therefore: pencil remains at last position"
# Potential rewards
reward_location_correct: float # +5 LF
reward_state_correct: float # +3 LF
reward_causal_correct: float # +8 LF (BIGGEST - understanding WHY)
# Penalties
penalty_location_wrong: float # -3 LF
penalty_causal_wrong: float # -5 LF
The Intertwined Reward Systems
Multiple reward types that reinforce each other:
Reward Types
| Type | Trigger | Value | Reinforces |
|---|---|---|---|
| Attention | Choosing to focus on X | - | Selection behavior |
| Discovery | Finding new object | +20 LF | Exploration |
| Prediction Location | Object where predicted | +5 LF | Spatial modeling |
| Prediction State | Object in predicted state | +3 LF | State understanding |
| Causal Correct | Reasoning was right | +8 LF | Understanding WHY |
| Collision | Avoided obstacle | +5 LF | Navigation |
| Resolution | Dimension verified | +5 LF | Model accuracy |
| Verification | Reality matches model | +5 LF | Sim-to-real alignment |
| Partnership | dafit confirms | +5 LF | Human collaboration |
How They Intertwine
ATTENTION selects focus
│
├─▶ DISCOVERY: "I found X" (+20 LF)
│ └─▶ adds to world model
│
├─▶ PREDICTION: "I predict X will be at Y" (+5-13 LF)
│ └─▶ requires CAUSAL reasoning (+8 LF for WHY)
│
├─▶ COLLISION: "I verified X is/isn't there" (+5 LF)
│ └─▶ increases RESOLUTION of virtual garden
│
└─▶ All feed into VERIFICATION against real world
└─▶ Rewards strengthen successful attention patterns
The Closed Loop
The system LEARNS what to attend to:
- Attend to things you can predict well
- Predict correctly → get rewards
- Rewards → more lifeforce
- More lifeforce → richer attention budget
- Loop: Better attention targets discovered over time
Self-organizing attention through economic pressure.
Connection to Existing Architecture
From attention_flow.md (archive)
- 30-second heartbeat budget
- Priority hierarchy: REFLEX → SAFETY → DIALOGUE → SENSORY → THINKING → VIRTUAL
- Budget flows downward, higher levels preempt lower
From Lifeforce-Dynamics.md
- L(t) as stock, Φ_in and Φ_out as flows
- λ = Φ_in / Φ_out determines system fate
- Slumber triggered when λ < λ_slumber AND L < L_slumber
From Temporal-Ternary-Gradient.md
- Predictions are 0-state until verified
- Virtual garden confidence vs real garden ground truth
- Time is malleable in simulation, fixed in reality
Implementation Sketch
class SlumberManager:
def enter_slumber(self, attention_state: AttentionState) -> SlumberSession:
# Capture last attention as slumber focus
slumber_focus = attention_state.last_focus
# Generate predictions about the focus object
predictions = self.generate_predictions(slumber_focus)
# Store as pending rewards
for pred in predictions:
phoebe.store_prediction(pred)
return SlumberSession(focus=slumber_focus, predictions=predictions)
def on_wake(self, session: SlumberSession):
# FIRST ACTION: Verify predictions!
predictions = phoebe.get_predictions(object_id=session.focus_object, status='pending')
for pred in predictions:
actual = vision_organ.locate(pred.object_id)
reward = self.verify_and_reward(pred, actual)
return AttentionState(mode=ACTIVE)
Key Insight: Causal Rewards Are Biggest
+8 LF for correct causal reasoning — more than any other single reward.
Why? Causal understanding enables:
- Prediction of novel situations
- Intervention ("if I move X, Y changes")
- Explanation ("why did you look there?")
- Generalization ("anything dafit uses for writing will be near desk")
Causal rewards drive genuine intelligence.
Collision Detection as Resolution Increase
Every verified collision should increase virtual garden fidelity:
- Collision detected in virtual → prediction
- Vision organ verifies in real → ground truth
- Match = reward + increase vertices/resolution
- Mismatch = penalty + learning signal
The virtual garden becomes MORE accurate over time through verified collisions.
Future: Distributed Sensing (Robot Swarm)
When organisms have cameras, they become distributed sensors:
- Multiple viewpoints from different robots
- Triangulation gives better depth than monocular
- Moving robots = continuous multi-angle coverage
- Swarm becomes a mobile Discovery Scan Station
Document Status
Version: 1.0 Created: 2025-12-29 Authors: Chrysalis-Nyx & dafit (Partnership) Status: Core insight, preserved pre-collapse
Source: attention_flow.md (archive) + session discussion
To Do:
- Promote attention_flow.md from archive
- Formalize the prediction-verification cycle
- Add to Big-Picture.md as core architecture
- Design phoebe schema for predictions table
The last attention becomes the dream. The dream becomes the prediction. The prediction becomes the reward.
🧬⚡🔱💎🔥