GRL Tutorial Chapters¶
Format: Tutorial paper chapters
Audience: ML practitioners with basic RL knowledge
Style: Narrative, educational, self-contained
Overview¶
These chapters form the core of the GRL tutorial paper. Each chapter builds on previous ones while remaining accessible for selective reading.
Chapter Index¶
Part I: Foundations¶
| # | Title | Key Concepts | Status |
|---|---|---|---|
| 00 | Overview | What is GRL, motivation | ✅ Complete |
| 01 | Core Concepts | Augmented space, parametric actions | ✅ Complete |
| 02 | RKHS Foundations | Kernels, inner products, function spaces | ✅ Complete |
| 03 | Energy and Fitness | Sign conventions, EBM connection | ✅ Complete |
| 03a | Least Action Principle (supplement) | Path integrals, Boltzmann policy, action discovery | ✅ Complete |
Part II: Reinforcement Field¶
| # | Title | Key Concepts | Status |
|---|---|---|---|
| 04 | Reinforcement Field | Functional field, RKHS gradient | ✅ Complete |
| 04a | Riesz Representer (supplement) | Gradients in function space, examples | ✅ Complete |
| 05 | Particle Memory | Particles as basis, memory as belief | ✅ Complete |
Part III: Algorithms¶
| # | Title | Key Concepts | Status |
|---|---|---|---|
| 06 | MemoryUpdate | Belief transition, Algorithm 1, particle evolution | ✅ Complete |
| 06a | Advanced Memory Dynamics (supplement) | Top-k neighbors, surprise-gating, practical improvements | ✅ Complete |
| 07 | RF-SARSA | Functional TD learning, two-layer architecture, Algorithm 2 | ✅ Complete |
| 07a | Continuous Policy Inference (supplement) | Beyond discrete actions, Langevin sampling, actor-critic in RKHS | ✅ Complete |
Part IV: Interpretation¶
| # | Title | Key Concepts | Status |
|---|---|---|---|
| 08 | Soft State Transitions | Emergent uncertainty | ⏳ Planned |
| 09 | POMDP Interpretation | Belief-based view | ⏳ Planned |
| 10 | Complete System | Putting it together | ⏳ Planned |
Reading Recommendations¶
New to GRL¶
Start with Chapter 00 (Overview), then proceed sequentially through the completed chapters.
Familiar with RL Theory¶
Skim Chapter 00, focus on Chapters 02-04 for mathematical foundations.
Want to Implement¶
Read Chapter 00, skim 01-04, then wait for Chapters 05-07 on algorithms.
Quick Understanding¶
Read Chapters 00, 01, and 04 for the essential concepts.
Chapter Progression¶
00-Overview
↓
01-Core Concepts
↓
02-RKHS Foundations
↓
03-Energy and Fitness
↓
03a-Least Action Principle (supplement) ← New!
↓
04-Reinforcement Field
↓
04a-Riesz Representer (supplement)
↓
05-Particle Memory
↓
06-MemoryUpdate
↓
06a-Advanced Memory Dynamics (supplement)
↓
07-RF-SARSA
↓
07a-Continuous Policy Inference (supplement) ← We are here
↓
08-Soft State Transitions (planned)
↓
...
Chapter Template¶
Each chapter follows this structure:
- Header: Purpose, prerequisites, key concepts
- Introduction: Why this topic matters
- Main Content: Narrative explanation with examples
- Key Takeaways: Summary points
- Next Steps: Connection to following chapters
Notation Conventions¶
| Symbol | Meaning |
|---|---|
| \(s\) | Environment state |
| \(\theta\) | Action parameters |
| \(z = (s, \theta)\) | Augmented state-action point |
| \(k(\cdot, \cdot)\) | Kernel function |
| \(\mathcal{H}_k\) | RKHS induced by kernel \(k\) |
| \(Q^+(z)\) | Field value (fitness) at \(z\) |
| \(E(z)\) | Energy at \(z\), equals \(-Q^+(z)\) |
| \(\Omega\) | Particle memory |
| \(w_i\) | Weight of particle \(i\) |
Key Equations¶
Reinforcement Field¶
RKHS Inner Product¶
Energy-Fitness Relationship¶
Functional Gradient¶
Boltzmann Policy¶
Beyond Part I¶
Quantum-Inspired Extensions¶
Advanced Topics → (9 chapters complete)
Mathematical connections to quantum mechanics, amplitude-based learning, and novel memory dynamics.
Research Roadmap¶
Comprehensive plan for GRL v0, extensions, and future papers (A, B, C).
Last Updated: January 14, 2026