← Back to Archives
Archive: 2026-01-31_4
Pattern Injection: UM → RNN
- Focus
- Injecting UM patterns into RNN weights via SVD. Why W_hh fails.
- Key Result
- ~1 bit/char head start (5.46 → 4.47 bpc). Q unbounded, f32 isn't.
- Builds On
- 2026-01-31_3 (pattern-injection.pdf)
Figures
fig-02: RNN Memory Depth
How far back does the RNN remember? Variance method. Predicted ~12 chars, measured ~flat.
fig-01: SVD Component Interpretation
What singular values mean. ASCII vs UTF-8, letters vs digits, bracket structure, phonotactics.
Reachability Map
Archive structure visualization. Forward and back links.
Papers
RNN Memory Depth: Experiment
Predicted ~12 chars, observed ~flat. Two methods, analysis, future work.
Embeddings and the Atomic Time Step
t_{n+1} = p × t_n becomes h_{n+1} = W × h_n. Dimension d = bandwidth.
Time, Frequency, Energy, and Bits
bits ∝ h/time. Depth limit as energy budget. Fourier ↔ Bayes/Thermo duality.
Unification: Bayes, Thermo, Quotient, Factor Maps
The quotient IS the luck. Commutative diagrams. Depth limit.
Pattern Injection: UM → RNN
Theory, SVD factorization, experiments, AC analogy. Light theme.
Dashboard View
Same content, dark theme. Tables, code blocks.
Empirical Results
Pattern Injection vs Random Init
Initialization Initial After 10 epochs
---------------------------------------------------
Random 5.46 bpc 4.81 bpc
Pattern Injection 4.47 bpc 4.53 bpc
Gap 0.99 bpc 0.28 bpc
SVD Factorization (H=64):
Full bigram: 3.84 bpc
Rank-64 approx: 3.92 bpc
Loss from rank: 0.08 bpc
Files
svd_interpret.py
SVD component analysis. Generates the interpretation above.
svd_viz.py
Full SVD visualization with log support tables.
memory_depth.py
Method 1: perturbation approach (shows why it fails).
memory_depth2.py
Method 2: conditional variance approach.
inject_train.py
Training comparison: random vs injected initialization.
export_weights.py
Export SVD weights to C header.
um_to_rnn.py
Complete UM → RNN pipeline.
Source (.tex)
memory-depth.tex
embeddings-time.tex
time-energy.tex
unification.tex
Navigation
→ Next: 20260131_5
Retrospective and predictions. Looking back, testable claims.
← Previous: 20260131_3
Memory traces, factor maps, pattern injection theory.