The Topological Transformer: Tauformer (domain-memory and faster attention)
Domain memory injected directly inside self-attention via a persistent Graph Laplacian (distilled knowledge graphs with arrowspace).
- Replaces the dot-product attention kernel with a topology-aware scalar signal (taumode / λ-distance), so attention is driven by distances in a domain manifold rather than raw geometry.
- Targets scaling pain points: ~50% KV-cache savings (values + λk instead of K+V) and ~20% faster time-per-token vs a nanoGPT baseline in the reported benchmarks.