Scientific References
Core Theory​
Hyperdimensional Computing (HDC)​
-
Kanerva, P. (2009). Hyperdimensional Computing: An Introduction to Computing in Distributed Representation with High-Dimensional Random Vectors. Cognitive Computation, 1(2), 139-159.
- Foundation of Vector Symbolic Architecture
- Binding, bundling, and permutation operations
-
Plate, T. A. (2003). Holographic Reduced Representation: Distributed Representation for Cognitive Structures. CSLI Publications.
- Circular convolution for binding
- Similarity-preserving operations
-
Rachkovskij, D. A., & Kussul, E. M. (2001). Binding and Normalization of Binary Sparse Distributed Representations by Context-Dependent Thinning. Neural Computation, 13(2), 411-452.
- Sparse distributed representations
- Context-dependent binding
Ternary Neural Networks​
-
Ma, S., Wang, H., Ma, L., et al. (2024). The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits. Microsoft Research. arXiv:2402.17764
- BitNet b1.58 architecture
- 1.58-bit quantization
- Key insight: Ternary weights (-1, 0, +1) achieve comparable accuracy to full-precision
-
Hubara, I., Courbariaux, M., Soudry, D., et al. (2017). Quantized Neural Networks: Training Neural Networks with Low Precision Weights and Activations. Journal of Machine Learning Research, 18(1), 6869-6898.
- Binary and ternary quantization
- Training methodologies
Mathematical Foundations​
-
Knuth, D. E. (1998). The Art of Computer Programming, Volume 2: Seminumerical Algorithms. Addison-Wesley.
- Balanced ternary arithmetic
- Radix economy analysis
- Key result: Base 3 is the most economical integer base
-
Hayes, B. (2001). Third Base. American Scientist, 89(6), 490-494.
- Popular introduction to balanced ternary
- Historical context (Setun computer)
Implementation References​
GGUF Format​
- ggerganov et al. (2023). GGUF: GGML Universal Format. GitHub: ggerganov/llama.cpp
- Model file format specification
- Quantization schemes (Q4, Q8, etc.)
SIMD Optimization​
-
Intel (2023). Intel Intrinsics Guide. Intel Developer Zone.
- AVX-512 operations
- Vectorized dot products
-
ARM (2023). ARM NEON Intrinsics Reference. ARM Developer.
- NEON SIMD operations
- SDOT instruction for int8 dot products
Compiler Technology​
-
Lattner, C., & Adve, V. (2004). LLVM: A Compilation Framework for Lifelong Program Analysis & Transformation. CGO 2004.
- LLVM IR
- JIT compilation
-
Zig Software Foundation (2024). Zig Language Reference. ziglang.org
- Comptime metaprogramming
- SIMD vectors
Trinity-Specific Papers​
VSA Operations​
Trinity VSA achieves 21x speedup over scalar baseline using
ARM NEON SIMD for 1024-dimensional ternary vectors.
Benchmark results (M3 chip):
- Scalar: 7.985 ms
- SIMD: 0.370 ms
- Speedup: 21.55x
Trinity Identity​
The mathematical foundation φ² + 1/φ² = 3:
φ = (1 + √5) / 2 ≈ 1.618033988749895
φ² = φ + 1 ≈ 2.618033988749895
1/φ² = 1/(φ + 1) ≈ 0.381966011250105
φ² + 1/φ² = 2.618... + 0.382... = 3.000
This connects to ternary computing:
- Base 3 optimal radix economy
- Three states: -1, 0, +1
- Three-way branching
Acknowledgments​
Trinity builds upon foundational work from:
- Pentti Kanerva (Stanford) - HDC/VSA theory
- Shuming Ma (Microsoft) - BitNet research
- Georgi Gerganov - llama.cpp and GGUF
- Andrew Kelley - Zig programming language
- Donald Knuth - Balanced ternary mathematics
Further Reading​
Online Resources​
Books​
- Kanerva, P. Sparse Distributed Memory. MIT Press, 1988.
- Plate, T. Holographic Reduced Representation. CSLI, 2003.
- Knuth, D. TAOCP Vol. 2. Addison-Wesley, 1998.
φ² + 1/φ² = 3 | Standing on the shoulders of giants