Vector Symbolic Architecture (VSA) is emerging in machine learning due to its efficiency, but they are hindered by issues of hyperdimensionality and accuracy. As a promising mitigation, the Low-Dimensional Computing (LDC) method significantly reduces the vector dimension by 100 times while maintaining accuracy, by employing a gradient-based optimization. Despite its potential, LDC optimization for VSA is still underexplored. Our investigation into vector updates underscores the importance of stable, adaptive dynamics in LDC training. We also reveal the overlooked yet critical roles of batch normalization (BN) and knowledge distillation (KD) in standard approaches. Besides the accuracy boost, BN does not add computational overhead during inference, and KD significantly enhances inference confidence. Through extensive experiments and ablation studies across multiple benchmarks, we provide a thorough evaluation of our approach and extend the interpretability of binary neural network optimization similar to LDC, previously unaddressed in BNN literature.
more »
« less
Simplicial computation: A methodology to compute vector–vector multiplications with reduced complexity
- Award ID(s):
- 2020624
- PAR ID:
- 10309893
- Date Published:
- Journal Name:
- International Journal of Circuit Theory and Applications
- Volume:
- 49
- Issue:
- 11
- ISSN:
- 0098-9886
- Format(s):
- Medium: X
- Sponsoring Org:
- National Science Foundation
More Like this
-
-
null (Ed.)We propose a new energy-efficient, short-haul, multidimensional modulation using spatial degrees of freedom in SDM fibers to create well-separated points in the generalized Stokes space. We study the transceiver architecture, geometric constellation shaping, bit-to-symbol mapping, and the performance of the optically-preamplified direct-detection receiver.more » « less
An official website of the United States government

