Exploring possible vector systems for faster training of neural networks with preconfigured latent spaces
By: Nikita Gabdullin
Potential Business Impact:
Makes computers learn faster with fewer details.
The overall neural network (NN) performance is closely related to the properties of its embedding distribution in latent space (LS). It has recently been shown that predefined vector systems, specifically An root system vectors, can be used as targets for latent space configurations (LSC) to ensure the desired LS structure. One of the main LSC advantage is the possibility of training classifier NNs without classification layers, which facilitates training NNs on datasets with extremely large numbers of classes. This paper provides a more general overview of possible vector systems for NN training along with their properties and methods for vector system construction. These systems are used to configure LS of encoders and visual transformers to significantly speed up ImageNet-1K and 50k-600k classes LSC training. It is also shown that using the minimum number of LS dimensions for a specific number of classes results in faster convergence. The latter has potential advantages for reducing the size of vector databases used to store NN embeddings.
Similar Papers
Using predefined vector systems as latent space configuration for neural network supervised training on data with arbitrarily large number of classes
Machine Learning (CS)
Teaches computers to learn from many more things.
Series of quasi-uniform scatterings with fast search, root systems and neural network classifications
Algebraic Geometry
Teaches computers to learn new things faster.
Native Logical and Hierarchical Representations with Subspace Embeddings
Machine Learning (CS)
Computers understand words and their meanings better.