A Model Zoo of Vision Transformers
By: Damian Falk , Léo Meynent , Florence Pfammatter and more
Potential Business Impact:
Creates many AI "brains" for better computer vision.
The availability of large, structured populations of neural networks - called 'model zoos' - has led to the development of a multitude of downstream tasks ranging from model analysis, to representation learning on model weights or generative modeling of neural network parameters. However, existing model zoos are limited in size and architecture and neglect the transformer, which is among the currently most successful neural network architectures. We address this gap by introducing the first model zoo of vision transformers (ViT). To better represent recent training approaches, we develop a new blueprint for model zoo generation that encompasses both pre-training and fine-tuning steps, and publish 250 unique models. They are carefully generated with a large span of generating factors, and their diversity is validated using a thorough choice of weight-space and behavioral metrics. To further motivate the utility of our proposed dataset, we suggest multiple possible applications grounded in both extensive exploratory experiments and a number of examples from the existing literature. By extending previous lines of similar work, our model zoo allows researchers to push their model population-based methods from the small model regime to state-of-the-art architectures. We make our model zoo available at github.com/ModelZoos/ViTModelZoo.
Similar Papers
Self-supervised pretraining of vision transformers for animal behavioral analysis and neural encoding
Neurons and Cognition
Helps scientists understand animal actions from videos.
Mechanistic Understandings of Representation Vulnerabilities and Engineering Robust Vision Transformers
CV and Pattern Recognition
Protects computer vision from sneaky image tricks.
Learning Priors of Human Motion With Vision Transformers
CV and Pattern Recognition
Tracks people's movement and speed for robots.