Vision transformer

From WikiMD's Food, Medicine & Wellness Encyclopedia

Vision_Transformer

Vision Transformer (ViT)

The Vision Transformer (ViT) is a type of artificial neural network architecture that applies transformer models, originally designed for natural language processing (NLP), to computer vision tasks. The ViT model was introduced by researchers at Google Research and has demonstrated state-of-the-art performance on various image recognition benchmarks.

Architecture[edit | edit source]

The Vision Transformer architecture leverages the self-attention mechanism of transformers to process image data. Unlike traditional convolutional neural networks (CNNs), which use convolutional layers to extract features from images, ViT divides an image into fixed-size patches and treats each patch as a token, similar to words in NLP.

Image Patching[edit | edit source]

An input image is divided into a grid of non-overlapping patches. Each patch is then flattened into a vector and linearly embedded into a lower-dimensional space. These embedded patches are combined with positional encodings to retain spatial information.

Transformer Encoder[edit | edit source]

The sequence of embedded patches is fed into a standard transformer encoder, which consists of multiple layers of multi-head self-attention and feed-forward neural networks. The self-attention mechanism allows the model to weigh the importance of different patches relative to each other, enabling it to capture global context.

Classification Head[edit | edit source]

For image classification tasks, a special classification token is prepended to the sequence of embedded patches. The output corresponding to this token is passed through a feed-forward network to produce the final class predictions.

Advantages[edit | edit source]

  • **Scalability**: ViT models can be scaled up more easily than CNNs by increasing the number of transformer layers or the size of the patches.
  • **Performance**: ViT has achieved competitive performance on several image classification benchmarks, such as ImageNet.
  • **Transfer Learning**: Pre-trained ViT models can be fine-tuned on specific tasks, similar to pre-trained models in NLP.

Challenges[edit | edit source]

  • **Data Efficiency**: ViT models typically require large amounts of training data to achieve optimal performance.
  • **Computational Resources**: Training ViT models can be computationally intensive, requiring significant GPU resources.

Applications[edit | edit source]

Vision Transformers have been applied to various computer vision tasks, including:

See Also[edit | edit source]

References[edit | edit source]

External Links[edit | edit source]

Template:Neural-net-stub

Wiki.png

Navigation: Wellness - Encyclopedia - Health topics - Disease Index‏‎ - Drugs - World Directory - Gray's Anatomy - Keto diet - Recipes

Search WikiMD


Ad.Tired of being Overweight? Try W8MD's physician weight loss program.
Semaglutide (Ozempic / Wegovy and Tirzepatide (Mounjaro / Zepbound) available.
Advertise on WikiMD

WikiMD is not a substitute for professional medical advice. See full disclaimer.

Credits:Most images are courtesy of Wikimedia commons, and templates Wikipedia, licensed under CC BY SA or similar.

Contributors: Prab R. Tumpati, MD