Meta AI and the University of Amsterdam unveil Pixel Transformers, a groundbreaking neural network architecture that treats individual pixels as tokens, outperforming traditional models in various computer vision tasks.
Pixel Transformers by Meta AI: Revolutionizing Image Processing
In Short
According to recent research from Meta AI and the University of Amsterdam, transformers are a common neural network architecture that can work directly on individual pixels in an image without depending on the locality inductive bias found in most contemporary computer vision models.
Vanilla Transformers are capable of producing extremely performant outcomes by treating every single pixel as a token in their operations. This design differs significantly from the widely used one in Vision Transformer, which treats each 16×16 patch as a token and preserves the inductive bias from ConvNets towards local neighbourhoods.
The efficiency of using pixels as tokens in three well-researched computer vision tasks: creating images using diffusion models, supervised learning for object categorization, and self-supervised learning through masked autoencoding.
Even if it is less computationally viable to manipulate individual pixels directly, researchers believe that the community should be aware of this surprising discovery to develop the next generation of computer vision neural networks.
The introduction of Pixel Transformers (PiTs) by researchers eliminated any presumptions regarding the 2D grid layout of images by treating each pixel as a separate token. Remarkably, PiTs performed remarkably well in a variety of activities.
Also Read: Apple Unveils ‘Apple Intelligence’ AI, Limited Developer Access This Summer
PiTs followed the Diffusion Transformers (DiTs) architecture and fared better than their locality-biased equivalents in quality metrics like Fréchet Inception Distance (FID) and Inception Score (IS) while operating on latent token spaces from VQGAN.
As per the research, the coverage and usefulness are still constrained, though. Because of the quadratic computation complexity, PiT is more of an investigative technique than an application-specific one.
However, we think this study has made it very evident—unfiltered—that pacification is just a helpful heuristic that compromises accuracy for performance and that locality is not essential.
Also Read: Oracle’s Initiative to Train 200,000 Indians in AI, Data Science, and Cloud
This post was last modified on June 17, 2024 11:40 pm
Google is launching The Android Show: I/O Edition, featuring Android ecosystem president Sameer Samat, to…
The top 11 generative AI companies in the world are listed below. These companies have…
Google has integrated Veo 2 video generation into the Gemini app for Advanced subscribers, enabling…
Perplexity's iOS app now makes its conversational AI voice assistant compatible with Apple devices, enabling…
Bhavish Aggarwal is in talks to raise $300 million for his AI company, Krutrim AI…
The Beijing Humanoid Robot Innovation Center won the Yizhuang Half-Marathon with the "Tiangong Ultra," a…