Meta AI and the University of Amsterdam unveil Pixel Transformers, a groundbreaking neural network architecture that treats individual pixels as tokens, outperforming traditional models in various computer vision tasks.
Pixel Transformers by Meta AI: Revolutionizing Image Processing
In Short
According to recent research from Meta AI and the University of Amsterdam, transformers are a common neural network architecture that can work directly on individual pixels in an image without depending on the locality inductive bias found in most contemporary computer vision models.
Vanilla Transformers are capable of producing extremely performant outcomes by treating every single pixel as a token in their operations. This design differs significantly from the widely used one in Vision Transformer, which treats each 16×16 patch as a token and preserves the inductive bias from ConvNets towards local neighbourhoods.
The efficiency of using pixels as tokens in three well-researched computer vision tasks: creating images using diffusion models, supervised learning for object categorization, and self-supervised learning through masked autoencoding.
Even if it is less computationally viable to manipulate individual pixels directly, researchers believe that the community should be aware of this surprising discovery to develop the next generation of computer vision neural networks.
The introduction of Pixel Transformers (PiTs) by researchers eliminated any presumptions regarding the 2D grid layout of images by treating each pixel as a separate token. Remarkably, PiTs performed remarkably well in a variety of activities.
Also Read: Apple Unveils ‘Apple Intelligence’ AI, Limited Developer Access This Summer
PiTs followed the Diffusion Transformers (DiTs) architecture and fared better than their locality-biased equivalents in quality metrics like Fréchet Inception Distance (FID) and Inception Score (IS) while operating on latent token spaces from VQGAN.
As per the research, the coverage and usefulness are still constrained, though. Because of the quadratic computation complexity, PiT is more of an investigative technique than an application-specific one.
However, we think this study has made it very evident—unfiltered—that pacification is just a helpful heuristic that compromises accuracy for performance and that locality is not essential.
Also Read: Oracle’s Initiative to Train 200,000 Indians in AI, Data Science, and Cloud
This post was last modified on June 17, 2024 11:40 pm
Rish Gupta is an Indian entrepreneur who serves as the chief executive officer (CEO) of…
Are you looking to advance your engineering career in the field of robotics? Check out…
Artificial intelligence is a topic that has recently made internet users all over the world…
Boost your learning journey with the power of AI communities. The article below highlights the…
Demystify the world of Artificial Intelligence with our comprehensive AI Glossary and Terminologies Cheat Sheet.…
Scott Wu is the co-founder and Chief Executive Officer of Cognition Labs, an artificial intelligence…