Google researchers discover new algorithm that can render 3D scenes just by a few snapshots
- The new tech can render a 3D model with just five snapshots.
- This algorithm can identify an object’s shape, size and colour in a given scene.
- Once it is perfected, the need to label images for
AItraining will be eliminated.
We all can perceive our surroundings by just a look, and now computers will be able to accomplish the same. This new type of
So, how does the technology work?
First, the computer is introduced into an environment and is allowed to take some snaps from a few different perspectives. After this, the GQN takes over and pieces together an object's appearance and creates an abstract of the scene to learn the essentials. And based on what it ‘learns’, the GQN predicts what the object would look like from another angle, one that is not included in the snapshots.
What can the new algorithm do?
According to the published research, the AI system can fully render a 3D scene based on just five separate virtual images. The algorithm can identify shapes, sizes and colours of all the objects in the scene and then integrate them to create an accurate 3D model.
With the help of one rendered scene, researchers can create entirely new scenes without the need to explicitly lay out which objects needs to go where. And, all this is done without any human assistance, supervision or training.
Advantages over existing technology
Earlier, in order to enable any AI model to perceive an object, the model needed to be trained with a set of images both of the object and of scenes that do not have the object in it. And the more difficult part of the process was that all these images needed to be manually labelled by humans - and AI training means many, many, MANY images.
The new algorithm can help generate training images (with and without the object and label them) and the need of human intervention is reduced to zero. Think of all the money and time you can save.
This algorithm, when applied to advanced technologies, can help extend the
Google scientists also think that this can give rise to machines that can autonomously learn about their surroundings without any help.
Currently, the GQN is not refined enough to be thrown open to the gallery and is being used to train AI models and work on their precision
All these findings were published in the journal Science.
- Delhi budget: Free Covid vaccine scheme at state-run hospitals; special women mohalla clinics
- Uttarakhand CM Trivendra Singh Rawat resigns
- Laxmi Organic IPO is slated to hit capital markets next week — price band set at ₹129-130 a share
- Tamil Nadu orders spot fine of ₹200 for not wearing a mask amid rising cases of COVID-19
- India's oldest voter gets first dose of COVID-19 vaccine