As the effort to build more and more complex models and machines continues, it is becoming increasingly expensive to find not only the resources to build such machines, but to collect, annotate and label the training data. Because as you know, without having properly labeled data, the model might as well not exist!
Often to train these complex models, we had to manually tag and annotate the images to be used by the algorithm. A group of researchers claim to have found a solution to improvise this time taking process. Google’s DeepMind research team has created an algorithm that can create 3D objects based on a few 2D snapshots. In other words, the AI algorithm is able to use the 2D images to understand or “imagine” how the object looks like from various angles (which are not seen in the image).
The team of researchers have published a research paper explaining how the algorithm works. The system, called Generative Query Network or GQN, can think and imagine like humans. It can render 3D objects without having to be trained on what the various angles of the object look like.
As mentioned by the team, the AI system has two parts:
The GQN has the ability to learn about the shape, size and color of the object independently and can then combine all these features to form an accurate 3D model. Furthermore, the researchers were able to use this algorithm to develop new scenes without having to explicitly train the system as to what object should go where.
The tests were conducted in a virtual room and the results have been shared by the team in the above mentioned research paper. The algorithm has been tested only on objects so far and is not yet developed enough to work with human faces.
You can read about this algorithm in more detail on DeepMind’s blog post.
Another computer vision algorithm! Annotating and labeling data is a major time staking and often expensive exercise so this is another welcome addition. While still in its nascent stages (DeepMind admit as much in the blog), the potential applications are HUGE. GQN is not limited to tagging and annotating images, it can further be used by autonomous robots to have a better understanding of their surroundings.
What else can you think of where we can apply this algorithm? Let me know your thoughts on this in the comments section below!