Researchers from Google’s DeepMind subsidiary have developed deep neural networks that “have a remarkable capacity to understand a scene, represent it in a compact format, and then ‘imagine’ what the same scene would look like from a perspective the network hasn’t seen before,” writes Timothy B. Lee via Ars Technica. From the report: A DeepMind team led by Ali Eslami and Danilo Rezende has developed software based on deep neural networks with these same capabilities — at least for simplified geometric scenes. Given a handful of “snapshots” of a virtual scene, the software — known as a generative query network (GQN) — uses a neural network to build a compact mathematical representation of that scene. It then uses that representation to render images of the room from new perspectives — perspectives the network hasn’t seen before.

Under the hood, the GQN is really two different deep neural networks connected together. On the left, the representation network takes in a collection of images representing a scene (together with data about the camera location for each image) and condenses these images down to a compact mathematical representation (essentially a vector of numbers) of the scene as a whole. Then it’s the job of the generation network to reverse this process: starting with the vector representing the scene, accepting a camera location as input, and generating an image representing how the scene would look like from that angle. The team used the standard machine learning technique of stochastic gradient descent to iteratively improve the two networks. The software feeds some training images into the network, generates an output image, and then observes how much this image diverged from the expected result. […] If the output doesn’t match the desired image, then the software back-propagates the errors, updating the numerical weights on the thousands of neurons to improve the network’s performance.

Share on Google+

of this story at Slashdot.

…read more

Source:: Slashdot