www.nanowerk.com 2015 04496.txt.txt

#The robot that learns everything from scratch (Nanowerk News) Two researchers at NTNU have made a robot that learns like a young child. At least, thats the idea. The machine starts with nothing it has to learn everything from scratch. Were still pretty far away from accurately modelling all aspects of a living childs brain, but the algorithms that handle sound and image processing are inspired by biology, says Professor yvind Brandtsegg at NTNU. The machine is called self..It analyses sound through a system based on the human ear, and learns to recognize images using a digital model of how nerve cells in the brain handle sensory impressions. It is designed to learn entirely from sensory input with no predefined knowledge database, so that its learning process will resemble that of a human child in early life. Weve given it almost no predefined knowledge on purpose, Brandtsegg says. Axel Tidemann, self. and yvind Brandtsegg. Photo: Ole Morten Melgrd, NTNU) Interdisciplinary project The we that Brandtsegg refers to is himself and postdoc Axel Tidemann because this is without a doubt an interdisciplinary project. The machine is so complex that cooperation between different research fields is an absolute necessity to get it to work. Brandtsegg is at the Department of music, while Tidemann is at the Department of Computer and Information science. But they have overlapping interests. We understand just enough of each others fields of study to see what is difficult, and why, Brandtsegg says. Naturally, his main interest is music. But he is accomplished also an programmer and uses this knowledge to make music. Conversely, Tidemann made a drumming robot for his doctoral project. The robot simulated the playing styles of living drummers. Knows nothing In the beginning, their robot knew nothing. It hears sounds from a person speaking, and can connect these to a simultaneous video feed of the speaker. The robot picks a sound that the person appears to be emphasizing, and responds by playing other sounds that it associates with this, while projecting a neural representation of its association between the sound and pictures. It doesnt show a video, but rather how its brain connects sounds and images. Learning The robot has already been on display in Trondheim and Arendal where visitors were able to affect its learning. It was in Trondheim for a month before Christmas, and in Arendal for two weeks in January. Interacting with a diverse audience allowed the researchers to see exactly how it learns. There was a lot of My name is and What is your name? from the audience, but some people sang, and others read poems. This resulted in a period where a lot of similar sounds and connected people got mixed up, a chaos of the machine making strange connections. But this changed the more it learned. The robot gradually absorbed more and more impressions of different people. Certain people, like guides, affected it more, because it saw them often. The robot also learned to filter input. If a word is said in a certain way five times, and then in a different way once, it learned to filter away the standout and concentrate on the most common way, which is presumably correct. This processing happens during the robots downtime. We say that the machine dreams at night, Brandtsegg says. After a while, the robot was able to connect words and pictures together in a more complex manner you could say that it associates sounds with images and connects them by itself. Development The robot is constantly under development, and Brandtsegg and Tidemann have lost a lot of sleep over it. The day before it was put on display in Trondheim, we worked through the night until eight in the morning. Then we went home ate breakfast, and went back to work at 11, Brandtsegg says. Between the two displays, they worked on improving the way the robot organizes its memories. Every little change we make takes a lot of time, at least if we want to make sure that we dont destroy any of the things it already has learned, Brandtsegg explains. The result is a robot that shows how it makes associations in a very pedagogical manner. It doesnt resemble any living organisms on purpose youre supposed to concentrate on its learning and the process behind it. The robot looks rough as Brandtsegg says. strong>Thinking on its own? self. is an art project, and raises questions that may be very relevant in the years to come. When is a robot thinking on its own? When is appropriate to call a machine living? Many say that intelligence can be determined by specific behaviour, Tidemann says. He names the Turing Test, in which a machine is considered to be thinking if it is able to convince a human that it is human as well, through text-based questions, at least thirty per cent of the time. Based on this definition, computers that play chess, like IBMS Deep Blue, can be defined as intelligent, because they are very good at chess. But this is symbolic reasoning that isnt necessarily transferrable to real-world scenarios. Specialized robots for things like precise machining and industrial work have been better at certain tasks than humans for decades. But these robots are far from being able to learn. Not to mention doing things like running up stairs or jumping rope. There is also no machine that is as good at analysing a football match or writing a novel as a human. Not in a vacuum Many artificial intelligence (AI) researchers, myself included, believe that true intelligence cant occur in a vacuum it is a consequence of adapting and living in a dynamic environment, Tidemann explains. You could see our intelligence as a byproduct of our adaptability. Because we have developed the ability to plan and remember, we get cognition as a sort of package deal, he says. Cognition is combined the ability to perceive your surroundings reason based on this perception, communicate with your surroundings and recall and act reasonably according to the information you have on hand. What is independent thinking? What is artificial life? These are the big questions, Tidemann says. But we believe that the right way to reach for the holy grail of AI is to implement biologically inspired models in a machine, let it operate in a physical environment and see if we can observe intelligent behaviour. Researchers use the phrase technological singularity to describe the point where human intellectual capacity is surpassed by machines. This is still a long time coming, however Brandtsegg and Tidemanns goal with self. is that it will be able to learn through interacting with humans as well as possible e


< Back - Next >


Overtext Web Module V3.0 Alpha
Copyright Semantic-Knowledge, 1994-2011