Gadgetory


All Cool Mind-blowing Gadgets You Love in One Place

Cracking the code of body language to build lifelike robots

2016-12-07
if you ever find yourself in Pittsburgh and want to catch a glimpse of the future Carnegie Mellon University is probably a good place to start to get inside the robotics lab you'll need to walk past tank he's a Robo ception a-- stand well actually tank is really important to this story but he isn't what I want to show you right now this is looking at this large geodesic dome covered in wires you can't help but feel like you're staring into a time machine and in a way you kind of are but rather than traveling through time this fin optics to do is capturing it using an array of cameras researchers here are able to capture and then replicate any action into a 3d model from a dance to a cello performance it can perfectly reconstruct these moments in time and while that's pretty impressive by itself it's just the starting point in a quest to decipher an age-old code so basically you can see that this is a kind of geodesic dome looks like a sphere we specifically designed this architecture so that we can put cameras as uniform as possible and you can see also that we specially mobilize the system so that we only design one panel and we can duplicate the panel's one panel is composed of 24 small cameras and one high resolution cameras and they are saved in the local hard drives and in the end all the local data are there transmitted to the NAS and so that sounds like a ton of data like how much data is in one minute of capture so one minute of data is about 600 petabytes that's huge data that's a big job in fact to this point the Panoptix to do has captured over 1 petabyte of data that's about a million gigabytes and the latest is well me yeah so while I'm dancing those 500 cameras are capturing every movement I make and after a couple of weeks of processing it looks something like this this point cloud view is not only showing a bunch of tracking points its tracing the past they could take while I'm moving around all right got a stretch before that now motion capture by itself isn't anything new similar technology has been used to bring digital characters to life and films for years now motion capture technology has gotten so good that an actor like Andy Serkis can really inject life into a fully digital character but to do so requires a full motion capture suit with tracking markers placed all over your body even paint it on your face now this works well enough for trying to bring a dragon to life but it's hard to feel natural in that kind of equipment the Panoptix to do in contrast allows for similar tracking but without any of the physical markers you can just freely move your body in a way that feels natural and unrestricted which is a necessity for them to capture what they are really looking for really good quote from Edward Sapir who's a linguist he says there's an elaborate and secret code that is known to no one but understood by all and he's referring to to all the gestures that we use to communicate and that's kind of the code that we're trying to sort of crack or understanding our goal essentially is to endow machines with the capacity to understand in social interactions so that's where tank comes back into play you remember tank the Robo ception esteem e via camera and react to my presence but he can't understand more than what I typed to him so when I start to get frustrated that he isn't helping me find the panopticon or react in a way that helped ease my tension which kind of makes him bad at his job machines have a very poor understanding of what all those things are so for example if if I'm not paying attention to you or or if I'm looking at you paying attention those are the kinds of things and machines just don't respond to right now and that's kind of a big deal studies have suggested that when we communicate with each other more than half of the message we're sending is based on the way we move our posture facial expressions and gestures can at times convey more than our words alone here let me show you so now you can see me and you can read what my body is telling you even if my words are saying something different maybe for my posture you can tell them in a cramped little sound booth or you can tell I'm excited about this story that I'm nervous about reading from a script that I'm hopelessly improvising into a camera to try and give you a sense of how expressive body language can be when we communicate each other we use many interesting hand gestures which is really important to communicate each other but it's very hard to understand why we are doing this motion when we are doing this motion right I don't know why I'm sitting I'm not trying to send you I mean I guess I'm sending you a signal but I don't know what it means exactly usually we are using our gestures but it's very hard to define why we are using it's not we are not doing this consciously right this is number adjusters I mean I'm just like trying to sit in a way that looks good okay I like your socks by the way thank you I'm trying to show off my socks that's why I'm sitting like this and that's the really weird thing about body language it's something we all do unconsciously and understand innately but if I were to stop and try to tell you why I'm holding my shoulders a certain way or moving my hands I wouldn't understand it and to be able to explain it so if we can't explain it to each other how can we possibly teach it to machines well it turns out we can teach them the same way we've been teaching a lot of other artificial intelligence systems machine learning specifically a technique called deep learning has produced a number of huge breakthroughs in the last few years a very simplistic view of it is you feed machines a lot of data about a certain subject like say driving a car feed in an update about how we drive and we'll be able to start understanding the system at play and react in a similar fashion the way we do most recently google's deepmind watched a few years worth of TV shows understand how to read human lips now it can do that better than most experts the panopticon enough visual data showing how humans interact it's deep learning system we'll be able to tease out the structure of that elaborate code making it possible for computers to read and write our body language possibly understanding a facet of our communication better than we do ourselves so if all this sounds familiar you are probably watching the new HBO series Westworld it's about a park filled with robots that can so vividly replicate human behavior it becomes impossible to distinguish who's a real human and who's not so Westworld is science fiction sure but to a quick youtube search for lifelike robots and you'll find we've actually made some incredible progress in the last few decades and some stuff to fill your nightmares find you and me be good friends but the benefits of these social robots could be immense look at Milo he's a robot that's being used to help kids with autism he could help these children identify social cues and emotional reactions that they can sometimes struggle to express and interpret themselves and robots will also help us care for the elderly so less Westworld and more Franken the robot beyond helping accomplish physical tasks future robots could become vital emotional supporters as well imagine instead of Robo ception estanque it helps someone suffering from Alzheimer's socialize on a more regular basis identifying their mood and reacting appropriately there's many components that need to happen before this works an AI in particular needs to get better right but I think that there will be a big advantage in terms of how we will treat machines once they start respecting at least for example whether we are paying attention to them or whether we are open to an interaction and other sort of social signals that that we use in our daily conversations but if all this talk of lifelike robots makes you nervous about the coming of some Skynet that'll wipe out humanity consider this teaching robots this social behavior might be the best way for us to foster a better understanding between humans and their creations so one way of putting it is that we use machines and even robots as tools currently but well what we will achieve when they interact with us as almost equals as we really find collaborators not so we won't start using machines as tools but rather as as press agents that we can collaborate with so I think in those terms that that will be very important so if we want to integrate machines and it's so eventually going to happen they should at least respond to us in a way that we appreciate this research is already being used by a company you probably know well Facebook the social networking giant and its oculus division are working with researchers from the Panoptix studio to find ways to make avatars in VR more accurate and social this could make communication with a friend more intimate or it could give Siri a body that can walk and talk just like a real person
We are a participant in the Amazon Services LLC Associates Program, an affiliate advertising program designed to provide a means for us to earn fees by linking to Amazon.com and affiliated sites.