Gadgetory


All Cool Mind-blowing Gadgets You Love in One Place

Project Tango Interview and Demo at Google I/O 2014

2014-06-26
hey it's Josh regard from android authority here at Google i/o at the project tango booth and I'm here with one of the engineers of project tango Johnny and I wanted to get some to get some extra information because I know that tango in some ways might be somewhat of a mystery to some of you out there so I wanted to get it from the actual people themselves what project tango is from their word so when we get a bit of an explanation about our project tango is and we have seen gaming examples of it but what else can project tango do sure they at a high level the goal our project tango is too advanced 3d sensing and tracking on mobile devices so we've been working with the hardware ecosystem and the software our partners to advance both the sensors and the algorithm is necessary to make these devices track their position and capture models of their space so one of the things that we've done it a lot recently is what you can do it just the tracking information where we take that camera and control either a unity basic 3d in my environment or an unreal engine based 3d environment we and we add the data from the depth sensor this is where it starts to get into a new code and new algorithms that aren't quite ready to put in the SDK yet but also why we haven't had so many demos but we have a few demos here that essentially as you walk through a space you're actually able to capture the geometry of the floor and the walls and the furniture and outside of gaming well inside of gaming that allows you to have characters that sort of know how to navigate through your hallways or potentially play hide-and-seek in your in your house and because it knows where your closet is but industrial applications and professional applications include real estate where you're trying to estimate the square footage of your home we have apps from Trimble which essentially start to explore well how much better could you make room scanning because rim scanning has ends up being important for shopping for furniture or being able to figure out how much corporate do you need or how many tiles that you want to put on the and then there's an industrial inspection so the trumbull also has another very early prototype of an experience where we've taken a 3d model of the most funny center and once we recognize the position of the device in the Moscone Center you can essentially get x-ray vision of where the elevators where's the service stations for this particular amount panel and without having to walk around the building sort of looking at the signs on the walls you actually can hold up the tablet and see sort of through the walls with sorry with a one one question I just came to mine for for various applications I say consumers would be using would the algorithms or the 3d schematic switch let's say would they be made available for the various applications or would the consumers be able to create those schematics themselves so the amount of data that the device can generate from scratch is relatively modest it basically is just the position of the device as you were holding it and then rough geometry information sort of i can pick out sort of rough walls and floors and some furniture it since it has a 3d sensor to capture that geometry it does not automatically know about say where does the hot water pipe run through the wall that requires having the CAD model in advance or some other source so and chances are people don't won't have that available in most buildings but for professional and industrial commercial structures that data may be available by the Builder so that it's a 7-inch tablet that has a Tegra k1 inside and our custom designed sensor so we have a fisheye camera that's designed for motion tracking which is in the middle and if you think about human vision we have a tremendously wide peripheral vision that expands up and gives us enormous amount of context about the way the environment is moving and so what this does is gives us a wide field of view of the environment and then we have a set of sensors on here that are depth sensors these are sensors that basically give us information about the floor and the walls to tell us what the geometry is in this particular tablet we have structured light depth sensor from a company called mantis but we also have a prototype hear of a depth sensor from a company PMD tensions that has a different principle of operation which is time of flight and the time of flight has slightly different properties and that it doesn't require a large separation of sensors and potentially is more robust to sunlight but as we evolve the hardware will continue to sort of push the push the envelope in terms of the performance of those deficits is true so let me give you a quick tour of the software that actually runs on a project tango device this is our Diagnostics tool which just gives us a view of the sensor data coming in so on the left side of the screen you can see this fisheye lens in fact I can still see that my fingers just around the tablet and so this is a super wide camera that that allows us to understand the motion of the device the left points are being tracked yes so those green points are essentially the image processing doing optical flow so this allows us to understand while everything moved in the world in a way that makes sense to the way that device was help we also have the gyro and accelerometer data right below the image that is essentially time synchronized with the with the camera so this allows all the sensor data to come into the same time Orwell timestamp so we can make an estimate of where the device is but if I actually move the tablet left and right or actually make a circle you can actually see that it's actually tracing out a circle on this one over on the screen yeah so I'm going to just walk around here a little bit and you'll see the tracking showing up on I'll point at it over here yeah so Johnny's gonna go off camera and you'll see the tracking right here is actually taking where he is and there it goes you can see it's actually mapping the entire area that he is walking through using all of those sensors and now we have this sort of this kind of janky figure eight right here oh that's great yeah so this is tracing out my path so one of the things we can do is insert that path into a game environment so this is one of the unity applications that we have will be part of the SDK which is how do you actually use this data for end user experience and so what we've done is we've connected that tracking data to the unity camera and this is a very simple game where essentially I just pick up the cube and put it on the switch but to actually finish this game I you can see that I can't quite reach the cube from where I am now so actually have to go walk forward to solve anything is that using the data that you just tracked that's right all right so so that this ability to track the physical motion of the device is now being piped into unity so if I pick up this green switch buck sorry I can see that the green switch is actually over there on the booth you hit them so I'll come back alright and there you go so this so this basically introduces motion into mobile gaming so if I move forward in the real world I actually move forward in the virtual space as well so this allows game developers application developers to start thinking about well how would I actually use my physical room either my my living room or my office space and actually use this inside the game which is if you can imagine now that I understand where I am in my house or my environment and I actually have a little bit of geometry data I can sort of transform my environment into a fantasy world so again i'm just using the tablet as a camera controller i can look at the trees and the stones but i also see on the ground there's a small wizard so if we crouched if I crouch down I can get down to the Wizards level and interact with him so essentially if I have the ability to place these assets i can start transforming parts of my living room into a fantasy environment so I can't imagine you know my living room is fantasy world versus the my bedroom is another sort of safe haven or you know in Iceland and you can actually use physical space as part of the game if we had game a game that actually understood the shape of the Moscone Center you can imagine actually having multiple people walk around the space competing with each other for territory you think that would not be very dangerous though like I actually have a personal desire to see a game at some point that causes two people to actively tackle each other but we'll see that actually comes to comes to fruition here's an app from a university partner that we just got recently that actually combines the real-time data of the tracking with the depth sensing that we have I will see you this runs this was an app that we just got a few days ago so what this is doing right now is actually combining both the tracking data and the depth sensor data to allow us to create a 3d capture of the space oh it's like oh that's actually imaging the yeah so this is like well I'm painting the environment with my camera and I can back up when i'll scan you into this place okay that's supposed to be me right there yeah so this is a you know a very simple mesh that we're building in real time but as the software improves and as the hardware improves in terms of accuracy we can imagine that the quality of this data continues to get higher and higher fidelity if we actually stored the data and did more post-processing on it app developers could actually get much much higher quality but this is just what we could process in real time so these are sort of the building blocks you can imagine allowing developers to start building use new user experiences that use both motion as well as physical geometry to create new eps ok do you think that next year maybe for i 0 20 15 or 20 or 16 you'll be able to create that game using the Moscone Center schematic I think it would be a really neat target to aim for oh cool we'll have to see there's a lot of things that have to sort of improve on both the software side and the hardware side to make it happen I'm sure it's gonna get there I mean yeah some great minds working on end with the way the word is already i'm almost sounded this is really wonderful like it's great all right well thank you very much and yeah you know keep it tuned here we have project tango as it as it happens we're going to be covering it as it comes along it's a wonderful wonderful sweet of technology that we have here and Johnny one of the minds behind it thank you so much for for fielding our questions and for giving us a demo of it thank you right take care all right this android authority coming to you live from google i/o 2014
We are a participant in the Amazon Services LLC Associates Program, an affiliate advertising program designed to provide a means for us to earn fees by linking to Amazon.com and affiliated sites.