Placeholder Image

Subtitles section Play video

  • [MUSIC PLAYING]

  • CHRIS KELLEY: Thank you so much for joining us.

  • My name is Chris.

  • I'm a designer and prototyper working

  • on immersive prototyping at Google,

  • and I'm joined by Ellie and Luca.

  • And today, we're going to talk about exploring AR interaction.

  • It's really awesome to be here.

  • We explore immersive computing through rapid prototyping

  • of AR and VR experiments.

  • Often, that's focused on use case exploration or app ideas.

  • We work fast, which means we fail fast,

  • but that means that we learn fast.

  • We spend a week or two on each prototyping sprint,

  • and at the end of the sprint, we end

  • with a functional prototype starting

  • from a tightly scoped question.

  • And then we put that prototype in people's hands

  • and we see what we can learn.

  • So this talk is going to be about takeaways we have

  • from those AR explorations.

  • But first, I want to set the table a little bit

  • and talk about what we mean when we say augmented reality.

  • When a lot of people think about AR,

  • the first thing they think about is bringing virtual objects

  • to users in the world.

  • And it is that.

  • That's part of it.

  • We call this the out of AR.

  • But AR also means more than that.

  • It means being able to understand the world visually

  • to bring information to users, and we call this understanding

  • the in of AR.

  • Many of the tools and techniques that

  • were created for computer vision and machine learning perfectly

  • complement tools like ARCore, which is Google's AR

  • development platform.

  • So when we explore AR, we build experiences

  • that include one of these approaches or both.

  • So this talk is going to be about three magic powers

  • that we've found for AR.

  • We think that these magic powers can help you build better AR

  • experiences for your users.

  • So we're going to talk about some prototypes that we've

  • built and share our learnings with you

  • during each of these three magic power areas during the talk.

  • First, I'll talk to you about context-driven superpowers.

  • That's about how we can combine visual and physical

  • understanding of the world to make magical AR experiences.

  • Then Ellie will talk to you about shared augmentations.

  • And this is really all about the different ways

  • that we can connect people together in AR,

  • and how we can empower them just by putting them together.

  • And then Luca will cover expressive inputs.

  • This is about how AR can help unlock

  • authentic and natural understanding for our users.

  • So let's start about context-driven superpowers.

  • What this really means is using AR technologies

  • to deeply understand the context of a device,

  • and then build experiences that directly leverage that context.

  • And there's two parts to an AR context.

  • One is visual understanding, and the other

  • is physical understanding.

  • With ARCore, this gives your phone

  • the ability to understand and sense

  • its environment physically.

  • But through computer vision and machine learning,

  • we can make sense of the world visually.

  • And by combining these results, we

  • get an authentic understanding of the scene,

  • which is a natural building block of magical AR.

  • So let's start with visual understanding.

  • The prototyping community has done some awesome explorations

  • here, and we've done a few of our own

  • that we're excited to share.

  • To start, we wondered if we could

  • trigger custom experiences from visual signals in the world.

  • Traditional apps today leverage all kinds of device

  • signals to trigger experiences.

  • GPS, the IMU, et cetera.

  • So could we use visual input as a signal as well?

  • We built a really basic implementation of this concept.

  • This uses ARCore and the Google Cloud Vision

  • API that detects any kind of snowman

  • in the scene, which triggers a particle system that

  • starts to snow.

  • So through visual understanding, we

  • were able to tailor an experience to specific cues

  • in the environment for users.

  • This enables adaptable and context aware applications.

  • So even though this example is a simple one,

  • the concept can be extended so much further.

  • For example, yesterday we announced the augmented images

  • API for ARCore.

  • So if you use this, you can make something

  • like an experience that reacts relative to device movement

  • around an image in the scene, or even

  • from a known distance to an object in the world.

  • If you think this concept is interesting,

  • I highly recommend checking out the AR VR demo tent.

  • They have some amazing augmented images demos there.

  • The next thing we wanted to know is

  • if we could bridge the gap between digital and physical,

  • and, for example, bring some of the most delightful features

  • of e-readers to physical books.

  • The digital age has brought all kinds of improvements

  • to some traditional human behaviors,

  • and e-readers have brought lots of cool new things to reading.

  • But if you're like me, sometimes you just

  • missed the tactility in holding a great book in your hands.

  • So we wanted to know if we could bridge that gap.

  • In this prototype, users highlight a passage or word

  • with their finger and they instantly

  • get back a definition.

  • This is a great example of a short-form-focused interaction

  • that required no setup for users.

  • It was an easy win only made possible

  • by visual understanding.

  • But as soon as we tried this prototype,

  • there were two downfalls that we noticed,

  • and they became immediately apparent when we used it.

  • The first is that it was really difficult to aim your finger

  • at a small moving target on a phone,

  • and maybe the page is moving as well,

  • and you're trying to target this little word.

  • That was really hard.

  • And the second was that when you're highlighting a word,

  • your finger is blocking the exact thing

  • that you're trying to see.

  • Now, these are easily solvable with a follow-up UX iteration,

  • but they illustrate a larger lesson.

  • And that's that with any kind of immersive computing,

  • you really have to try it before you can judge it.

  • An interaction might sound great when you talk about it

  • and it might even look good in a visual mock,

  • but until you have it in your hand

  • and you can feel it and try it, you're

  • not going to know if it works or not.

  • You really have to put it in a prototype

  • so you can create your own facts.

  • Another thing we think about a lot

  • is, can we help people learn more effectively?

  • Could we use AR to make learning better?

  • There's many styles of learning, and if you

  • combine these styles of learning,

  • it often results in faster and higher-quality learning.

  • In this prototype, we combined visual, oral, verbal,

  • and kinesthetic learning to teach people how

  • to make the perfect espresso.

  • The videos explain--

  • I'm sorry.

  • We placed videos around the espresso machine

  • in the physical locations where that step occurs.

  • So if you were learning how to use the grinder,

  • the video for the grinder is right next to it.

  • Now, for users to trigger that video,

  • they move their phone to the area

  • and then they can watch the lesson.

  • That added physical component of the physical proximity

  • of the video and the actual device

  • made a huge difference in general understanding.

  • In our studies, users who had never used an espresso machine

  • before easily made an espresso after using this prototype.

  • So for some kinds of learning, this

  • can be really beneficial for users.

  • Now, unfortunately for our prototype,

  • one thing that we learned here was

  • that it's actually really hard to hold your phone

  • and make an espresso at the same time.

  • So you need to be really mindful of the fact

  • that your users might be splitting

  • their physical resources between the phone and the world.

  • And so as it applies to your use case,

  • try building experiences that are really

  • snackable and hands-free.

  • Speaking of combining learning and superpowers together,

  • we wondered if AR could help us learn

  • from hidden information that's layered in the world

  • all around us.

  • This is a prototype that we built

  • that's an immersive language learning app.

  • We showed translations roughly next to objects of interest

  • and positioned these labels by taking a point cloud

  • sample from around the object and putting the label sort

  • of in the middle of the points.

  • Users found this kind of immersive learning really fun,

  • and we saw users freely exploring

  • the world looking for other things to learn about.

  • So we found that if you give people

  • the freedom to roam and tools that are simple and flexible,

  • the experiences that you build for them

  • can create immense value.

  • We now have physical understanding.

  • This is AR's ability to extract and infer

  • information and meaning from the world around you.

  • When a device knows exactly where it is, not only in space,

  • but also relative to other devices,

  • we can start to do things that really

  • feel like you have superpowers.

  • For example, we can start to make

  • interactions that are extremely physical, natural,

  • and delightful.

  • Humans have been physically interacting

  • with each other for a really long time,

  • but digital life has abstracted some of those interactions.

  • We wondered if we could swing the pendulum

  • back the other direction a little bit using AR.

  • So in this prototype, much like a carnival milk bottle game,

  • you fling a baseball out of the top of your phone

  • and it hits milk bottles that are shown on other devices.

  • You just point the ball where you want to go, and it goes.

  • We did this by putting multiple devices

  • in a shared coordinate system, which

  • you could do using the new Google Cloud Anchors API

  • that we announced for ARCore yesterday.

  • And one thing you'll notice here is

  • that we aren't even showing users past their camera.

  • Now, we did that deliberately because we really

  • wanted to stretch and see how far we

  • could take this concept of physical interaction.

  • And one thing we learned was that once people learned

  • to do it, they found it really natural

  • and actually had a lot of fun with it.

  • But almost every user that tried it had to be not only

  • told how to do it, but shown how to do it.

  • People actually had to flip this mental switch

  • of the expectations they have for how a 2D smartphone

  • interaction works.

  • So you really need to be mindful of the context that people

  • are bringing in and the mental models they have

  • for 2D smartphone interactions.

  • We also wanted to know if we could help someone visualize

  • the future in a way that would let them make better decisions.

  • Humans pay attention to the things that matter to us.

  • And in a literal sense, the imagery

  • that appears in our peripheral vision

  • takes a lower cognitive priority than the things

  • we're focused on.

  • Would smartphone AR be any different?

  • In this experiment, we overlaid the architectural mesh

  • of the homeowner's remodel on top of the active construction

  • project.

  • The homeowner could visualize in context

  • what the changes to their home was going to look like.

  • Now, at the time that this prototype was created,

  • we had to do actual manual alignment of this model

  • on top of the house.

  • You could do it today.

  • If I rebuilt it, I would use the augmented images API

  • that we announced yesterday.

  • It would be much easier to put a fixed

  • image in a location, the house, and sync them together.

  • But even with that initial friction for the UX,

  • the homeowner got tremendous value out of this.

  • In fact, they went back to their architect after seeing this

  • and changed the design of their new home

  • because they found out that they weren't going

  • to have enough space in the upstairs bathroom-- something

  • they hadn't noticed in the plans before.

  • So the lesson is that if you provide people high-quality,

  • personally relevant content, you can create ways

  • that people will find really valuable and attention grabbing

  • experiences.

  • But when does modifying the real environment

  • start to break down?

  • You may be familiar with the uncanny valley.

  • It's a concept that suggests when

  • things that are really familiar to humans

  • are almost right but just a little bit off,

  • it makes us feel uneasy.

  • Subtle manipulations of the real environment in AR

  • can sometimes feel similar.

  • It can be difficult to get right.

  • In this specific example, we tried

  • removing things from the world.

  • We created this AR invisibility cloak for the plant.

  • What we did was we created a point cloud around the object,

  • attached little cubes to the point cloud,

  • applied a material to those points,

  • and extracted the texture from the surrounding environment.

  • That worked pretty well in uniform environments,

  • but unfortunately, the world doesn't have too many of those.

  • It's made up of dynamic lighting and subtle patterns,

  • so this always ended up looking a little bit weird.

  • Remember to be thoughtful about the way that you add

  • or remove things from the environment.

  • People are really perceptive, and so you

  • need to strive to build experiences

  • that align with their expectations,

  • or at the very least, don't defy them.

  • But is physical understanding always critical?

  • All points in the section have their place,

  • but, ultimately, you have to be guided by your critical user

  • journeys.

  • In this example, we wanted to build

  • a viewer for this amazing 3D model by Damon [INAUDIBLE]..

  • It was important that people could see the model in 3D

  • and move around to discover the object.

  • A challenge, though, was that the camera feed

  • was creating a lot of visual noise and distraction.

  • People were having a hard time appreciating

  • the nuances of the model.

  • We adopted concepts from filmmaking and guided users

  • by using focus and depth of field,

  • all which were controlled by the user's motion.

  • This resulted in people feeling encouraged to explore,

  • and they really stopped getting distracted

  • by the physical environment.

  • So humans are already great at so many things.

  • AR really allows us to leverage those existing capabilities

  • to make interactions feel invisible.

  • If we leverage visual and physical understanding

  • together, we can build experiences that

  • really give people superpowers.

  • With that, Ellie is going to talk to you

  • about special opportunities we have in shared augmentations.

  • ELLIE NATTINGER: Thanks, Chris.

  • So I'm Ellie Nattinger.

  • I'm a software engineer and prototyper

  • on Google's VR and AR team.

  • Chris has talked about the kinds of experiences

  • you start to have when your devices can understand

  • the world around you, and I'm going

  • to talk about what happens when you can share those experiences

  • with the people around you.

  • We're interested not only in adding AR augmentations

  • to your own reality, but also in sharing those augmentations.

  • If you listened to the developer keynote yesterday,

  • you know that shared AR experiences

  • is a really big topic for us these days.

  • For one thing, a shared reality lets

  • people be immersed in the same experience.

  • Think about a movie theater.

  • Why do movie theaters exist?

  • Everybody's watching a movie that they could probably

  • watch at home on their television or their computer

  • by themselves much more comfortably not

  • having to go anywhere, but it feels qualitatively

  • different to be in a space with other people sharing

  • that experience.

  • And beyond those kinds of shared passive experiences,

  • having a shared reality lets you collaborate, lets you learn,

  • lets you build and play together.

  • We think you should be able to share your augmented realities

  • with your friends, and your families, and your colleagues,

  • so we've done a variety of explorations

  • about how do you build those kinds of shared

  • realities in AR.

  • First, there's kind of a technical question.

  • How do you get people aligned in a shared AR space?

  • There's a number of ways we've tried.

  • If you don't need a lot of accuracy,

  • you could just start your apps with all the devices

  • in approximately the same location.

  • You could use markers or augmented images

  • so multiple users can all point their devices at one picture

  • and get a common point of reference--

  • cures the zero, zero, zero of my virtual world.

  • And you can even use the new ARCore Cloud Anchors API

  • that we just announced yesterday to localize

  • multiple devices against the visual features

  • of a particular space.

  • In addition to the technical considerations,

  • we've found three axes of experience

  • that we think are really useful to consider

  • when you're designing these kinds of shared

  • augmented experiences.

  • First of those is co-located versus remote.

  • Are your users in the same physical space

  • or different physical spaces?

  • Second is, how much precision is required, or is it optional?

  • Do you have to have everybody see the virtual bunny

  • at exactly the same point in the world,

  • or do you have a little bit of flexibility about that?

  • And the third is whether your experience

  • is synchronous or asynchronous.

  • Is everybody participating in this augmented experience

  • at exactly the same time, or at slightly different times?

  • And we see these not as necessarily binary axes,

  • but more of a continuum that you can

  • consider when you're designing these multi-person AR

  • experiences.

  • So let's talk about some prototypes and apps that

  • fall on different points of the spectrum and the lessons

  • we've learned from them.

  • To start with, we've found that when

  • you've got a group that's interacting

  • with the same content in the same space,

  • you really need shared, precise, spatial registration.

  • For example, let's say you're in a classroom.

  • Imagine if a group of students who

  • are doing a unit on the solar system could all look at

  • and walk around the globe, or an asteroid field,

  • or look at the sun.

  • In Expeditions AR, one of Google's initial AR

  • experiences, all the students can point their devices

  • to a marker, they calibrate themselves

  • against a shared location, they see the object

  • in the same place, and then what this allows

  • is for a teacher to be able to point out

  • particular parts of the object.

  • Oh, if you all come over and look at this side of the sun,

  • you see a cut-out into its core.

  • Over here on the Earth, you can see a hurricane.

  • Everybody starts get a spatial understanding

  • of the parts of the object and where they are in the world.

  • So when does it matter that your shared

  • space has a lot of precision?

  • When you have multiple people who

  • are all in the same physical space

  • interacting with or looking at the exact same

  • augmented objects at the same time.

  • We were also curious--

  • how much can we take advantage of people's existing spatial

  • awareness when you're working in high-precision shared spaces?

  • We experimented with this in this multi-person construction

  • application, where you've got multiple people who

  • are all building onto a shared AR object in the same space.

  • Adding blocks to each other, everybody's

  • being able to coordinate.

  • And you want to be able to tell what part of the object someone

  • is working on.

  • Have your physical movement support that collaboration.

  • Like, if Chris is over here and he's

  • placing some green blocks in the real world,

  • I'm not going to step in front of him

  • and start putting yellow blocks there instead.

  • We've got a natural sense of how to collaborate, how to arrange,

  • how to coordinate ourselves in space.

  • People already have that sense.

  • So we can keep that in a shared AR

  • if we've got our virtual objects precisely lined up enough.

  • We also found it helpful to notice

  • that because you can see both the digital object but also

  • the other people through the pass-through camera,

  • you are able to get a pretty good sense of what people

  • were looking at as well as what they were interacting with.

  • We've also wondered what would it

  • feel like to have a shared AR experience for multiple people

  • in the same space, but who aren't necessarily interacting

  • with the same things?

  • So think of this more like an AR LAN party.

  • Where we're all in the same space,

  • or maybe could be different spaces,

  • we're seeing connected things, and we're

  • having a shared experience.

  • So this prototype's a competitive quiz guessing game

  • where you look at the map and you

  • have to figure out where on the globe you think is represented

  • and stick your pushpin in, get points

  • depending on how close you are.

  • We've got the state synced, so we know who's winning.

  • But the location of where that globe is

  • doesn't actually need to be synchronized.

  • And maybe you don't want it to be synchronized because I don't

  • want anybody to get a clue based on where I'm sticking

  • my pushpin into the globe.

  • It's fun to be together, even when we're not looking

  • at exactly the same AR things.

  • And do we always need our spaces to align exactly?

  • Sometimes it's enough just to be in the same room.

  • This prototype example's of an AR boat race.

  • You blow on the microphone of your phone,

  • and it creates the wind that propels your boat

  • down the little AR track.

  • By us being next to each other when we start the app

  • and spawn the track, we get a shared physical experience

  • even though our AR worlds might not perfectly align.

  • We get to keep all the elements of the social game play--

  • talking to each other, our physical presence--

  • but we're not necessarily touching the same objects.

  • Another super interesting area we've been playing with

  • is how audio can be a way to include

  • multiple people in a single device AR experience.

  • If you think of the standard Magic Window device AR,

  • it's a pretty personal experience.

  • I'm looking at this thing through my phone.

  • But now, imagine you can leave a sound in AR that

  • has a 3D position like any other virtual thing,

  • and now you start to be able to hear it,

  • even if you're not necessarily looking at it.

  • And other people can hear the sound

  • from your device at the same time.

  • So for an example, let's say you could leave little notes

  • all over your space.

  • Might look something like this.

  • I'm a plant.

  • I'm a plant.

  • I'm a plant.

  • I'm elephant.

  • I'm elephant.

  • I'm elephant.

  • This is a chair.

  • This is a chair.

  • This is a chair.

  • I'm a plant.

  • I'm a plant.

  • I'm elephant.

  • I'm elephant.

  • This is a chair.

  • This is a chair.

  • So notice, you don't have to be the one with a phone

  • to get a sense of where these audio annotations start

  • to live in physical space.

  • Another question we've asked--

  • if you have a synchronous AR experience with multiple people

  • who are in different places, what kind of representation

  • do you need of the other person?

  • So let's imagine you have maybe a shared AR photos app

  • where multiple people can look at photos

  • that are arranged in space.

  • So I'm taking pictures in one location,

  • I'm viewing them arranged around me in AR,

  • and then I want to share my AR experience

  • with Luca, who comes in and joins me

  • from a remote location.

  • What we found-- we needed a couple of things to make

  • us feel like we were connected and sharing the same AR

  • experience, even though we were in different places.

  • We needed to have a voice connection so we could actually

  • talk about the pictures, and we needed

  • to know where the other person was looking.

  • See which picture you're paying attention to when

  • you're talking about it.

  • But what was interesting is we didn't actually

  • need to know where the other person was, as long as we had

  • that shared frame of reference.

  • We're all here, here's what I'm looking at,

  • here's what Luca's looking at.

  • We've also been curious about asymmetric experiences.

  • What happens when users share the same space

  • and the same augmentations, but they've got different roles

  • in the experience?

  • So for instance, in this prototype,

  • Chris is using his phone as a controller to draw in space,

  • but he's not actually seeing the AR annotations he's drawing.

  • The other person sees the same AR content

  • and uses their phone to take a video.

  • They're playing different roles in the same experience.

  • Kind of artist versus cinematographer.

  • And we found there could be some challenges

  • to asymmetric experiences if there's

  • a lack of information about what the other person is

  • experiencing.

  • For instance, Chris can't tell what Luca's filming

  • or see how his drawing looks from far away.

  • So as we mentioned previously, these kinds

  • of different combinations of space,

  • and time, and precision are relevant for multi-person AR

  • experiences, and they have different technical and

  • experiential needs.

  • If you have multiple people in the same space

  • with the same augmentations at the same time,

  • then you need a way of sharing.

  • You need a way of common localization.

  • That's why we created the new Cloud Anchors API.

  • If you've got multiple people in the same space

  • with different augmentations at the same time,

  • the kind of AR LAN party model, you

  • need some way to share data.

  • And if you've got multiple people

  • in different spaces interacting with the same augmentations

  • at the same time, you need sharing

  • in some kind of representation of that interaction.

  • So shared AR experiences is a big area.

  • We've explored some parts of the space.

  • We'd love to see what you all come up with.

  • So Chris has talked about examples

  • where your device understands your surroundings

  • and gives you special powers, I talked about examples

  • where you've got multiple people who

  • can collaborate and interact.

  • Now Luca will talk about what happens when your devices have

  • a better understanding of you and allow

  • for more expressive inputs.

  • Luca?

  • LUCA PRASSO: Thank you, Ellie.

  • My name is Luca Prasso, and I'm a prototyper

  • and a technical artist working in the Google AR and VR team.

  • So let's talk about the device that you

  • carry with you every day and the ones that are all around you,

  • and how they can provide the meaningful and authentic

  • signals that we can use in our augmented experiences.

  • So ARCore tracks the device motion

  • as we move to the real world and provides some understanding

  • of the environment.

  • And these signals can be used to create powerful, and creative,

  • and expressive tools, and offer new ways for us

  • to interact with digital content.

  • So the data represents who we are, what we know,

  • and what we have.

  • And we were interested in understanding

  • if the user can connect more deeply if the data is displayed

  • around them in 3D, and through AR and physical aspirations,

  • they can look at this data.

  • So we took a database of several thousand world cities,

  • and we mapped it in an area that's

  • wide as a football field.

  • We assign a dot to every city and we scale the dot based

  • on the population of the city.

  • And each country has a different color.

  • So now you can walk to this data field.

  • And as ARCore tracks the motion of the user,

  • we play footsteps in sync.

  • You take a step and you hear a step.

  • And [INAUDIBLE] sound fields surrounds the user

  • and enhances the experience and the sense of exploration

  • of this data forest.

  • And flight paths are displayed up in the sky.

  • And the pass-through camera is heavily tinted

  • so that we can allow the user to focus on the data

  • and then still give a sense of presence.

  • And what happens is the user, as he walks to the physical space,

  • he starts mapping, and pairing, and creating

  • this mental map between the data and the physical location.

  • And starts understanding better, in this particular case,

  • the relative distance between the places.

  • And what we discover is also that the gestures that

  • are a part of our digital life every day, a pinch to zoom,

  • it's now in AR something more traditional.

  • It's actually moving closer to the digital object

  • and inspecting it like we do with a real object.

  • And pan and drag means taking a couple of steps

  • to the right to look at the information.

  • So physical exploration like this is very fascinating,

  • but we need to take into account all the different users

  • and provide the alternative move and affordances.

  • So in AR, a user can move everywhere,

  • but what if he cannot or he doesn't want to move?

  • What if he's sitting?

  • So in this particular case, we allow the user to simply point

  • the phone everywhere they want to go, tap on the screen

  • anywhere, and the application will move the point of view

  • in that direction.

  • At the same time, we still have to provide audio, haptics,

  • and color effects to enhance the sense of physical space

  • the user has to have while traveling.

  • And so we found that this is a powerful mechanism

  • to explore a certain type of data that makes sense in the 3D

  • space and to allow the user to discover hidden patterns.

  • But can we go beyond the pixels that you

  • can find on your screen?

  • We're fascinated by the spatial audio and a way

  • to incorporate audio into an AR experience.

  • So we combine ARCore and the Google Resonance SDK.

  • And Resonance is this very powerful spatial audio engine

  • that recently Google open-sourced.

  • And you should check it out because it's great.

  • And so now I can take audio sources

  • and place them into the 3D locations,

  • and animate them, and describe the properties of the walls,

  • and the ceilings, and the floor, and all the obstacles.

  • And now as the ARCore moves the point of view,

  • it carries with it the digital ears,

  • the Resonance used to render accurately

  • the sounds in the scene.

  • So what can we do with this?

  • So we imagine, what if I can sit next to a performer

  • during an acoustic concert, or a classical concert, or a jazz

  • performance?

  • What if I can be onstage with actors,

  • and listen to their play, and be there?

  • So we took two amazing actors, Chris and Ellie,

  • and we asked them to record separately

  • lines from Shakespeare.

  • And we placed these audio sources a few feet apart

  • and we surrounded the environment

  • with an ambisonic sound field of a rain forest, of the raining.

  • And then later on, we switched to a room with a lot of reverb

  • into the walls.

  • CHRIS KELLEY: Thou told'st me they were stolen unto this

  • wood, and here I am, and wode within this wood,

  • because I cannot meet my Hermia.

  • Hence, get thee gone, and follow me no more.

  • ELLIE NATTINGER: You draw me, you hard-hearted adamant,

  • but yet you draw not iron, for my heart is true as steel.

  • Leave you your power to draw, and I shall

  • have no power to follow you.

  • CHRIS KELLEY: Do I entice you?

  • Do I speak to you fair?

  • Or rather, do I not in plainest truth tell you, I do not,

  • nor I cannot love you?

  • LUCA PRASSO: So now the user can walk around,

  • maybe with his eyes closed, a nice pair of headphones,

  • and it's like being on stage with these actors.

  • So we took this example and we extended it.

  • We observed that we can build in real-time a 2D map of where

  • the user has been so far with his phone

  • as he's walking around.

  • And so at any given time when the user hits a button,

  • we can programmatically place audio recording in space

  • where we know that the user can reach with the phone

  • and with their ears.

  • [MUSIC PLAYING]

  • And suddenly, the user becomes the human mixer

  • of this experience.

  • And different instruments can populate

  • your squares, and your rooms, and your schools.

  • And this opens the door to an amazing amount

  • of opportunities with AR audio-first experiments.

  • So let's go back to visual understanding.

  • Chris mentioned that the computer vision and machine

  • learning can interpret the things that are around us,

  • and this is also important to understand the body in turning

  • into an expressive controller.

  • So in real life, we are surrounded

  • by a lot of sound sources for all of the places.

  • And naturally, our body and our head

  • moves to mix and focus on what we like

  • and what we want to listen to.

  • So can we take this intuition into the way we watch movies

  • or play video games on a mobile device?

  • So what we did, we took the phone camera signal,

  • fed it to Google Mobile Vision.

  • That gave us a head position and head orientation.

  • And we fed it to Google Resonance SDK.

  • And we said, OK, you're watching a scene in which actors

  • are in a forest, and they're all around you, and it's raining.

  • So now as I leave my phone far away from my head,

  • I hear the forest.

  • As I'm taking the phone closer to my face,

  • I start hearing the actors playing.

  • I warn you, this is an Oscar performance.

  • [THUNDER RUMBLES]

  • ELLIE NATTINGER: Our company here.

  • CHRIS KELLEY: My man, according to the script.

  • ELLIE NATTINGER: Here is the scroll of every man's name

  • which is thought fit through all Athens to play in our interlude

  • before the duke and the duchess on his [INAUDIBLE]

  • LUCA PRASSO: So now what is interesting is

  • that the tiny little motions that we

  • can do when we're watching and we're playing this experience,

  • it can be turned into subtle changes in the user

  • experience that we can control.

  • So we talk about how the changes in poses

  • can become a trigger to drive interaction.

  • In this Google Research app called [INAUDIBLE],,

  • we actually exploit the opposite--

  • the absence of motion.

  • And when the user--

  • in this case, my kids--

  • stop posing, the app takes a picture.

  • And so the simple mechanism that is triggered by computer vision

  • creates the incredible, delightful opportunities

  • that, apparently, my kids love.

  • And Research is doing incredible progress

  • in looking at an RGB image and understanding where

  • the body pose and skeleton is.

  • And you should check out the Google Research blog post

  • because their post estimation research is amazing.

  • So we took Ellie's video and we fed it to the machine computer

  • algorithm.

  • And we got back, a bunch of 3D poses and segmentation

  • masks of Ellie.

  • And this opens the door to a lot of variety

  • of experiments with creative filters

  • that we can apply to this.

  • But what's more interesting for us

  • is that it also allows us to understand better

  • the intent and the context of the user.

  • So we took this pose estimation technology

  • and we added a digital character.

  • Now it tries to mimic what the human character is doing.

  • And this allows [INAUDIBLE] now to bring your family

  • and friends--

  • in this case, my son, Noah--

  • into the scene so that he can act and create a nice video.

  • But this also, like Ellie mentioned before,

  • we should consider the situation,

  • because this is an asymmetric experience.

  • What you don't see here is how frustrated

  • my son was after a few minutes because he

  • couldn't see what was going on.

  • I was the one having fun taking picture and video him,

  • and he didn't see much.

  • He could only hear the lion roaring.

  • So we need to be extremely mindful

  • as the developer about this unbalance of delight.

  • And so maybe I should have passed the image of the phone

  • to a nearby TV so I can make my son first-class citizen

  • in this experience.

  • So all this AR technology and the physical

  • and the visual understanding are ingredients

  • that allow us to unlock all kinds of new expressive input

  • mechanisms.

  • And we are still exploring.

  • We're just at the beginning of this journey.

  • But we are excited to hear what you think

  • and what you want to come up with.

  • So to summarize, we shared a bunch

  • of ways in which we think about AR and various aspirations

  • that we have done.

  • We talked about expanding our definition of AR.

  • Putting content into the world, but also pulling information

  • from the world.

  • And these are all ingredients that we

  • use to create these magical AR superpowers to enhance

  • the social interactions and to express yourself

  • in this new digital medium.

  • So we combined ARCore capabilities

  • with different Google technologies,

  • and this gives us the opportunity

  • to explore all these new interaction models.

  • And we encourage you, developers,

  • to stretch your definition of AR.

  • But we want to do this together.

  • We're going to keep exploring, but what

  • we want to hear what tickled you,

  • what tickled your curiosity.

  • So we can wait to see what you build next.

  • Thank you very much for coming.

  • [MUSIC PLAYING]

[MUSIC PLAYING]

Subtitles and vocabulary

Click the word to look it up Click the word to find further inforamtion about it