a908d668c7
GitOrigin-RevId: 4cee4a2c2317fb190680c17e31ebbb03bb73b71c
123 lines
7.0 KiB
Markdown
123 lines
7.0 KiB
Markdown
---
|
|
layout: default
|
|
title: Instant Motion Tracking
|
|
parent: Solutions
|
|
nav_order: 9
|
|
---
|
|
|
|
# MediaPipe Instant Motion Tracking
|
|
{: .no_toc }
|
|
|
|
1. TOC
|
|
{:toc}
|
|
---
|
|
|
|
## Overview
|
|
|
|
Augmented Reality (AR) technology creates fun, engaging, and immersive user
|
|
experiences. The ability to perform AR tracking across devices and platforms,
|
|
without initialization, remains important to power AR applications at scale.
|
|
|
|
MediaPipe Instant Motion Tracking provides AR tracking across devices and
|
|
platforms without initialization or calibration. It is built upon the
|
|
[MediaPipe Box Tracking](./box_tracking.md) solution. With Instant Motion
|
|
Tracking, you can easily place virtual 2D and 3D content on static or moving
|
|
surfaces, allowing them to seamlessly interact with the real-world environment.
|
|
|
|
![instant_motion_tracking_android_small](../images/mobile/instant_motion_tracking_android_small.gif) |
|
|
:-----------------------------------------------------------------------: |
|
|
*Fig 1. Instant Motion Tracking is used to augment the world with a 3D sticker.* |
|
|
|
|
## Pipeline
|
|
|
|
The Instant Motion Tracking pipeline is implemented as a MediaPipe
|
|
[graph](https://github.com/google/mediapipe/tree/master/mediapipe/graphs/instant_motion_tracking/instant_motion_tracking.pbtxt),
|
|
which internally utilizes a
|
|
[RegionTrackingSubgraph](https://github.com/google/mediapipe/tree/master/mediapipe/graphs/instant_motion_tracking/subgraphs/region_tracking.pbtxt)
|
|
in order to perform anchor tracking for each individual 3D sticker.
|
|
|
|
We first use a
|
|
[StickerManagerCalculator](https://github.com/google/mediapipe/tree/master/mediapipe/graphs/instant_motion_tracking/calculators/sticker_manager_calculator.cc)
|
|
to prepare the individual sticker data for the rest of the application. This
|
|
information is then sent to the
|
|
[RegionTrackingSubgraph](https://github.com/google/mediapipe/tree/master/mediapipe/graphs/instant_motion_tracking/subgraphs/region_tracking.pbtxt)
|
|
that performs 3D region tracking for sticker placement and rendering. Once
|
|
acquired, our tracked sticker regions are sent with user transformations (i.e.
|
|
gestures from the user to rotate and zoom the sticker) and IMU data to the
|
|
[MatricesManagerCalculator](https://github.com/google/mediapipe/tree/master/mediapipe/graphs/instant_motion_tracking/calculators/matrices_manager_calculator.cc),
|
|
which turns all our sticker transformation data into a set of model matrices.
|
|
This data is handled directly by our
|
|
[GlAnimationOverlayCalculator](https://github.com/google/mediapipe/tree/master/mediapipe/graphs/object_detection_3d/calculators/gl_animation_overlay_calculator.cc)
|
|
as an input stream, which will render the provided texture and object file using
|
|
our matrix specifications. The output of
|
|
[GlAnimationOverlayCalculator](https://github.com/google/mediapipe/tree/master/mediapipe/graphs/object_detection_3d/calculators/gl_animation_overlay_calculator.cc)
|
|
is a video stream depicting the virtual 3D content rendered on top of the real
|
|
world, creating immersive AR experiences for users.
|
|
|
|
## Using Instant Motion Tracking
|
|
|
|
With the Instant Motion Tracking MediaPipe [graph](https://github.com/google/mediapipe/tree/master/mediapipe/graphs/instant_motion_tracking/instant_motion_tracking.pbtxt),
|
|
an application can create an interactive and realistic AR experience by
|
|
specifying the required input streams, side packets, and output streams.
|
|
The input streams are the following:
|
|
|
|
* Input Video (GpuBuffer): Video frames to render augmented stickers onto.
|
|
* Rotation Matrix (9-element Float Array): The 3x3 row-major rotation
|
|
matrix from the device IMU to determine proper orientation of the device.
|
|
* Sticker Proto String (String): A string representing the
|
|
serialized [sticker buffer protobuf message](https://github.com/google/mediapipe/tree/master/mediapipe/graphs/instant_motion_tracking/calculators/sticker_buffer.proto),
|
|
containing a list of all stickers and their attributes.
|
|
* Each sticker in the Protobuffer has a unique ID to find associated
|
|
anchors and transforms, an initial anchor placement in a normalized [0.0, 1.0]
|
|
3D space, a user rotation and user scaling transform on the sticker,
|
|
and an integer indicating which type of objects to render for the
|
|
sticker (e.g. 3D asset or GIF).
|
|
* Sticker Sentinel (Integer): When an anchor must be initially placed or
|
|
repositioned, this value must be changed to the ID of the anchor to reset from
|
|
the sticker buffer protobuf message. If no valid ID is provided, the system
|
|
will simply maintain tracking.
|
|
|
|
Side packets are also an integral part of the Instant Motion Tracking solution
|
|
to provide device-specific information for the rendering system:
|
|
|
|
* Field of View (Float): The field of view of the camera in radians.
|
|
* Aspect Ratio (Float): The aspect ratio (width / height) of the camera frames
|
|
(this ratio corresponds to the image frames themselves, not necessarily the
|
|
screen bounds).
|
|
* Object Asset (String): The
|
|
[GlAnimationOverlayCalculator](https://github.com/google/mediapipe/tree/master/mediapipe/graphs/object_detection_3d/calculators/gl_animation_overlay_calculator.cc)
|
|
must be provided with an associated asset file name pointing to the 3D model
|
|
to render in the viewfinder.
|
|
* (Optional) Texture (ImageFrame on Android, GpuBuffer on iOS): Textures for
|
|
the
|
|
[GlAnimationOverlayCalculator](https://github.com/google/mediapipe/tree/master/mediapipe/graphs/object_detection_3d/calculators/gl_animation_overlay_calculator.cc)
|
|
can be provided either via an input stream (dynamic texturing) or as a side
|
|
packet (unchanging texture).
|
|
|
|
The rendering system for the Instant Motion Tracking is powered by OpenGL. For
|
|
more information regarding the structure of model matrices and OpenGL rendering,
|
|
please visit [OpenGL Wiki](https://www.khronos.org/opengl/wiki/). With the
|
|
specifications above, the Instant Motion Tracking capabilities can be adapted to
|
|
any device that is able to run the MediaPipe framework with a working IMU system
|
|
and connected camera.
|
|
|
|
## Example Apps
|
|
|
|
Please first see general instructions for
|
|
[Android](../getting_started/building_examples.md#android) on how to build
|
|
MediaPipe examples.
|
|
|
|
* Graph: [mediapipe/graphs/instant_motion_tracking/instant_motion_tracking.pbtxt](https://github.com/google/mediapipe/tree/master/mediapipe/graphs/instant_motion_tracking/instant_motion_tracking.pbtxt)
|
|
|
|
* Android target (or download prebuilt [ARM64 APK](https://drive.google.com/file/d/1KnaBBoKpCHR73nOBJ4fL_YdWVTAcwe6L/view?usp=sharing)):
|
|
[`mediapipe/examples/android/src/java/com/google/mediapipe/apps/instantmotiontracking:instantmotiontracking`](https://github.com/google/mediapipe/tree/master/mediapipe/examples/android/src/java/com/google/mediapipe/apps/instantmotiontracking/BUILD)
|
|
|
|
## Resources
|
|
|
|
* Google Developers Blog:
|
|
[Instant Motion Tracking With MediaPipe](https://developers.googleblog.com/2020/08/instant-motion-tracking-with-mediapipe.html)
|
|
* Google AI Blog:
|
|
[The Instant Motion Tracking Behind Motion Stills AR](https://ai.googleblog.com/2018/02/the-instant-motion-tracking-behind.html)
|
|
* Paper:
|
|
[Instant Motion Tracking and Its Applications to Augmented Reality](https://arxiv.org/abs/1907.06796)
|