Merge pull request #4422 from priankakariatyml:ios-gesture-recognizer-files

PiperOrigin-RevId: 533471803
This commit is contained in:
Copybara-Service 2023-05-19 09:19:20 -07:00
commit 9044d62f61
11 changed files with 602 additions and 20 deletions

View File

@ -38,3 +38,20 @@ objc_library(
"//mediapipe/tasks/ios/vision/core:MPPRunningMode", "//mediapipe/tasks/ios/vision/core:MPPRunningMode",
], ],
) )
objc_library(
name = "MPPGestureRecognizer",
hdrs = ["sources/MPPGestureRecognizer.h"],
copts = [
"-ObjC++",
"-std=c++17",
"-x objective-c++",
],
module_name = "MPPGestureRecognizer",
deps = [
":MPPGestureRecognizerOptions",
":MPPGestureRecognizerResult",
"//mediapipe/tasks/ios/core:MPPTaskOptions",
"//mediapipe/tasks/ios/vision/core:MPPImage",
],
)

View File

@ -0,0 +1,165 @@
// Copyright 2023 The MediaPipe Authors.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#import <Foundation/Foundation.h>
#import "mediapipe/tasks/ios/core/sources/MPPTaskOptions.h"
#import "mediapipe/tasks/ios/vision/core/sources/MPPImage.h"
#import "mediapipe/tasks/ios/vision/gesture_recognizer/sources/MPPGestureRecognizerOptions.h"
#import "mediapipe/tasks/ios/vision/gesture_recognizer/sources/MPPGestureRecognizerResult.h"
NS_ASSUME_NONNULL_BEGIN
/**
* @brief Performs gesture recognition on images.
*
* This API expects a pre-trained TFLite hand gesture recognizer model or a custom one created using
* MediaPipe Solutions Model Maker. See
* https://developers.google.com/mediapipe/solutions/model_maker.
*/
NS_SWIFT_NAME(GestureRecognizer)
@interface MPPGestureRecognizer : NSObject
/**
* Creates a new instance of `MPPGestureRecognizer` from an absolute path to a TensorFlow Lite model
* file stored locally on the device and the default `MPPGestureRecognizerOptions`.
*
* @param modelPath An absolute path to a TensorFlow Lite model file stored locally on the device.
* @param error An optional error parameter populated when there is an error in initializing the
* gesture recognizer.
*
* @return A new instance of `MPPGestureRecognizer` with the given model path. `nil` if there is an
* error in initializing the gesture recognizer.
*/
- (nullable instancetype)initWithModelPath:(NSString *)modelPath error:(NSError **)error;
/**
* Creates a new instance of `MPPGestureRecognizer` from the given `MPPGestureRecognizerOptions`.
*
* @param options The options of type `MPPGestureRecognizerOptions` to use for configuring the
* `MPPGestureRecognizer`.
* @param error An optional error parameter populated when there is an error in initializing the
* gesture recognizer.
*
* @return A new instance of `MPPGestureRecognizer` with the given options. `nil` if there is an
* error in initializing the gesture recognizer.
*/
- (nullable instancetype)initWithOptions:(MPPGestureRecognizerOptions *)options
error:(NSError **)error NS_DESIGNATED_INITIALIZER;
/**
* Performs gesture recognition on the provided MPPImage using the whole image as region of
* interest. Rotation will be applied according to the `orientation` property of the provided
* `MPPImage`. Only use this method when the `MPPGestureRecognizer` is created with
* `MPPRunningModeImage`.
* This method supports gesture recognition of RGBA images. If your `MPPImage` has a source type of
* `MPPImageSourceTypePixelBuffer` or `MPPImageSourceTypeSampleBuffer`, the underlying pixel buffer
* must have one of the following pixel format types:
* 1. kCVPixelFormatType_32BGRA
* 2. kCVPixelFormatType_32RGBA
*
* If your `MPPImage` has a source type of `MPPImageSourceTypeImage` ensure that the color space is
* RGB with an Alpha channel.
*
* @param image The `MPPImage` on which gesture recognition is to be performed.
* @param error An optional error parameter populated when there is an error in performing gesture
* recognition on the input image.
*
* @return An `MPPGestureRecognizerResult` object that contains the hand gesture recognition
* results.
*/
- (nullable MPPGestureRecognizerResult *)recognizeImage:(MPPImage *)image
error:(NSError **)error
NS_SWIFT_NAME(recognize(image:));
/**
* Performs gesture recognition on the provided video frame of type `MPPImage` using the whole
* image as region of interest. Rotation will be applied according to the `orientation` property of
* the provided `MPPImage`. Only use this method when the `MPPGestureRecognizer` is created with
* `MPPRunningModeVideo`.
*
* It's required to provide the video frame's timestamp (in milliseconds). The input timestamps must
* be monotonically increasing.
*
* This method supports gesture recognition of RGBA images. If your `MPPImage` has a source type of
* `MPPImageSourceTypePixelBuffer` or `MPPImageSourceTypeSampleBuffer`, the underlying pixel buffer
* must have one of the following pixel format types:
* 1. kCVPixelFormatType_32BGRA
* 2. kCVPixelFormatType_32RGBA
*
* If your `MPPImage` has a source type of `MPPImageSourceTypeImage` ensure that the color space is
* RGB with an Alpha channel.
*
* @param image The `MPPImage` on which gesture recognition is to be performed.
* @param timestampInMilliseconds The video frame's timestamp (in milliseconds). The input
* timestamps must be monotonically increasing.
* @param error An optional error parameter populated when there is an error in performing gesture
* recognition on the input video frame.
*
* @return An `MPPGestureRecognizerResult` object that contains the hand gesture recognition
* results.
*/
- (nullable MPPGestureRecognizerResult *)recognizeVideoFrame:(MPPImage *)image
timestampInMilliseconds:(NSInteger)timestampInMilliseconds
error:(NSError **)error
NS_SWIFT_NAME(recognize(videoFrame:timestampInMilliseconds:));
/**
* Sends live stream image data of type `MPPImage` to perform gesture recognition using the whole
* image as region of interest. Rotation will be applied according to the `orientation` property of
* the provided `MPPImage`. Only use this method when the `MPPGestureRecognizer` is created with
* `MPPRunningModeLiveStream`.
*
* The object which needs to be continuously notified of the available results of gesture
* recognition must confirm to `MPPGestureRecognizerLiveStreamDelegate` protocol and implement the
* `gestureRecognizer:didFinishRecognitionWithResult:timestampInMilliseconds:error:`
* delegate method.
*
* It's required to provide a timestamp (in milliseconds) to indicate when the input image is sent
* to the gesture recognizer. The input timestamps must be monotonically increasing.
*
* This method supports gesture recognition of RGBA images. If your `MPPImage` has a source type of
* `MPPImageSourceTypePixelBuffer` or `MPPImageSourceTypeSampleBuffer`, the underlying pixel buffer
* must have one of the following pixel format types:
* 1. kCVPixelFormatType_32BGRA
* 2. kCVPixelFormatType_32RGBA
*
* If the input `MPPImage` has a source type of `MPPImageSourceTypeImage` ensure that the color
* space is RGB with an Alpha channel.
*
* If this method is used for performing gesture recognition on live camera frames using
* `AVFoundation`, ensure that you request `AVCaptureVideoDataOutput` to output frames in
* `kCMPixelFormat_32RGBA` using its `videoSettings` property.
*
* @param image A live stream image data of type `MPPImage` on which gesture recognition is to be
* performed.
* @param timestampInMilliseconds The timestamp (in milliseconds) which indicates when the input
* image is sent to the gesture recognizer. The input timestamps must be monotonically increasing.
* @param error An optional error parameter populated when there is an error in performing gesture
* recognition on the input live stream image data.
*
* @return `YES` if the image was sent to the task successfully, otherwise `NO`.
*/
- (BOOL)recognizeAsyncImage:(MPPImage *)image
timestampInMilliseconds:(NSInteger)timestampInMilliseconds
error:(NSError **)error
NS_SWIFT_NAME(recognizeAsync(image:timestampInMilliseconds:));
- (instancetype)init NS_UNAVAILABLE;
+ (instancetype)new NS_UNAVAILABLE;
@end
NS_ASSUME_NONNULL_END

View File

@ -21,22 +21,73 @@
NS_ASSUME_NONNULL_BEGIN NS_ASSUME_NONNULL_BEGIN
@class MPPGestureRecognizer;
/**
* This protocol defines an interface for the delegates of `MPPGestureRecognizer` object to receive
* results of performing asynchronous gesture recognition on images (i.e, when `runningMode` =
* `MPPRunningModeLiveStream`).
*
* The delegate of `MPPGestureRecognizer` must adopt `MPPGestureRecognizerLiveStreamDelegate`
* protocol. The methods in this protocol are optional.
*/
NS_SWIFT_NAME(GestureRecognizerLiveStreamDelegate)
@protocol MPPGestureRecognizerLiveStreamDelegate <NSObject>
@optional
/**
* This method notifies a delegate that the results of asynchronous gesture recognition of
* an image submitted to the `MPPGestureRecognizer` is available.
*
* This method is called on a private serial dispatch queue created by the `MPPGestureRecognizer`
* for performing the asynchronous delegates calls.
*
* @param gestureRecognizer The gesture recognizer which performed the gesture recognition.
* This is useful to test equality when there are multiple instances of `MPPGestureRecognizer`.
* @param result The `MPPGestureRecognizerResult` object that contains a list of detections, each
* detection has a bounding box that is expressed in the unrotated input frame of reference
* coordinates system, i.e. in `[0,image_width) x [0,image_height)`, which are the dimensions of the
* underlying image data.
* @param timestampInMilliseconds The timestamp (in milliseconds) which indicates when the input
* image was sent to the gesture recognizer.
* @param error An optional error parameter populated when there is an error in performing gesture
* recognition on the input live stream image data.
*
*/
- (void)gestureRecognizer:(MPPGestureRecognizer *)gestureRecognizer
didFinishRecognitionWithResult:(nullable MPPGestureRecognizerResult *)result
timestampInMilliseconds:(NSInteger)timestampInMilliseconds
error:(nullable NSError *)error
NS_SWIFT_NAME(gestureRecognizer(_:didFinishGestureRecognition:timestampInMilliseconds:error:));
@end
/** Options for setting up a `MPPGestureRecognizer`. */ /** Options for setting up a `MPPGestureRecognizer`. */
NS_SWIFT_NAME(GestureRecognizerOptions) NS_SWIFT_NAME(GestureRecognizerOptions)
@interface MPPGestureRecognizerOptions : MPPTaskOptions <NSCopying> @interface MPPGestureRecognizerOptions : MPPTaskOptions <NSCopying>
/**
* Running mode of the gesture recognizer task. Defaults to `MPPRunningModeImage`.
* `MPPGestureRecognizer` can be created with one of the following running modes:
* 1. `MPPRunningModeImage`: The mode for performing gesture recognition on single image inputs.
* 2. `MPPRunningModeVideo`: The mode for performing gesture recognition on the decoded frames of a
* video.
* 3. `MPPRunningModeLiveStream`: The mode for performing gesture recognition on a live stream of
* input data, such as from the camera.
*/
@property(nonatomic) MPPRunningMode runningMode; @property(nonatomic) MPPRunningMode runningMode;
/** /**
* The user-defined result callback for processing live stream data. The result callback should only * An object that confirms to `MPPGestureRecognizerLiveStreamDelegate` protocol. This object must
* be specified when the running mode is set to the live stream mode. * implement `gestureRecognizer:didFinishRecognitionWithResult:timestampInMilliseconds:error:` to
* TODO: Add parameter `MPPImage` in the callback. * receive the results of performing asynchronous gesture recognition on images (i.e, when
* `runningMode` = `MPPRunningModeLiveStream`).
*/ */
@property(nonatomic, copy) void (^completion) @property(nonatomic, weak, nullable) id<MPPGestureRecognizerLiveStreamDelegate>
(MPPGestureRecognizerResult *result, NSInteger timestampMs, NSError *error); gestureRecognizerLiveStreamDelegate;
/** Sets the maximum number of hands can be detected by the GestureRecognizer. */ /** Sets the maximum number of hands can be detected by the GestureRecognizer. */
@property(nonatomic) NSInteger numHands; @property(nonatomic) NSInteger numberOfHands NS_SWIFT_NAME(numHands);
/** Sets minimum confidence score for the hand detection to be considered successful */ /** Sets minimum confidence score for the hand detection to be considered successful */
@property(nonatomic) float minHandDetectionConfidence; @property(nonatomic) float minHandDetectionConfidence;

View File

@ -19,7 +19,7 @@
- (instancetype)init { - (instancetype)init {
self = [super init]; self = [super init];
if (self) { if (self) {
_numHands = 1; _numberOfHands = 1;
_minHandDetectionConfidence = 0.5f; _minHandDetectionConfidence = 0.5f;
_minHandPresenceConfidence = 0.5f; _minHandPresenceConfidence = 0.5f;
_minTrackingConfidence = 0.5f; _minTrackingConfidence = 0.5f;
@ -31,8 +31,9 @@
MPPGestureRecognizerOptions *gestureRecognizerOptions = [super copyWithZone:zone]; MPPGestureRecognizerOptions *gestureRecognizerOptions = [super copyWithZone:zone];
gestureRecognizerOptions.runningMode = self.runningMode; gestureRecognizerOptions.runningMode = self.runningMode;
gestureRecognizerOptions.completion = self.completion; gestureRecognizerOptions.gestureRecognizerLiveStreamDelegate =
gestureRecognizerOptions.numHands = self.numHands; self.gestureRecognizerLiveStreamDelegate;
gestureRecognizerOptions.numberOfHands = self.numberOfHands;
gestureRecognizerOptions.minHandDetectionConfidence = self.minHandDetectionConfidence; gestureRecognizerOptions.minHandDetectionConfidence = self.minHandDetectionConfidence;
gestureRecognizerOptions.minHandPresenceConfidence = self.minHandPresenceConfidence; gestureRecognizerOptions.minHandPresenceConfidence = self.minHandPresenceConfidence;
gestureRecognizerOptions.minTrackingConfidence = self.minTrackingConfidence; gestureRecognizerOptions.minTrackingConfidence = self.minTrackingConfidence;

View File

@ -25,7 +25,7 @@ NS_SWIFT_NAME(GestureRecognizerResult)
@interface MPPGestureRecognizerResult : MPPTaskResult @interface MPPGestureRecognizerResult : MPPTaskResult
/** Hand landmarks of detected hands. */ /** Hand landmarks of detected hands. */
@property(nonatomic, readonly) NSArray<NSArray<MPPLandmark *> *> *landmarks; @property(nonatomic, readonly) NSArray<NSArray<MPPNormalizedLandmark *> *> *landmarks;
/** Hand landmarks in world coordniates of detected hands. */ /** Hand landmarks in world coordniates of detected hands. */
@property(nonatomic, readonly) NSArray<NSArray<MPPLandmark *> *> *worldLandmarks; @property(nonatomic, readonly) NSArray<NSArray<MPPLandmark *> *> *worldLandmarks;
@ -54,10 +54,10 @@ NS_SWIFT_NAME(GestureRecognizerResult)
* landmarks, handedness and gestures. * landmarks, handedness and gestures.
* *
*/ */
- (instancetype)initWithLandmarks:(NSArray<NSArray<MPPLandmark *> *> *)landmarks - (instancetype)initWithGestures:(NSArray<NSArray<MPPCategory *> *> *)gestures
worldLandmarks:(NSArray<NSArray<MPPLandmark *> *> *)worldLandmarks
handedness:(NSArray<NSArray<MPPCategory *> *> *)handedness handedness:(NSArray<NSArray<MPPCategory *> *> *)handedness
gestures:(NSArray<NSArray<MPPCategory *> *> *)gestures landmarks:(NSArray<NSArray<MPPNormalizedLandmark *> *> *)landmarks
worldLandmarks:(NSArray<NSArray<MPPLandmark *> *> *)worldLandmarks
timestampInMilliseconds:(NSInteger)timestampInMilliseconds; timestampInMilliseconds:(NSInteger)timestampInMilliseconds;
@end @end

View File

@ -16,10 +16,10 @@
@implementation MPPGestureRecognizerResult @implementation MPPGestureRecognizerResult
- (instancetype)initWithLandmarks:(NSArray<NSArray<MPPLandmark *> *> *)landmarks - (instancetype)initWithGestures:(NSArray<NSArray<MPPCategory *> *> *)gestures
worldLandmarks:(NSArray<NSArray<MPPLandmark *> *> *)worldLandmarks
handedness:(NSArray<NSArray<MPPCategory *> *> *)handedness handedness:(NSArray<NSArray<MPPCategory *> *> *)handedness
gestures:(NSArray<NSArray<MPPCategory *> *> *)gestures landmarks:(NSArray<NSArray<MPPNormalizedLandmark *> *> *)landmarks
worldLandmarks:(NSArray<NSArray<MPPLandmark *> *> *)worldLandmarks
timestampInMilliseconds:(NSInteger)timestampInMilliseconds { timestampInMilliseconds:(NSInteger)timestampInMilliseconds {
self = [super initWithTimestampInMilliseconds:timestampInMilliseconds]; self = [super initWithTimestampInMilliseconds:timestampInMilliseconds];
if (self) { if (self) {

View File

@ -0,0 +1,51 @@
# Copyright 2023 The MediaPipe Authors.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
package(default_visibility = ["//mediapipe/tasks:internal"])
licenses(["notice"])
objc_library(
name = "MPPGestureRecognizerOptionsHelpers",
srcs = ["sources/MPPGestureRecognizerOptions+Helpers.mm"],
hdrs = ["sources/MPPGestureRecognizerOptions+Helpers.h"],
deps = [
"//mediapipe/framework:calculator_options_cc_proto",
"//mediapipe/tasks/cc/vision/gesture_recognizer/proto:gesture_classifier_graph_options_cc_proto",
"//mediapipe/tasks/cc/vision/gesture_recognizer/proto:gesture_recognizer_graph_options_cc_proto",
"//mediapipe/tasks/cc/vision/gesture_recognizer/proto:hand_gesture_recognizer_graph_options_cc_proto",
"//mediapipe/tasks/cc/vision/hand_detector/proto:hand_detector_graph_options_cc_proto",
"//mediapipe/tasks/cc/vision/hand_landmarker/proto:hand_landmarker_graph_options_cc_proto",
"//mediapipe/tasks/cc/vision/hand_landmarker/proto:hand_landmarks_detector_graph_options_cc_proto",
"//mediapipe/tasks/ios/common/utils:NSStringHelpers",
"//mediapipe/tasks/ios/components/processors/utils:MPPClassifierOptionsHelpers",
"//mediapipe/tasks/ios/core:MPPTaskOptionsProtocol",
"//mediapipe/tasks/ios/core/utils:MPPBaseOptionsHelpers",
"//mediapipe/tasks/ios/vision/gesture_recognizer:MPPGestureRecognizerOptions",
],
)
objc_library(
name = "MPPGestureRecognizerResultHelpers",
srcs = ["sources/MPPGestureRecognizerResult+Helpers.mm"],
hdrs = ["sources/MPPGestureRecognizerResult+Helpers.h"],
deps = [
"//mediapipe/framework:packet",
"//mediapipe/framework/formats:classification_cc_proto",
"//mediapipe/framework/formats:landmark_cc_proto",
"//mediapipe/tasks/ios/components/containers/utils:MPPCategoryHelpers",
"//mediapipe/tasks/ios/components/containers/utils:MPPLandmarkHelpers",
"//mediapipe/tasks/ios/vision/gesture_recognizer:MPPGestureRecognizerResult",
],
)

View File

@ -0,0 +1,32 @@
// Copyright 2023 The MediaPipe Authors.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#include "mediapipe/framework/calculator_options.pb.h"
#import "mediapipe/tasks/ios/core/sources/MPPTaskOptionsProtocol.h"
#import "mediapipe/tasks/ios/vision/gesture_recognizer/sources/MPPGestureRecognizerOptions.h"
NS_ASSUME_NONNULL_BEGIN
@interface MPPGestureRecognizerOptions (Helpers) <MPPTaskOptionsProtocol>
/**
* Populates the provided `CalculatorOptions` proto container with the current settings.
*
* @param optionsProto The `CalculatorOptions` proto object to copy the settings to.
*/
- (void)copyToProto:(::mediapipe::CalculatorOptions *)optionsProto;
@end
NS_ASSUME_NONNULL_END

View File

@ -0,0 +1,92 @@
// Copyright 2023 The MediaPipe Authors.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#import "mediapipe/tasks/ios/vision/gesture_recognizer/utils/sources/MPPGestureRecognizerOptions+Helpers.h"
#import "mediapipe/tasks/ios/common/utils/sources/NSString+Helpers.h"
#import "mediapipe/tasks/ios/components/processors/utils/sources/MPPClassifierOptions+Helpers.h"
#import "mediapipe/tasks/ios/core/utils/sources/MPPBaseOptions+Helpers.h"
#include "mediapipe/tasks/cc/vision/gesture_recognizer/proto/gesture_classifier_graph_options.pb.h"
#include "mediapipe/tasks/cc/vision/gesture_recognizer/proto/gesture_recognizer_graph_options.pb.h"
#include "mediapipe/tasks/cc/vision/gesture_recognizer/proto/hand_gesture_recognizer_graph_options.pb.h"
#include "mediapipe/tasks/cc/vision/hand_detector/proto/hand_detector_graph_options.pb.h"
#include "mediapipe/tasks/cc/vision/hand_landmarker/proto/hand_landmarker_graph_options.pb.h"
#include "mediapipe/tasks/cc/vision/hand_landmarker/proto/hand_landmarks_detector_graph_options.pb.h"
namespace {
using CalculatorOptionsProto = mediapipe::CalculatorOptions;
using GestureClassifierGraphOptionsProto =
::mediapipe::tasks::vision::gesture_recognizer::proto::GestureClassifierGraphOptions;
using GestureRecognizerGraphOptionsProto =
::mediapipe::tasks::vision::gesture_recognizer::proto::GestureRecognizerGraphOptions;
using HandGestureRecognizerGraphOptionsProto =
::mediapipe::tasks::vision::gesture_recognizer::proto::HandGestureRecognizerGraphOptions;
using HandLandmarkerGraphOptionsProto =
::mediapipe::tasks::vision::hand_landmarker::proto::HandLandmarkerGraphOptions;
using HandDetectorGraphOptionsProto =
::mediapipe::tasks::vision::hand_detector::proto::HandDetectorGraphOptions;
using HandLandmarksDetectorGraphOptionsProto =
::mediapipe::tasks::vision::hand_landmarker::proto::HandLandmarksDetectorGraphOptions;
using ClassifierOptionsProto = ::mediapipe::tasks::components::processors::proto::ClassifierOptions;
} // namespace
@implementation MPPGestureRecognizerOptions (Helpers)
- (void)copyToProto:(CalculatorOptionsProto *)optionsProto {
GestureRecognizerGraphOptionsProto *gestureRecognizerGraphOptionsProto =
optionsProto->MutableExtension(GestureRecognizerGraphOptionsProto::ext);
gestureRecognizerGraphOptionsProto->Clear();
[self.baseOptions copyToProto:gestureRecognizerGraphOptionsProto->mutable_base_options()
withUseStreamMode:self.runningMode != MPPRunningModeImage];
HandLandmarkerGraphOptionsProto *handLandmarkerGraphOptionsProto =
gestureRecognizerGraphOptionsProto->mutable_hand_landmarker_graph_options();
handLandmarkerGraphOptionsProto->Clear();
handLandmarkerGraphOptionsProto->set_min_tracking_confidence(self.minTrackingConfidence);
HandDetectorGraphOptionsProto *handDetectorGraphOptionsProto =
handLandmarkerGraphOptionsProto->mutable_hand_detector_graph_options();
handDetectorGraphOptionsProto->Clear();
handDetectorGraphOptionsProto->set_num_hands(self.numberOfHands);
handDetectorGraphOptionsProto->set_min_detection_confidence(self.minHandDetectionConfidence);
HandLandmarksDetectorGraphOptionsProto *handLandmarksDetectorGraphOptionsProto =
handLandmarkerGraphOptionsProto->mutable_hand_landmarks_detector_graph_options();
handLandmarksDetectorGraphOptionsProto->Clear();
handLandmarksDetectorGraphOptionsProto->set_min_detection_confidence(
self.minHandPresenceConfidence);
HandGestureRecognizerGraphOptionsProto *handGestureRecognizerGraphOptionsProto =
gestureRecognizerGraphOptionsProto->mutable_hand_gesture_recognizer_graph_options();
if (self.cannedGesturesClassifierOptions) {
GestureClassifierGraphOptionsProto *cannedGesturesClassifierOptionsProto =
handGestureRecognizerGraphOptionsProto->mutable_canned_gesture_classifier_graph_options();
cannedGesturesClassifierOptionsProto->Clear();
[self.cannedGesturesClassifierOptions
copyToProto:cannedGesturesClassifierOptionsProto->mutable_classifier_options()];
}
if (self.customGesturesClassifierOptions) {
GestureClassifierGraphOptionsProto *customGesturesClassifierOptionsProto =
handGestureRecognizerGraphOptionsProto->mutable_custom_gesture_classifier_graph_options();
customGesturesClassifierOptionsProto->Clear();
[self.customGesturesClassifierOptions
copyToProto:customGesturesClassifierOptionsProto->mutable_classifier_options()];
}
}
@end

View File

@ -0,0 +1,46 @@
// Copyright 2023 The MediaPipe Authors.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#import "mediapipe/tasks/ios/vision/gesture_recognizer/sources/MPPGestureRecognizerResult.h"
#include "mediapipe/framework/packet.h"
NS_ASSUME_NONNULL_BEGIN
static const int kMicroSecondsPerMilliSecond = 1000;
@interface MPPGestureRecognizerResult (Helpers)
/**
* Creates an `MPPGestureRecognizerResult` from hand gestures, handedness, hand landmarks and world
* landmarks packets.
*
* @param handGesturesPacket a MediaPipe packet wrapping a`std::vector<ClassificationListProto>`.
* @param handednessPacket a MediaPipe packet wrapping a`std::vector<ClassificationListProto>`.
* @param handLandmarksPacket a MediaPipe packet wrapping
* a`std::vector<NormalizedlandmarkListProto>`.
* @param handLandmarksPacket a MediaPipe packet wrapping a`std::vector<LandmarkListProto>`.
*
* @return An `MPPGestureRecognizerResult` object that contains the hand gesture recognition
* results.
*/
+ (MPPGestureRecognizerResult *)
gestureRecognizerResultWithHandGesturesPacket:(const mediapipe::Packet &)handGesturesPacket
handednessPacket:(const mediapipe::Packet &)handednessPacket
handLandmarksPacket:(const mediapipe::Packet &)handLandmarksPacket
worldLandmarksPacket:(const mediapipe::Packet &)worldLandmarksPacket;
@end
NS_ASSUME_NONNULL_END

View File

@ -0,0 +1,127 @@
// Copyright 2023 The MediaPipe Authors.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#import "mediapipe/tasks/ios/vision/gesture_recognizer/utils/sources/MPPGestureRecognizerResult+Helpers.h"
#import "mediapipe/tasks/ios/components/containers/utils/sources/MPPCategory+Helpers.h"
#import "mediapipe/tasks/ios/components/containers/utils/sources/MPPLandmark+Helpers.h"
#include "mediapipe/framework/formats/classification.pb.h"
#include "mediapipe/framework/formats/landmark.pb.h"
#include "mediapipe/framework/packet.h"
namespace {
using ClassificationListProto = ::mediapipe::ClassificationList;
using LandmarkListProto = ::mediapipe::LandmarkList;
using NormalizedLandmarkListProto = ::mediapipe::NormalizedLandmarkList;
using ::mediapipe::Packet;
} // namespace
@implementation MPPGestureRecognizerResult (Helpers)
+ (MPPGestureRecognizerResult *)
gestureRecognizerResultWithHandGesturesPacket:(const Packet &)handGesturesPacket
handednessPacket:(const Packet &)handednessPacket
handLandmarksPacket:(const Packet &)handLandmarksPacket
worldLandmarksPacket:(const Packet &)worldLandmarksPacket {
NSInteger timestampInMilliseconds =
(NSInteger)(handGesturesPacket.Timestamp().Value() / kMicroSecondsPerMilliSecond);
if (handGesturesPacket.IsEmpty()) {
return [[MPPGestureRecognizerResult alloc] initWithGestures:@[]
handedness:@[]
landmarks:@[]
worldLandmarks:@[]
timestampInMilliseconds:timestampInMilliseconds];
}
if (!handGesturesPacket.ValidateAsType<std::vector<ClassificationListProto>>().ok() ||
!handednessPacket.ValidateAsType<std::vector<ClassificationListProto>>().ok() ||
!handLandmarksPacket.ValidateAsType<std::vector<NormalizedLandmarkListProto>>().ok() ||
!worldLandmarksPacket.ValidateAsType<std::vector<LandmarkListProto>>().ok()) {
return nil;
}
const std::vector<ClassificationListProto> &handGesturesClassificationListProtos =
handGesturesPacket.Get<std::vector<ClassificationListProto>>();
NSMutableArray<NSMutableArray<MPPCategory *> *> *multiHandGestures =
[NSMutableArray arrayWithCapacity:(NSUInteger)handGesturesClassificationListProtos.size()];
for (const auto &classificationListProto : handGesturesClassificationListProtos) {
NSMutableArray<MPPCategory *> *gestures = [NSMutableArray
arrayWithCapacity:(NSUInteger)classificationListProto.classification().size()];
for (const auto &classificationProto : classificationListProto.classification()) {
MPPCategory *category = [MPPCategory categoryWithProto:classificationProto];
[gestures addObject:category];
}
[multiHandGestures addObject:gestures];
}
const std::vector<ClassificationListProto> &handednessClassificationListProtos =
handednessPacket.Get<std::vector<ClassificationListProto>>();
NSMutableArray<NSMutableArray<MPPCategory *> *> *multiHandHandedness =
[NSMutableArray arrayWithCapacity:(NSUInteger)handednessClassificationListProtos.size()];
for (const auto &classificationListProto : handednessClassificationListProtos) {
NSMutableArray<MPPCategory *> *handedness = [NSMutableArray
arrayWithCapacity:(NSUInteger)classificationListProto.classification().size()];
for (const auto &classificationProto : classificationListProto.classification()) {
MPPCategory *category = [MPPCategory categoryWithProto:classificationProto];
[handedness addObject:category];
}
[multiHandHandedness addObject:handedness];
}
const std::vector<NormalizedLandmarkListProto> &handLandmarkListProtos =
handLandmarksPacket.Get<std::vector<NormalizedLandmarkListProto>>();
NSMutableArray<NSMutableArray<MPPNormalizedLandmark *> *> *multiHandLandmarks =
[NSMutableArray arrayWithCapacity:(NSUInteger)handLandmarkListProtos.size()];
for (const auto &handLandmarkListProto : handLandmarkListProtos) {
NSMutableArray<MPPNormalizedLandmark *> *handLandmarks =
[NSMutableArray arrayWithCapacity:(NSUInteger)handLandmarkListProto.landmark().size()];
for (const auto &normalizedLandmarkProto : handLandmarkListProto.landmark()) {
MPPNormalizedLandmark *normalizedLandmark =
[MPPNormalizedLandmark normalizedLandmarkWithProto:normalizedLandmarkProto];
[handLandmarks addObject:normalizedLandmark];
}
[multiHandLandmarks addObject:handLandmarks];
}
const std::vector<LandmarkListProto> &worldLandmarkListProtos =
worldLandmarksPacket.Get<std::vector<LandmarkListProto>>();
NSMutableArray<NSMutableArray<MPPLandmark *> *> *multiHandWorldLandmarks =
[NSMutableArray arrayWithCapacity:(NSUInteger)worldLandmarkListProtos.size()];
for (const auto &worldLandmarkListProto : worldLandmarkListProtos) {
NSMutableArray<MPPLandmark *> *worldLandmarks =
[NSMutableArray arrayWithCapacity:(NSUInteger)worldLandmarkListProto.landmark().size()];
for (const auto &landmarkProto : worldLandmarkListProto.landmark()) {
MPPLandmark *landmark = [MPPLandmark landmarkWithProto:landmarkProto];
[worldLandmarks addObject:landmark];
}
[multiHandWorldLandmarks addObject:worldLandmarks];
}
MPPGestureRecognizerResult *gestureRecognizerResult =
[[MPPGestureRecognizerResult alloc] initWithGestures:multiHandGestures
handedness:multiHandHandedness
landmarks:multiHandLandmarks
worldLandmarks:multiHandWorldLandmarks
timestampInMilliseconds:timestampInMilliseconds];
return gestureRecognizerResult;
}
@end