release facemesh sdk ios
This commit is contained in:
parent
156038543c
commit
9dac04ca46
|
@ -375,6 +375,14 @@ http_archive(
|
||||||
url = "https://github.com/opencv/opencv/releases/download/3.2.0/opencv-3.2.0-ios-framework.zip",
|
url = "https://github.com/opencv/opencv/releases/download/3.2.0/opencv-3.2.0-ios-framework.zip",
|
||||||
)
|
)
|
||||||
|
|
||||||
|
# http_archive(
|
||||||
|
# name = "ios_opencv",
|
||||||
|
# sha256 = "c23e92c4a0cd343f73d4056e66c961cdf68e73ca699b1129638537a931c4ffc8",
|
||||||
|
# build_file = "@//third_party:opencv_ios.BUILD",
|
||||||
|
# type = "zip",
|
||||||
|
# url = "https://github.com/opencv/opencv/releases/download/4.7.0/opencv-4.7.0-ios-framework.zip",
|
||||||
|
# )
|
||||||
|
|
||||||
# Building an opencv.xcframework from the OpenCV 4.5.3 sources is necessary for
|
# Building an opencv.xcframework from the OpenCV 4.5.3 sources is necessary for
|
||||||
# MediaPipe iOS Task Libraries to be supported on arm64(M1) Macs. An
|
# MediaPipe iOS Task Libraries to be supported on arm64(M1) Macs. An
|
||||||
# `opencv.xcframework` archive has not been released and it is recommended to
|
# `opencv.xcframework` archive has not been released and it is recommended to
|
||||||
|
|
|
@ -1,5 +1,7 @@
|
||||||
#!/bin/sh
|
#!/bin/sh
|
||||||
|
|
||||||
|
sudo rm -rf frameworkbuild
|
||||||
|
|
||||||
# Create output directories~
|
# Create output directories~
|
||||||
mkdir -p ./frameworkbuild/FaceMeshSDK/arm64
|
mkdir -p ./frameworkbuild/FaceMeshSDK/arm64
|
||||||
# XCFramework is how we're going to use it.
|
# XCFramework is how we're going to use it.
|
||||||
|
|
BIN
frameworkbuild/FaceMeshSDK/arm64/FaceMeshSDK.framework/FaceMeshSDK
Executable file
BIN
frameworkbuild/FaceMeshSDK/arm64/FaceMeshSDK.framework/FaceMeshSDK
Executable file
Binary file not shown.
|
@ -0,0 +1,58 @@
|
||||||
|
#import <CoreVideo/CoreVideo.h>
|
||||||
|
#import <Foundation/Foundation.h>
|
||||||
|
|
||||||
|
@interface IntPoint : NSObject
|
||||||
|
@property(nonatomic) NSInteger x;
|
||||||
|
@property(nonatomic) NSInteger y;
|
||||||
|
|
||||||
|
- (instancetype)initWithX:(NSInteger)x y:(NSInteger)y;
|
||||||
|
@end
|
||||||
|
|
||||||
|
@interface NSValue (IntPoint)
|
||||||
|
+ (instancetype)valuewithIntPoint:(IntPoint *)value;
|
||||||
|
@property (readonly) IntPoint* intPointValue;
|
||||||
|
@end
|
||||||
|
|
||||||
|
@interface FaceMeshLandmarkPoint : NSObject
|
||||||
|
@property(nonatomic) float x;
|
||||||
|
@property(nonatomic) float y;
|
||||||
|
@property(nonatomic) float z;
|
||||||
|
@end
|
||||||
|
|
||||||
|
@interface FaceMeshNormalizedRect : NSObject
|
||||||
|
@property(nonatomic) float centerX;
|
||||||
|
@property(nonatomic) float centerY;
|
||||||
|
@property(nonatomic) float height;
|
||||||
|
@property(nonatomic) float width;
|
||||||
|
@property(nonatomic) float rotation;
|
||||||
|
@end
|
||||||
|
|
||||||
|
@protocol FaceMeshDelegate <NSObject>
|
||||||
|
@optional
|
||||||
|
/**
|
||||||
|
* Array of faces, with faces represented by arrays of face landmarks
|
||||||
|
*/
|
||||||
|
- (void)didReceiveFaces:(NSArray<NSArray<FaceMeshLandmarkPoint *> *> *)faces;
|
||||||
|
|
||||||
|
- (void)didSavedRegions:(NSArray<NSURL *> *)foreheadURLs
|
||||||
|
leftcheekURLs:(NSArray<NSURL *> *)leftcheekURLs
|
||||||
|
rightcheekURLs:(NSArray<NSURL *> *)rightcheekURLs;
|
||||||
|
@end
|
||||||
|
|
||||||
|
@interface FaceMesh : NSObject
|
||||||
|
- (instancetype)init;
|
||||||
|
- (void)startGraph;
|
||||||
|
- (void)processVideoFrame:(CVPixelBufferRef)imageBuffer;
|
||||||
|
- (CVPixelBufferRef)resize:(CVPixelBufferRef)pixelBuffer
|
||||||
|
width:(int)width
|
||||||
|
height:(int)height;
|
||||||
|
- (uint8_t **)buffer2Array2D:(CVPixelBufferRef)pixelBuffer;
|
||||||
|
- (void)extractRegions:(NSURL *)fileName
|
||||||
|
foreheadBoxes:(NSArray<NSArray<IntPoint *> *> *)foreheadBoxes
|
||||||
|
leftCheekBoxes:(NSArray<NSArray<IntPoint *> *> *)leftCheekBoxes
|
||||||
|
rightCheekBoxes:(NSArray<NSArray<IntPoint *> *> *)rightCheekBoxes
|
||||||
|
totalFramesNeedProcess:(NSInteger)totalFramesNeedProcess
|
||||||
|
skipNFirstFrames:(NSInteger)skipNFirstFrames;
|
||||||
|
@property(weak, nonatomic) id<FaceMeshDelegate> delegate;
|
||||||
|
@property(nonatomic) size_t timestamp;
|
||||||
|
@end
|
|
@ -0,0 +1,16 @@
|
||||||
|
//
|
||||||
|
// FaceMeshSDK.h
|
||||||
|
// FaceMeshSDK
|
||||||
|
//
|
||||||
|
|
||||||
|
#import <Foundation/Foundation.h>
|
||||||
|
|
||||||
|
//! Project version number for FaceMeshSDK.
|
||||||
|
FOUNDATION_EXPORT double FaceMeshSDKVersionNumber;
|
||||||
|
|
||||||
|
//! Project version string for FaceMeshSDK.
|
||||||
|
FOUNDATION_EXPORT const unsigned char FaceMeshSDKVersionString[];
|
||||||
|
|
||||||
|
// In this header, you should import all the public headers of your framework using statements like #import <FaceMeshSDK/PublicHeader.h>
|
||||||
|
|
||||||
|
#import "FaceMesh.h"
|
BIN
frameworkbuild/FaceMeshSDK/arm64/FaceMeshSDK.framework/Info.plist
Executable file
BIN
frameworkbuild/FaceMeshSDK/arm64/FaceMeshSDK.framework/Info.plist
Executable file
Binary file not shown.
|
@ -0,0 +1,21 @@
|
||||||
|
framework module FaceMeshSDK {
|
||||||
|
umbrella header "FaceMeshSDK.h"
|
||||||
|
|
||||||
|
export *
|
||||||
|
module * { export * }
|
||||||
|
|
||||||
|
link framework "AVFoundation"
|
||||||
|
link framework "Accelerate"
|
||||||
|
link framework "AssetsLibrary"
|
||||||
|
link framework "CoreFoundation"
|
||||||
|
link framework "CoreGraphics"
|
||||||
|
link framework "CoreImage"
|
||||||
|
link framework "CoreMedia"
|
||||||
|
link framework "CoreVideo"
|
||||||
|
link framework "GLKit"
|
||||||
|
link framework "Metal"
|
||||||
|
link framework "MetalKit"
|
||||||
|
link framework "OpenGLES"
|
||||||
|
link framework "QuartzCore"
|
||||||
|
link framework "UIKit"
|
||||||
|
}
|
Binary file not shown.
BIN
frameworkbuild/FaceMeshSDK/arm64/FaceMeshSDK.framework/face_landmark.tflite
Executable file
BIN
frameworkbuild/FaceMeshSDK/arm64/FaceMeshSDK.framework/face_landmark.tflite
Executable file
Binary file not shown.
Binary file not shown.
|
@ -0,0 +1,5 @@
|
||||||
|
|
||||||
|
€ConstantSidePacketCalculator2PACKET:with_attentionBI
|
||||||
|
Atype.googleapis.com/mediapipe.ConstantSidePacketCalculatorOptions
|
||||||
|
|
||||||
|
¬FaceLandmarkFrontGpuIMAGE:input_video"LANDMARKS:multi_face_landmarks"-ROIS_FROM_LANDMARKS:face_rects_from_landmarks*NUM_FACES:num_faces*WITH_ATTENTION:with_attentionRinput_videozmulti_face_landmarks‚ num_faces
|
|
@ -0,0 +1,3 @@
|
||||||
|
|
||||||
|
<EFBFBD>FaceLandmarkFrontGpuIMAGE:input_video"LANDMARKS:multi_face_landmarks"-ROIS_FROM_LANDMARKS:face_rects_from_landmarks*NUM_FACES:num_faces
|
||||||
|
SPacketPresenceCalculatorPACKET:multi_face_landmarks"PRESENCE:landmark_presenceRinput_videozmulti_face_landmarkszface_rects_from_landmarkszlandmark_presence‚ num_faces
|
|
@ -0,0 +1,25 @@
|
||||||
|
<?xml version="1.0" encoding="UTF-8"?>
|
||||||
|
<!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd">
|
||||||
|
<plist version="1.0">
|
||||||
|
<dict>
|
||||||
|
<key>AvailableLibraries</key>
|
||||||
|
<array>
|
||||||
|
<dict>
|
||||||
|
<key>LibraryIdentifier</key>
|
||||||
|
<string>ios-arm64</string>
|
||||||
|
<key>LibraryPath</key>
|
||||||
|
<string>FaceMeshSDK.framework</string>
|
||||||
|
<key>SupportedArchitectures</key>
|
||||||
|
<array>
|
||||||
|
<string>arm64</string>
|
||||||
|
</array>
|
||||||
|
<key>SupportedPlatform</key>
|
||||||
|
<string>ios</string>
|
||||||
|
</dict>
|
||||||
|
</array>
|
||||||
|
<key>CFBundlePackageType</key>
|
||||||
|
<string>XFWK</string>
|
||||||
|
<key>XCFrameworkFormatVersion</key>
|
||||||
|
<string>1.0</string>
|
||||||
|
</dict>
|
||||||
|
</plist>
|
Binary file not shown.
|
@ -0,0 +1,58 @@
|
||||||
|
#import <CoreVideo/CoreVideo.h>
|
||||||
|
#import <Foundation/Foundation.h>
|
||||||
|
|
||||||
|
@interface IntPoint : NSObject
|
||||||
|
@property(nonatomic) NSInteger x;
|
||||||
|
@property(nonatomic) NSInteger y;
|
||||||
|
|
||||||
|
- (instancetype)initWithX:(NSInteger)x y:(NSInteger)y;
|
||||||
|
@end
|
||||||
|
|
||||||
|
@interface NSValue (IntPoint)
|
||||||
|
+ (instancetype)valuewithIntPoint:(IntPoint *)value;
|
||||||
|
@property (readonly) IntPoint* intPointValue;
|
||||||
|
@end
|
||||||
|
|
||||||
|
@interface FaceMeshLandmarkPoint : NSObject
|
||||||
|
@property(nonatomic) float x;
|
||||||
|
@property(nonatomic) float y;
|
||||||
|
@property(nonatomic) float z;
|
||||||
|
@end
|
||||||
|
|
||||||
|
@interface FaceMeshNormalizedRect : NSObject
|
||||||
|
@property(nonatomic) float centerX;
|
||||||
|
@property(nonatomic) float centerY;
|
||||||
|
@property(nonatomic) float height;
|
||||||
|
@property(nonatomic) float width;
|
||||||
|
@property(nonatomic) float rotation;
|
||||||
|
@end
|
||||||
|
|
||||||
|
@protocol FaceMeshDelegate <NSObject>
|
||||||
|
@optional
|
||||||
|
/**
|
||||||
|
* Array of faces, with faces represented by arrays of face landmarks
|
||||||
|
*/
|
||||||
|
- (void)didReceiveFaces:(NSArray<NSArray<FaceMeshLandmarkPoint *> *> *)faces;
|
||||||
|
|
||||||
|
- (void)didSavedRegions:(NSArray<NSURL *> *)foreheadURLs
|
||||||
|
leftcheekURLs:(NSArray<NSURL *> *)leftcheekURLs
|
||||||
|
rightcheekURLs:(NSArray<NSURL *> *)rightcheekURLs;
|
||||||
|
@end
|
||||||
|
|
||||||
|
@interface FaceMesh : NSObject
|
||||||
|
- (instancetype)init;
|
||||||
|
- (void)startGraph;
|
||||||
|
- (void)processVideoFrame:(CVPixelBufferRef)imageBuffer;
|
||||||
|
- (CVPixelBufferRef)resize:(CVPixelBufferRef)pixelBuffer
|
||||||
|
width:(int)width
|
||||||
|
height:(int)height;
|
||||||
|
- (uint8_t **)buffer2Array2D:(CVPixelBufferRef)pixelBuffer;
|
||||||
|
- (void)extractRegions:(NSURL *)fileName
|
||||||
|
foreheadBoxes:(NSArray<NSArray<IntPoint *> *> *)foreheadBoxes
|
||||||
|
leftCheekBoxes:(NSArray<NSArray<IntPoint *> *> *)leftCheekBoxes
|
||||||
|
rightCheekBoxes:(NSArray<NSArray<IntPoint *> *> *)rightCheekBoxes
|
||||||
|
totalFramesNeedProcess:(NSInteger)totalFramesNeedProcess
|
||||||
|
skipNFirstFrames:(NSInteger)skipNFirstFrames;
|
||||||
|
@property(weak, nonatomic) id<FaceMeshDelegate> delegate;
|
||||||
|
@property(nonatomic) size_t timestamp;
|
||||||
|
@end
|
|
@ -0,0 +1,16 @@
|
||||||
|
//
|
||||||
|
// FaceMeshSDK.h
|
||||||
|
// FaceMeshSDK
|
||||||
|
//
|
||||||
|
|
||||||
|
#import <Foundation/Foundation.h>
|
||||||
|
|
||||||
|
//! Project version number for FaceMeshSDK.
|
||||||
|
FOUNDATION_EXPORT double FaceMeshSDKVersionNumber;
|
||||||
|
|
||||||
|
//! Project version string for FaceMeshSDK.
|
||||||
|
FOUNDATION_EXPORT const unsigned char FaceMeshSDKVersionString[];
|
||||||
|
|
||||||
|
// In this header, you should import all the public headers of your framework using statements like #import <FaceMeshSDK/PublicHeader.h>
|
||||||
|
|
||||||
|
#import "FaceMesh.h"
|
Binary file not shown.
|
@ -0,0 +1,21 @@
|
||||||
|
framework module FaceMeshSDK {
|
||||||
|
umbrella header "FaceMeshSDK.h"
|
||||||
|
|
||||||
|
export *
|
||||||
|
module * { export * }
|
||||||
|
|
||||||
|
link framework "AVFoundation"
|
||||||
|
link framework "Accelerate"
|
||||||
|
link framework "AssetsLibrary"
|
||||||
|
link framework "CoreFoundation"
|
||||||
|
link framework "CoreGraphics"
|
||||||
|
link framework "CoreImage"
|
||||||
|
link framework "CoreMedia"
|
||||||
|
link framework "CoreVideo"
|
||||||
|
link framework "GLKit"
|
||||||
|
link framework "Metal"
|
||||||
|
link framework "MetalKit"
|
||||||
|
link framework "OpenGLES"
|
||||||
|
link framework "QuartzCore"
|
||||||
|
link framework "UIKit"
|
||||||
|
}
|
Binary file not shown.
Binary file not shown.
Binary file not shown.
|
@ -0,0 +1,5 @@
|
||||||
|
|
||||||
|
€ConstantSidePacketCalculator2PACKET:with_attentionBI
|
||||||
|
Atype.googleapis.com/mediapipe.ConstantSidePacketCalculatorOptions
|
||||||
|
|
||||||
|
¬FaceLandmarkFrontGpuIMAGE:input_video"LANDMARKS:multi_face_landmarks"-ROIS_FROM_LANDMARKS:face_rects_from_landmarks*NUM_FACES:num_faces*WITH_ATTENTION:with_attentionRinput_videozmulti_face_landmarks‚ num_faces
|
|
@ -0,0 +1,3 @@
|
||||||
|
|
||||||
|
<EFBFBD>FaceLandmarkFrontGpuIMAGE:input_video"LANDMARKS:multi_face_landmarks"-ROIS_FROM_LANDMARKS:face_rects_from_landmarks*NUM_FACES:num_faces
|
||||||
|
SPacketPresenceCalculatorPACKET:multi_face_landmarks"PRESENCE:landmark_presenceRinput_videozmulti_face_landmarkszface_rects_from_landmarkszlandmark_presence‚ num_faces
|
BIN
mediapipe/.DS_Store
vendored
Normal file
BIN
mediapipe/.DS_Store
vendored
Normal file
Binary file not shown.
BIN
mediapipe/examples/.DS_Store
vendored
Normal file
BIN
mediapipe/examples/.DS_Store
vendored
Normal file
Binary file not shown.
BIN
mediapipe/examples/ios/.DS_Store
vendored
Normal file
BIN
mediapipe/examples/ios/.DS_Store
vendored
Normal file
Binary file not shown.
BIN
mediapipe/examples/ios/facemeshioslib/.DS_Store
vendored
Normal file
BIN
mediapipe/examples/ios/facemeshioslib/.DS_Store
vendored
Normal file
Binary file not shown.
|
@ -56,6 +56,8 @@ objc_library(
|
||||||
"//mediapipe/objc:mediapipe_framework_ios",
|
"//mediapipe/objc:mediapipe_framework_ios",
|
||||||
"//mediapipe/objc:mediapipe_input_sources_ios",
|
"//mediapipe/objc:mediapipe_input_sources_ios",
|
||||||
"//mediapipe/calculators/core:packet_presence_calculator",
|
"//mediapipe/calculators/core:packet_presence_calculator",
|
||||||
|
"//mediapipe/framework/port:opencv_video",
|
||||||
|
"//mediapipe/framework/port:opencv_imgcodecs",
|
||||||
# "//mediapipe/objc:mediapipe_layer_renderer", # no need for layer renderer since I don't render
|
# "//mediapipe/objc:mediapipe_layer_renderer", # no need for layer renderer since I don't render
|
||||||
] + select({
|
] + select({
|
||||||
"//conditions:default": [
|
"//conditions:default": [
|
||||||
|
|
|
@ -1,6 +1,18 @@
|
||||||
#import <CoreVideo/CoreVideo.h>
|
#import <CoreVideo/CoreVideo.h>
|
||||||
#import <Foundation/Foundation.h>
|
#import <Foundation/Foundation.h>
|
||||||
|
|
||||||
|
@interface IntPoint : NSObject
|
||||||
|
@property(nonatomic) NSInteger x;
|
||||||
|
@property(nonatomic) NSInteger y;
|
||||||
|
|
||||||
|
- (instancetype)initWithX:(NSInteger)x y:(NSInteger)y;
|
||||||
|
@end
|
||||||
|
|
||||||
|
@interface NSValue (IntPoint)
|
||||||
|
+ (instancetype)valuewithIntPoint:(IntPoint *)value;
|
||||||
|
@property (readonly) IntPoint* intPointValue;
|
||||||
|
@end
|
||||||
|
|
||||||
@interface FaceMeshLandmarkPoint : NSObject
|
@interface FaceMeshLandmarkPoint : NSObject
|
||||||
@property(nonatomic) float x;
|
@property(nonatomic) float x;
|
||||||
@property(nonatomic) float y;
|
@property(nonatomic) float y;
|
||||||
|
@ -21,6 +33,10 @@
|
||||||
* Array of faces, with faces represented by arrays of face landmarks
|
* Array of faces, with faces represented by arrays of face landmarks
|
||||||
*/
|
*/
|
||||||
- (void)didReceiveFaces:(NSArray<NSArray<FaceMeshLandmarkPoint *> *> *)faces;
|
- (void)didReceiveFaces:(NSArray<NSArray<FaceMeshLandmarkPoint *> *> *)faces;
|
||||||
|
|
||||||
|
- (void)didSavedRegions:(NSArray<NSURL *> *)foreheadURLs
|
||||||
|
leftcheekURLs:(NSArray<NSURL *> *)leftcheekURLs
|
||||||
|
rightcheekURLs:(NSArray<NSURL *> *)rightcheekURLs;
|
||||||
@end
|
@end
|
||||||
|
|
||||||
@interface FaceMesh : NSObject
|
@interface FaceMesh : NSObject
|
||||||
|
@ -30,6 +46,13 @@
|
||||||
- (CVPixelBufferRef)resize:(CVPixelBufferRef)pixelBuffer
|
- (CVPixelBufferRef)resize:(CVPixelBufferRef)pixelBuffer
|
||||||
width:(int)width
|
width:(int)width
|
||||||
height:(int)height;
|
height:(int)height;
|
||||||
|
- (uint8_t **)buffer2Array2D:(CVPixelBufferRef)pixelBuffer;
|
||||||
|
- (void)extractRegions:(NSURL *)fileName
|
||||||
|
foreheadBoxes:(NSArray<NSArray<IntPoint *> *> *)foreheadBoxes
|
||||||
|
leftCheekBoxes:(NSArray<NSArray<IntPoint *> *> *)leftCheekBoxes
|
||||||
|
rightCheekBoxes:(NSArray<NSArray<IntPoint *> *> *)rightCheekBoxes
|
||||||
|
totalFramesNeedProcess:(NSInteger)totalFramesNeedProcess
|
||||||
|
skipNFirstFrames:(NSInteger)skipNFirstFrames;
|
||||||
@property(weak, nonatomic) id<FaceMeshDelegate> delegate;
|
@property(weak, nonatomic) id<FaceMeshDelegate> delegate;
|
||||||
@property(nonatomic) size_t timestamp;
|
@property(nonatomic) size_t timestamp;
|
||||||
@end
|
@end
|
|
@ -8,11 +8,17 @@
|
||||||
|
|
||||||
#include "mediapipe/framework/port/opencv_core_inc.h"
|
#include "mediapipe/framework/port/opencv_core_inc.h"
|
||||||
#include "mediapipe/framework/port/opencv_imgproc_inc.h"
|
#include "mediapipe/framework/port/opencv_imgproc_inc.h"
|
||||||
|
#include "mediapipe/framework/port/opencv_video_inc.h"
|
||||||
|
#include "mediapipe/framework/port/opencv_imgcodecs_inc.h"
|
||||||
|
#include "opencv2/imgcodecs/ios.h"
|
||||||
|
|
||||||
|
#import <UIKit/UIKit.h>
|
||||||
|
|
||||||
//#import "mediapipe/objc/MPPLayerRenderer.h"
|
//#import "mediapipe/objc/MPPLayerRenderer.h"
|
||||||
|
|
||||||
// The graph name specified is supposed to be the same as in the pb file (binarypb?)
|
// The graph name specified is supposed to be the same as in the pb file (binarypb?)
|
||||||
static NSString* const kGraphName = @"face_mesh_ios_lib_gpu";
|
static NSString* const kGraphName = @"face_mesh_ios_lib_gpu";
|
||||||
|
// static NSString* const kGraphName = @"pure_face_mesh_mobile_gpu";
|
||||||
|
|
||||||
static const char* kInputStream = "input_video";
|
static const char* kInputStream = "input_video";
|
||||||
static const char* kNumFacesInputSidePacket = "num_faces";
|
static const char* kNumFacesInputSidePacket = "num_faces";
|
||||||
|
@ -157,73 +163,235 @@ static const int kNumFaces = 1;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Resize the CVPixelBufferRef with INTER_AREA.
|
- (void)extractRegions:(NSURL *)fileName
|
||||||
- (CVPixelBufferRef)resize:(CVPixelBufferRef)pixelBuffer
|
foreheadBoxes:(NSArray<NSArray<IntPoint *> *> *)foreheadBoxes
|
||||||
width:(int)width
|
leftCheekBoxes:(NSArray<NSArray<IntPoint *> *> *)leftCheekBoxes
|
||||||
height:(int)height {
|
rightCheekBoxes:(NSArray<NSArray<IntPoint *> *> *)rightCheekBoxes
|
||||||
|
totalFramesNeedProcess:(NSInteger)totalFramesNeedProcess
|
||||||
|
skipNFirstFrames:(NSInteger)skipNFirstFrames {
|
||||||
|
|
||||||
OSType srcType = CVPixelBufferGetPixelFormatType(pixelBuffer);
|
NSString *filename = fileName.path;
|
||||||
size_t channels = 2;
|
std::string filePath = [filename UTF8String];
|
||||||
if (srcType == kCVPixelFormatType_32ARGB || srcType == kCVPixelFormatType_32BGRA) {
|
|
||||||
channels = 4;
|
cv::VideoCapture vid(filePath);
|
||||||
|
|
||||||
|
if (!vid.isOpened()) {
|
||||||
|
printf("@Error Opening video file");
|
||||||
|
}
|
||||||
|
else {
|
||||||
|
printf("File Opened AAAA");
|
||||||
|
|
||||||
|
// NSMutableArray *foreheadURLs = [NSMutableArray new];
|
||||||
|
// NSMutableArray *leftcheekURLs = [NSMutableArray new];
|
||||||
|
// NSMutableArray *rightcheekURLs = [NSMutableArray new];
|
||||||
|
|
||||||
|
int startFrame = int(vid.get(cv::CAP_PROP_POS_FRAMES));
|
||||||
|
int totalFrame = int(vid.get(cv::CAP_PROP_FRAME_COUNT));
|
||||||
|
// int nframes = totalFrame - startFrame;
|
||||||
|
|
||||||
|
// if (totalFramesNeedProcess > nframes) {
|
||||||
|
// NSLog(@"Video too short");
|
||||||
|
// return;
|
||||||
|
// }
|
||||||
|
|
||||||
|
if (skipNFirstFrames < 0 || totalFramesNeedProcess < 0) {
|
||||||
|
vid.release();
|
||||||
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Lock the CVPixelBuffer
|
int frameIdx = skipNFirstFrames;
|
||||||
CVPixelBufferLockBaseAddress(pixelBuffer, kCVPixelBufferLock_ReadOnly);
|
int maxFrameIndex = totalFramesNeedProcess;
|
||||||
|
|
||||||
// Get the pixel buffer attributes
|
if (skipNFirstFrames > 0) {
|
||||||
size_t srcWidth = CVPixelBufferGetWidth(pixelBuffer);
|
maxFrameIndex += skipNFirstFrames;
|
||||||
size_t srcHeight = CVPixelBufferGetHeight(pixelBuffer);
|
|
||||||
size_t bytesPerRow = CVPixelBufferGetBytesPerRow(pixelBuffer);
|
|
||||||
|
|
||||||
// Get the base address of the pixel buffer
|
|
||||||
unsigned char *baseAddress = (unsigned char *)CVPixelBufferGetBaseAddress(pixelBuffer);
|
|
||||||
|
|
||||||
// Create a cv::Mat without copying the data
|
|
||||||
cv::Mat argbImage(srcHeight, srcWidth, CV_8UC(channels), baseAddress, bytesPerRow);
|
|
||||||
|
|
||||||
// Create a cv::Mat to hold the resized image
|
|
||||||
cv::Mat resizedImage;
|
|
||||||
|
|
||||||
// Resize the image using cv::resize
|
|
||||||
cv::resize(argbImage, resizedImage, cv::Size(width, height), 0, 0, cv::INTER_AREA);
|
|
||||||
|
|
||||||
// Unlock the CVPixelBuffer
|
|
||||||
CVPixelBufferUnlockBaseAddress(pixelBuffer, kCVPixelBufferLock_ReadOnly);
|
|
||||||
|
|
||||||
// Create a new CVPixelBuffer with the desired size and format
|
|
||||||
CVPixelBufferRef resizedPixelBuffer;
|
|
||||||
CVReturn result = CVPixelBufferCreate(NULL, width, height, srcType, NULL, &resizedPixelBuffer);
|
|
||||||
if (result != kCVReturnSuccess) {
|
|
||||||
NSLog(@"Failed to create CVPixelBuffer. Error: %d", result);
|
|
||||||
return nil;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Lock the resized CVPixelBuffer
|
// Process forehead
|
||||||
CVPixelBufferLockBaseAddress(resizedPixelBuffer, 0);
|
std::vector<cv::Mat> foreheads;
|
||||||
|
std::vector<cv::Mat> leftcheeks;
|
||||||
|
std::vector<cv::Mat> rightcheeks;
|
||||||
|
|
||||||
// Get the base address and bytes per row of the resized pixel buffer
|
while (frameIdx < maxFrameIndex) {
|
||||||
void *resizedBaseAddress = CVPixelBufferGetBaseAddress(resizedPixelBuffer);
|
cv::Mat frame;
|
||||||
size_t resizedBytesPerRow = CVPixelBufferGetBytesPerRow(resizedPixelBuffer);
|
|
||||||
|
|
||||||
// Create a cv::Mat wrapper for the resized pixel buffer
|
if (!vid.read(frame)) {
|
||||||
cv::Mat resizedPixelBufferMat(height, width, CV_8UC(channels), resizedBaseAddress, resizedBytesPerRow);
|
NSLog(@"Failed to read frame.");
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
cv::cvtColor(frame, frame, cv::COLOR_BGR2RGB);
|
||||||
|
// Process the frame (e.g., display or save it)
|
||||||
|
|
||||||
// Convert the resized image (cv::Mat) to the resized pixel buffer (CVPixelBuffer)
|
NSLog(@"frame index: %d", frameIdx);
|
||||||
resizedImage.copyTo(resizedPixelBufferMat);
|
|
||||||
|
|
||||||
// Unlock the resized CVPixelBuffer
|
NSMutableArray *rowForehead = [foreheadBoxes objectAtIndex:(frameIdx - skipNFirstFrames)];
|
||||||
CVPixelBufferUnlockBaseAddress(resizedPixelBuffer, 0);
|
NSMutableArray *rowLeftCheek = [leftCheekBoxes objectAtIndex:(frameIdx - skipNFirstFrames)];
|
||||||
|
NSMutableArray *rowRightCheek = [rightCheekBoxes objectAtIndex:(frameIdx - skipNFirstFrames)];
|
||||||
|
|
||||||
// Return the resized CVPixelBuffer
|
cv::Mat forehead = extractRegion(frame, rowForehead);
|
||||||
return resizedPixelBuffer;
|
cv::Mat leftCheek = extractRegion(frame, rowLeftCheek);
|
||||||
|
cv::Mat rightCheek = extractRegion(frame, rowRightCheek);
|
||||||
|
|
||||||
|
foreheads.push_back(forehead);
|
||||||
|
leftcheeks.push_back(leftCheek);
|
||||||
|
rightcheeks.push_back(rightCheek);
|
||||||
|
|
||||||
|
frameIdx++;
|
||||||
|
}
|
||||||
|
|
||||||
|
NSLog(@"total foreheads: %d", foreheads.size());
|
||||||
|
NSLog(@"total leftCheeks: %d", leftcheeks.size());
|
||||||
|
NSLog(@"total rightCheeks: %d", rightcheeks.size());
|
||||||
|
|
||||||
|
// for (int i = 0; i < foreheads.size(); i++) {
|
||||||
|
// cv::Mat rgbaImage;
|
||||||
|
// cv::cvtColor(foreheads[i], rgbaImage, cv::COLOR_RGB2RGBA);
|
||||||
|
// saveCVImageAsPNG(MatToUIImage(rgbaImage), @"forehead");
|
||||||
|
// }
|
||||||
|
|
||||||
|
NSMutableArray *foreheadURLs = saveCVImagesAsPNGs(foreheads, @"forehead");
|
||||||
|
NSMutableArray *leftcheekURLs = saveCVImagesAsPNGs(leftcheeks, @"leftcheek");
|
||||||
|
NSMutableArray *rightcheekURLs = saveCVImagesAsPNGs(rightcheeks, @"rightcheek");
|
||||||
|
|
||||||
|
// cv::cvtColor(leftCheeks[0], rgbaImage, cv::COLOR_RGB2RGBA);
|
||||||
|
// NSData *firstData = [NSData dataWithBytes:rgbaImage.data length:rgbaImage.total() * rgbaImage.elemSize()];
|
||||||
|
// saveCVImageAsPNG([UIImage imageWithData:firstData], @"leftcheek");
|
||||||
|
|
||||||
|
// cv::cvtColor(rightCheeks[0], rgbaImage, cv::COLOR_RGB2RGBA);
|
||||||
|
// NSData *firstData = [NSData dataWithBytes:rgbaImage.data length:rgbaImage.total() * rgbaImage.elemSize()];
|
||||||
|
// saveCVImageAsPNG([UIImage imageWithData:firstData], @"rightcheek");
|
||||||
|
|
||||||
|
if([self.delegate respondsToSelector:@selector(didSavedRegions:leftcheekURLs:rightcheekURLs:)]) {
|
||||||
|
NSLog(@"nguyencse ==> has didSavedRegions");
|
||||||
|
[self.delegate didSavedRegions:foreheadURLs leftcheekURLs:leftcheekURLs rightcheekURLs:rightcheekURLs];
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
vid.release();
|
||||||
|
}
|
||||||
|
|
||||||
|
cv::Mat extractRegion(cv::Mat img, NSArray<IntPoint *> *box) {
|
||||||
|
IntPoint* point0 = [box objectAtIndex:0];
|
||||||
|
IntPoint* point1 = [box objectAtIndex:1];
|
||||||
|
IntPoint* point2 = [box objectAtIndex:2];
|
||||||
|
IntPoint* point3 = [box objectAtIndex:3];
|
||||||
|
|
||||||
|
// LEFT TOP --> RIGHT TOP --> RIGHT BOTTOM --> LEFT BOTTOM
|
||||||
|
int frameWidth = point1.x - point0.x;
|
||||||
|
int frameHeight = point3.y - point0.y;
|
||||||
|
cv::Mat region = cropROI(img, cv::Rect(point0.x, point0.y, frameWidth, frameHeight));
|
||||||
|
|
||||||
|
// square
|
||||||
|
region = square(region);
|
||||||
|
|
||||||
|
// resize to 32x32
|
||||||
|
region = resizeWithAreaInterpolation(region, cv::Size(32, 32));
|
||||||
|
|
||||||
|
return region;
|
||||||
|
}
|
||||||
|
|
||||||
|
cv::Mat cropROI(cv::Mat src, cv::Rect roi) {
|
||||||
|
// Crop the full image to that image contained by the rectangle myROI
|
||||||
|
// Note that this doesn't copy the data
|
||||||
|
cv::Mat croppedRef(src, roi);
|
||||||
|
|
||||||
|
cv::Mat cropped;
|
||||||
|
// Copy the data into new matrix
|
||||||
|
croppedRef.copyTo(cropped);
|
||||||
|
|
||||||
|
return cropped;
|
||||||
|
}
|
||||||
|
|
||||||
|
cv::Mat square(cv::Mat frame) {
|
||||||
|
if (frame.rows < frame.cols) {
|
||||||
|
int diff = frame.cols - frame.rows;
|
||||||
|
cv::Mat pad(frame.rows, diff, frame.type(), cv::Scalar(0));
|
||||||
|
cv::hconcat(frame, pad, frame);
|
||||||
|
} else if (frame.rows > frame.cols) {
|
||||||
|
int diff = frame.rows - frame.cols;
|
||||||
|
cv::Mat pad(diff, frame.cols, frame.type(), cv::Scalar(0));
|
||||||
|
cv::vconcat(frame, pad, frame);
|
||||||
|
}
|
||||||
|
|
||||||
|
return frame;
|
||||||
|
}
|
||||||
|
|
||||||
|
cv::Mat resizeWithAreaInterpolation(cv::Mat image, cv::Size newShape) {
|
||||||
|
cv::Size originalShape = image.size();
|
||||||
|
cv::resize(image, image, newShape, 0, 0, cv::INTER_AREA);
|
||||||
|
return image;
|
||||||
|
}
|
||||||
|
|
||||||
|
// NSURL *saveCVImageAsPNG(UIImage *image, NSString *fileName) {
|
||||||
|
// // Create a unique file name for each image
|
||||||
|
// NSString *fileNameIndex = [fileName stringByAppendingFormat:@"_%d", 0];
|
||||||
|
// NSString *filePath = [NSTemporaryDirectory() stringByAppendingPathComponent:[fileNameIndex stringByAppendingString:@".png"]];
|
||||||
|
// NSURL *fileURL = [NSURL fileURLWithPath:filePath];
|
||||||
|
|
||||||
|
// NSData *pngData = UIImagePNGRepresentation(image);
|
||||||
|
|
||||||
|
// if ([pngData writeToFile:filePath atomically:YES]) {
|
||||||
|
// NSLog(@"PNG file saved successfully at path: %@", filePath);
|
||||||
|
// } else {
|
||||||
|
// NSLog(@"Failed to save PNG file at path: %@", filePath);
|
||||||
|
// }
|
||||||
|
|
||||||
|
// return fileURL;
|
||||||
|
// }
|
||||||
|
|
||||||
|
NSArray<NSURL *> *saveCVImagesAsPNGs(std::vector<cv::Mat> frames, NSString *folderName) {
|
||||||
|
NSMutableArray<NSURL *> *fileURLs = [NSMutableArray arrayWithCapacity:frames.size()];
|
||||||
|
|
||||||
|
for (int i = 0; i < frames.size(); i++) {
|
||||||
|
// Create a unique file name for each image
|
||||||
|
NSString *fileNameIndex = [folderName stringByAppendingFormat:@"_%d", i];
|
||||||
|
NSString *filePath = [NSTemporaryDirectory() stringByAppendingPathComponent:[fileNameIndex stringByAppendingString:@".png"]];
|
||||||
|
NSURL *fileURL = [NSURL fileURLWithPath:filePath];
|
||||||
|
|
||||||
|
// NSLog(@"File URL: %@", fileURL);
|
||||||
|
|
||||||
|
// Do NOT compress pixel
|
||||||
|
std::vector<int> compressionParams;
|
||||||
|
compressionParams.push_back(cv::IMWRITE_PNG_COMPRESSION);
|
||||||
|
compressionParams.push_back(0);
|
||||||
|
|
||||||
|
const char * cpath = [filePath cStringUsingEncoding:NSUTF8StringEncoding];
|
||||||
|
const cv::String newPath = (const cv::String)cpath;
|
||||||
|
cv::imwrite(newPath, frames[i], compressionParams);
|
||||||
|
|
||||||
|
// Add file URL to the array
|
||||||
|
[fileURLs addObject:fileURL];
|
||||||
|
}
|
||||||
|
|
||||||
|
return [fileURLs copy];
|
||||||
}
|
}
|
||||||
|
|
||||||
@end
|
@end
|
||||||
|
|
||||||
|
|
||||||
|
@implementation NSValue (IntPoint)
|
||||||
|
+ (instancetype)valuewithIntPoint:(IntPoint *)value {
|
||||||
|
return [self valueWithBytes:&value objCType:@encode(IntPoint)];
|
||||||
|
}
|
||||||
|
- (IntPoint *) intPointValue {
|
||||||
|
IntPoint* value;
|
||||||
|
[self getValue:&value];
|
||||||
|
return value;
|
||||||
|
}
|
||||||
|
@end
|
||||||
|
|
||||||
@implementation FaceMeshLandmarkPoint
|
@implementation FaceMeshLandmarkPoint
|
||||||
@end
|
@end
|
||||||
|
|
||||||
@implementation FaceMeshNormalizedRect
|
@implementation FaceMeshNormalizedRect
|
||||||
@end
|
@end
|
||||||
|
|
||||||
|
@implementation IntPoint
|
||||||
|
- (instancetype)initWithX:(NSInteger)x y:(NSInteger)y {
|
||||||
|
self = [super init];
|
||||||
|
if (self) {
|
||||||
|
_x = x;
|
||||||
|
_y = y;
|
||||||
|
}
|
||||||
|
return self;
|
||||||
|
}
|
||||||
|
@end
|
BIN
mediapipe/graphs/.DS_Store
vendored
BIN
mediapipe/graphs/.DS_Store
vendored
Binary file not shown.
BIN
mediapipe/graphs/face_mesh/.DS_Store
vendored
Normal file
BIN
mediapipe/graphs/face_mesh/.DS_Store
vendored
Normal file
Binary file not shown.
|
@ -74,3 +74,10 @@ mediapipe_binary_graph(
|
||||||
output_name = "face_mesh_ios_lib_gpu.binarypb",
|
output_name = "face_mesh_ios_lib_gpu.binarypb",
|
||||||
deps = [":mobile_calculators"],
|
deps = [":mobile_calculators"],
|
||||||
)
|
)
|
||||||
|
|
||||||
|
mediapipe_binary_graph(
|
||||||
|
name = "pure_face_mesh_mobile_gpu_binary_graph",
|
||||||
|
graph = "pure_face_mesh_mobile.pbtxt",
|
||||||
|
output_name = "pure_face_mesh_mobile_gpu.binarypb",
|
||||||
|
deps = [":mobile_calculators"],
|
||||||
|
)
|
||||||
|
|
74
mediapipe/graphs/face_mesh/pure_face_mesh_mobile.pbtxt
Normal file
74
mediapipe/graphs/face_mesh/pure_face_mesh_mobile.pbtxt
Normal file
|
@ -0,0 +1,74 @@
|
||||||
|
# MediaPipe graph that performs face mesh with TensorFlow Lite on GPU.
|
||||||
|
# Edited from face_mesh_mobile.pbtxt because I don't want babysitting of auto throttling and such. I'll do it myself
|
||||||
|
|
||||||
|
# GPU buffer. (GpuBuffer)
|
||||||
|
input_stream: "input_video"
|
||||||
|
|
||||||
|
# Max number of faces to detect/process. (int)
|
||||||
|
input_side_packet: "num_faces"
|
||||||
|
|
||||||
|
# Output image with rendered results. (GpuBuffer)
|
||||||
|
# nope no rendering
|
||||||
|
# output_stream: "output_video"
|
||||||
|
|
||||||
|
# Collection of detected/processed faces, each represented as a list of
|
||||||
|
# landmarks. (std::vector<NormalizedLandmarkList>)
|
||||||
|
output_stream: "multi_face_landmarks"
|
||||||
|
|
||||||
|
# Regions of interest calculated based on landmarks.
|
||||||
|
# (For more info see mediapipe/modules/face_landmark/face_landmark_front_gpu.pbtxt)
|
||||||
|
# (std::vector<NormalizedRect>)
|
||||||
|
# For typings see "mediapipe/framework/formats/rect.pb.h"
|
||||||
|
output_stream: "face_rects_from_landmarks"
|
||||||
|
|
||||||
|
# The detections from the box model
|
||||||
|
# see detection.proto
|
||||||
|
# Regions of interest calculated based on face detections.
|
||||||
|
# (std::vector<NormalizedRect>)
|
||||||
|
# output_stream: "face_rects_from_detections"
|
||||||
|
|
||||||
|
# Extra outputs (for debugging, for instance).
|
||||||
|
# Detected faces. (std::vector<Detection>)
|
||||||
|
# (std::vector<Detections>)
|
||||||
|
# output_stream: "face_detections"
|
||||||
|
|
||||||
|
# Landmark presence (needed because whole graph won't emit anything if no faces are detected)
|
||||||
|
output_stream: "landmark_presence"
|
||||||
|
|
||||||
|
# screw the throttling, we do that ourselves.
|
||||||
|
# *throttling node code was deleted from here*
|
||||||
|
|
||||||
|
# Subgraph that detects faces and corresponding landmarks.
|
||||||
|
node {
|
||||||
|
calculator: "FaceLandmarkFrontGpu"
|
||||||
|
# the IMAGE: part is saying, pipe this data into the input with the name `image`
|
||||||
|
input_stream: "IMAGE:input_video"
|
||||||
|
input_side_packet: "NUM_FACES:num_faces"
|
||||||
|
output_stream: "LANDMARKS:multi_face_landmarks"
|
||||||
|
output_stream: "ROIS_FROM_LANDMARKS:face_rects_from_landmarks"
|
||||||
|
# face_detections is the stream that comes out from face_detection_short_range_common
|
||||||
|
# output_stream: "DETECTIONS:face_detections"
|
||||||
|
|
||||||
|
# output_stream: "ROIS_FROM_DETECTIONS:face_rects_from_detections"
|
||||||
|
}
|
||||||
|
|
||||||
|
# See this thread here https://github.com/google/mediapipe/issues/850#issuecomment-683268033
|
||||||
|
# "if there are no packets in the corresponding output stream, it is designed to wait until the packet comes in"
|
||||||
|
# That means that we'd get absolutely nothing to work with and won't know if our frame had anythin!
|
||||||
|
# So we add PacketPresenceCalculator
|
||||||
|
node {
|
||||||
|
calculator: "PacketPresenceCalculator"
|
||||||
|
input_stream: "PACKET:multi_face_landmarks"
|
||||||
|
output_stream: "PRESENCE:landmark_presence"
|
||||||
|
}
|
||||||
|
|
||||||
|
# nope not rendering.
|
||||||
|
# Subgraph that renders face-landmark annotation onto the input image.
|
||||||
|
# node {
|
||||||
|
# calculator: "FaceRendererGpu"
|
||||||
|
# input_stream: "IMAGE:throttled_input_video"
|
||||||
|
# input_stream: "LANDMARKS:multi_face_landmarks"
|
||||||
|
# input_stream: "NORM_RECTS:face_rects_from_landmarks"
|
||||||
|
# input_stream: "DETECTIONS:face_detections"
|
||||||
|
# output_stream: "IMAGE:output_video"
|
||||||
|
#}
|
BIN
third_party/.DS_Store
vendored
Normal file
BIN
third_party/.DS_Store
vendored
Normal file
Binary file not shown.
Loading…
Reference in New Issue
Block a user