image_style full

This commit is contained in:
mslight 2022-08-03 14:31:05 +04:00
parent c22ffa0012
commit 5fbb1988f5
13 changed files with 1462 additions and 601 deletions

View File

@ -18,14 +18,24 @@ licenses(["notice"])
package(default_visibility = ["//visibility:public"]) package(default_visibility = ["//visibility:public"])
mediapipe_proto_library(
name = "fast_utils_calculator_proto",
srcs = ["fast_utils_calculator.proto"],
visibility = ["//visibility:public"],
deps = [
"//mediapipe/framework:calculator_options_proto",
"//mediapipe/framework:calculator_proto",
],
)
cc_library( cc_library(
name = "fast_utils_calculator", name = "fast_utils_calculator",
srcs = ["fast_utils_calculator.cc"], srcs = ["fast_utils_calculator.cc"],
visibility = ["//visibility:public"], visibility = ["//visibility:public"],
deps = [ deps = [
":fast_utils_calculator_cc_proto",
"//mediapipe/framework:calculator_options_cc_proto", "//mediapipe/framework:calculator_options_cc_proto",
"//mediapipe/framework/formats:image_format_cc_proto", "//mediapipe/framework/formats:image_format_cc_proto",
"//mediapipe/util:color_cc_proto",
"@com_google_absl//absl/strings", "@com_google_absl//absl/strings",
"//mediapipe/framework:calculator_framework", "//mediapipe/framework:calculator_framework",
"//mediapipe/framework/formats:image_frame", "//mediapipe/framework/formats:image_frame",
@ -38,8 +48,32 @@ cc_library(
"//mediapipe/framework/port:opencv_highgui", "//mediapipe/framework/port:opencv_highgui",
"//mediapipe/framework/port:status", "//mediapipe/framework/port:status",
"//mediapipe/framework/port:vector", "//mediapipe/framework/port:vector",
"//mediapipe/util:annotation_renderer", ],
"//mediapipe/util:render_data_cc_proto", alwayslink = 1,
)
cc_library(
name = "apply_mask_calculator",
srcs = ["apply_mask_calculator.cc"],
visibility = ["//visibility:public"],
deps = [
"//mediapipe/framework:calculator_options_cc_proto",
"//mediapipe/framework/formats:image_format_cc_proto",
"@com_google_absl//absl/strings",
"//mediapipe/framework:calculator_framework",
"//mediapipe/framework/formats:image_frame",
"//mediapipe/framework/formats:image_frame_opencv",
"//mediapipe/framework/formats:video_stream_header",
"//mediapipe/framework/formats:landmark_cc_proto",
"//mediapipe/framework/port:logging",
"//mediapipe/framework/port:opencv_core",
"//mediapipe/framework/port:opencv_imgproc",
"//mediapipe/framework/port:opencv_highgui",
"//mediapipe/framework/port:status",
"//mediapipe/framework/port:vector",
"//mediapipe/framework/deps:file_path",
"//mediapipe/framework/port:file_helpers",
"//mediapipe/util:resource_util",
], ],
alwayslink = 1, alwayslink = 1,
) )
@ -49,3 +83,5 @@ cc_library(

View File

@ -0,0 +1,305 @@
// Copyright 2019 The MediaPipe Authors.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#include <math.h>
#include <string>
#include <memory>
#include "absl/strings/str_cat.h"
#include "mediapipe/framework/calculator_framework.h"
#include "mediapipe/framework/formats/image_format.pb.h"
#include "mediapipe/framework/formats/image_frame.h"
#include "mediapipe/framework/formats/image_frame_opencv.h"
#include "mediapipe/framework/formats/video_stream_header.h"
#include "mediapipe/framework/formats/landmark.pb.h"
#include "mediapipe/framework/port/logging.h"
#include "mediapipe/framework/port/opencv_core_inc.h"
#include "mediapipe/framework/port/opencv_imgproc_inc.h"
#include "mediapipe/framework/port/opencv_highgui_inc.h"
#include "mediapipe/framework/port/status.h"
#include "mediapipe/framework/port/logging.h"
#include "mediapipe/framework/port/vector.h"
using namespace std;
namespace mediapipe
{
namespace
{
static const std::vector<cv::Point2f> FFHQ_NORM_LM = {
{638.68525475 / 1024, 486.24604922 / 1024},
{389.31496114 / 1024, 485.8921848 / 1024},
{513.67979275 / 1024, 620.8915371 / 1024},
{405.50932642 / 1024, 756.52797927 / 1024},
{622.55630397 / 1024, 756.15509499 / 1024}};
constexpr char kImageFrameTag[] = "IMAGE";
constexpr char kFakeBgTag[] = "FAKE_BG";
constexpr char kLmMaskTag[] = "LM_MASK";
inline bool HasImageTag(mediapipe::CalculatorContext *cc) { return false; }
cv::Mat blend_mask(cv::Mat mask_face, cv::Mat mask_bbox, int kernel_size = 33, int reduce_size = 128)
{
int k_sz = kernel_size;
auto [width, height] = mask_face.size();
cv::Mat mask_face_0 = mask_face.clone();
double K = (double)reduce_size / std::min(height, width);
cv::resize(mask_face, mask_face, {(int)(width * K), (int)(height * K)});
mask_face.convertTo(mask_face, CV_32F);
cv::GaussianBlur(mask_face, mask_face, {k_sz, k_sz}, 0);
mask_face *= 2;
cv::threshold(mask_face, mask_face, 1, 255, CV_THRESH_TRUNC);
cv::resize(mask_bbox, mask_bbox, {(int)(width * K), (int)(height * K)});
mask_bbox.convertTo(mask_bbox, CV_32F);
cv::GaussianBlur(mask_bbox, mask_bbox, {k_sz, k_sz}, 0);
cv::Mat mask_bbox_3ch;
cv::merge(std::vector{mask_bbox, mask_bbox, mask_bbox}, mask_bbox_3ch);
cv::Mat mask = mask_bbox_3ch.mul(mask_face);
cv::Mat img_out;
cv::resize(mask, img_out, {width, height});
for (int i = 1; i < mask_face_0.rows; i++)
{
for (int j = 1; j < mask_face_0.cols; j++)
{
if (mask_face_0.at<uchar>(i, j) > 0)
img_out.at<cv::Vec3b>(i, j) = 1;
}
}
return img_out;
}
} // namespace
class ApplyMaskCalculator : public CalculatorBase
{
public:
ApplyMaskCalculator() = default;
~ApplyMaskCalculator() override = default;
static absl::Status GetContract(CalculatorContract *cc);
// From Calculator.
absl::Status Open(CalculatorContext *cc) override;
absl::Status Process(CalculatorContext *cc) override;
absl::Status Close(CalculatorContext *cc) override;
private:
absl::Status CreateRenderTargetCpu(CalculatorContext *cc,
std::unique_ptr<cv::Mat> &image_mat,
std::string_view tag,
ImageFormat::Format *target_format);
absl::Status RenderToCpu(
CalculatorContext *cc, const ImageFormat::Format &target_format,
uchar *data_image, std::unique_ptr<cv::Mat> &image_mat);
// Indicates if image frame is available as input.
bool image_frame_available_ = false;
int image_width_;
int image_height_;
};
REGISTER_CALCULATOR(ApplyMaskCalculator);
absl::Status ApplyMaskCalculator::GetContract(CalculatorContract *cc)
{
CHECK_GE(cc->Inputs().NumEntries(), 1);
if (cc->Inputs().HasTag(kImageFrameTag))
{
cc->Inputs().Tag(kImageFrameTag).Set<ImageFrame>();
CHECK(cc->Outputs().HasTag(kImageFrameTag));
}
if (cc->Inputs().HasTag(kFakeBgTag))
{
cc->Inputs().Tag(kFakeBgTag).Set<ImageFrame>();
}
if (cc->Inputs().HasTag(kLmMaskTag))
{
cc->Inputs().Tag(kLmMaskTag).Set<ImageFrame>();
}
if (cc->Outputs().HasTag(kImageFrameTag))
{
cc->Outputs().Tag(kImageFrameTag).Set<ImageFrame>();
}
return absl::OkStatus();
}
absl::Status ApplyMaskCalculator::Open(CalculatorContext *cc)
{
cc->SetOffset(TimestampDiff(0));
if (cc->Inputs().HasTag(kImageFrameTag) || HasImageTag(cc))
{
image_frame_available_ = true;
}
// Set the output header based on the input header (if present).
const char *tag = kImageFrameTag;
if (image_frame_available_ && !cc->Inputs().Tag(tag).Header().IsEmpty())
{
const auto &input_header =
cc->Inputs().Tag(tag).Header().Get<VideoHeader>();
auto *output_video_header = new VideoHeader(input_header);
cc->Outputs().Tag(tag).SetHeader(Adopt(output_video_header));
}
return absl::OkStatus();
}
absl::Status ApplyMaskCalculator::Process(CalculatorContext *cc)
{
if (cc->Inputs().HasTag(kImageFrameTag) &&
cc->Inputs().Tag(kImageFrameTag).IsEmpty())
{
return absl::OkStatus();
}
// Initialize render target, drawn with OpenCV.
ImageFormat::Format target_format;
std::unique_ptr<cv::Mat> image_mat;
MP_RETURN_IF_ERROR(CreateRenderTargetCpu(cc, image_mat, kImageFrameTag, &target_format));
if (((cc->Inputs().HasTag(kFakeBgTag) &&
!cc->Inputs().Tag(kFakeBgTag).IsEmpty())) &&
((cc->Inputs().HasTag(kLmMaskTag) &&
!cc->Inputs().Tag(kLmMaskTag).IsEmpty())))
{
// Initialize render target, drawn with OpenCV.
std::unique_ptr<cv::Mat> fake_bg;
std::unique_ptr<cv::Mat> lm_mask_ptr;
MP_RETURN_IF_ERROR(CreateRenderTargetCpu(cc, fake_bg, kFakeBgTag, &target_format));
MP_RETURN_IF_ERROR(CreateRenderTargetCpu(cc, lm_mask_ptr, kLmMaskTag, &target_format));
cv::Mat mat_fake_bg_ = *fake_bg.get();
cv::Mat mat_image_ = *image_mat.get();
cv::Mat lm_mask = *lm_mask_ptr.get();
image_width_ = image_mat->cols;
image_height_ = image_mat->rows;
cv::Mat roi_mask = mat_image_.clone();
cv::transform(roi_mask, roi_mask, cv::Matx13f(1, 1, 1));
cv::threshold(roi_mask, roi_mask, 1, 255, CV_THRESH_TRUNC);
cv::Mat mask = blend_mask(lm_mask, roi_mask, 33);
mat_image_.convertTo(mat_image_, CV_32F);
mat_fake_bg_.convertTo(mat_fake_bg_, CV_32F);
cv::resize(mat_fake_bg_, mat_fake_bg_, {image_width_, image_height_});
cv::Mat im_out = mat_fake_bg_.mul(cv::Scalar::all(1) - mask) + mat_image_.mul(mask);
im_out.convertTo(*image_mat, CV_8U);
}
uchar *image_mat_ptr = image_mat->data;
MP_RETURN_IF_ERROR(RenderToCpu(cc, target_format, image_mat_ptr, image_mat));
return absl::OkStatus();
}
absl::Status ApplyMaskCalculator::Close(CalculatorContext *cc)
{
return absl::OkStatus();
}
absl::Status ApplyMaskCalculator::RenderToCpu(
CalculatorContext *cc, const ImageFormat::Format &target_format,
uchar *data_image, std::unique_ptr<cv::Mat> &image_mat)
{
auto output_frame = absl::make_unique<ImageFrame>(
target_format, image_mat->cols, image_mat->rows);
output_frame->CopyPixelData(target_format, image_mat->cols, image_mat->rows, data_image,
ImageFrame::kDefaultAlignmentBoundary);
if (cc->Outputs().HasTag(kImageFrameTag))
{
cc->Outputs()
.Tag(kImageFrameTag)
.Add(output_frame.release(), cc->InputTimestamp());
}
return absl::OkStatus();
}
absl::Status ApplyMaskCalculator::CreateRenderTargetCpu(
CalculatorContext *cc, std::unique_ptr<cv::Mat> &image_mat, std::string_view tag,
ImageFormat::Format *target_format)
{
if (image_frame_available_)
{
const auto &input_frame =
cc->Inputs().Tag(tag).Get<ImageFrame>();
int target_mat_type;
switch (input_frame.Format())
{
case ImageFormat::SRGBA:
*target_format = ImageFormat::SRGBA;
target_mat_type = CV_8UC4;
break;
case ImageFormat::SRGB:
*target_format = ImageFormat::SRGB;
target_mat_type = CV_8UC3;
break;
case ImageFormat::GRAY8:
*target_format = ImageFormat::SRGB;
target_mat_type = CV_8UC3;
break;
default:
return absl::UnknownError("Unexpected image frame format.");
break;
}
image_mat = absl::make_unique<cv::Mat>(
input_frame.Height(), input_frame.Width(), target_mat_type);
auto input_mat = formats::MatView(&input_frame);
if (input_frame.Format() == ImageFormat::GRAY8)
{
cv::Mat rgb_mat;
cv::cvtColor(input_mat, rgb_mat, CV_GRAY2RGB);
rgb_mat.copyTo(*image_mat);
}
else
{
input_mat.copyTo(*image_mat);
}
}
else
{
image_mat = absl::make_unique<cv::Mat>(
1920, 1080, CV_8UC4,
cv::Scalar::all(255));
*target_format = ImageFormat::SRGBA;
}
return absl::OkStatus();
}
} // namespace mediapipe

View File

@ -22,6 +22,7 @@
#include <memory> #include <memory>
#include "absl/strings/str_cat.h" #include "absl/strings/str_cat.h"
#include "mediapipe/calculators/image_style/fast_utils_calculator.pb.h"
#include "mediapipe/framework/calculator_framework.h" #include "mediapipe/framework/calculator_framework.h"
#include "mediapipe/framework/calculator_options.pb.h" #include "mediapipe/framework/calculator_options.pb.h"
#include "mediapipe/framework/formats/image_format.pb.h" #include "mediapipe/framework/formats/image_format.pb.h"
@ -41,7 +42,7 @@ namespace mediapipe
{ {
namespace namespace
{ {
static const std::vector<cv::Point2f> FFHQ_NORM_LM = { const std::vector<cv::Point2f> FFHQ_NORM_LM = {
{638.68525475 / 1024, 486.24604922 / 1024}, {638.68525475 / 1024, 486.24604922 / 1024},
{389.31496114 / 1024, 485.8921848 / 1024}, {389.31496114 / 1024, 485.8921848 / 1024},
{513.67979275 / 1024, 620.8915371 / 1024}, {513.67979275 / 1024, 620.8915371 / 1024},
@ -52,16 +53,8 @@ namespace mediapipe
constexpr char kVectorTag[] = "VECTOR"; constexpr char kVectorTag[] = "VECTOR";
constexpr char kLandmarksTag[] = "LANDMARKS"; constexpr char kLandmarksTag[] = "LANDMARKS";
constexpr char kNormLandmarksTag[] = "NORM_LANDMARKS"; constexpr char kNormLandmarksTag[] = "NORM_LANDMARKS";
constexpr char kLmMaskTag[] = "LM_MASK";
std::tuple<int, int> _normalized_to_pixel_coordinates(float normalized_x, constexpr char kSizeTag[] = "SIZE";
float normalized_y, int image_width, int image_height)
{
// Converts normalized value pair to pixel coordinates
int x_px = std::min<int>(floor(normalized_x * image_width), image_width - 1);
int y_px = std::min<int>(floor(normalized_y * image_height), image_height - 1);
return {x_px, y_px};
};
static const std::vector<cv::Point> FACEMESH_FACE_OVAL{ static const std::vector<cv::Point> FACEMESH_FACE_OVAL{
{10, 338}, {338, 297}, {297, 332}, {332, 284}, {284, 251}, {251, 389}, {389, 356}, {356, 454}, {454, 323}, {323, 361}, {361, 288}, {288, 397}, {397, 365}, {365, 379}, {379, 378}, {378, 400}, {400, 377}, {377, 152}, {152, 148}, {148, 176}, {176, 149}, {149, 150}, {150, 136}, {136, 172}, {172, 58}, {58, 132}, {132, 93}, {93, 234}, {234, 127}, {127, 162}, {162, 21}, {21, 54}, {54, 103}, {103, 67}, {67, 109}, {109, 10}}; {10, 338}, {338, 297}, {297, 332}, {332, 284}, {284, 251}, {251, 389}, {389, 356}, {356, 454}, {454, 323}, {323, 361}, {361, 288}, {288, 397}, {397, 365}, {365, 379}, {379, 378}, {378, 400}, {400, 377}, {377, 152}, {152, 148}, {148, 176}, {176, 149}, {149, 150}, {150, 136}, {136, 172}, {172, 58}, {58, 132}, {132, 93}, {93, 234}, {234, 127}, {127, 162}, {162, 21}, {21, 54}, {54, 103}, {103, 67}, {67, 109}, {109, 10}};
@ -113,16 +106,15 @@ namespace mediapipe
cv::Mat &source, cv::Mat &source,
cv::Mat &target, float eps = 1e-7) cv::Mat &target, float eps = 1e-7)
{ {
cv::Mat source_mean_mat, target_mean_mat, source1ch, target1ch; cv::Mat source_mean_mat, target_mean_mat;
cv::reduce(source, source_mean_mat, 0, CV_REDUCE_AVG, CV_32F); cv::reduce(source, source_mean_mat, 0, CV_REDUCE_AVG, CV_32F);
cv::reduce(target, target_mean_mat, 0, CV_REDUCE_AVG, CV_32F); cv::reduce(target, target_mean_mat, 0, CV_REDUCE_AVG, CV_32F);
source -= {source_mean_mat.at<float>(0, 0), source_mean_mat.at<float>(0, 1)}; source -= {source_mean_mat.at<float>(0, 0), source_mean_mat.at<float>(0, 1)};
target -= {target_mean_mat.at<float>(0, 0), target_mean_mat.at<float>(0, 1)}; target -= {target_mean_mat.at<float>(0, 0), target_mean_mat.at<float>(0, 1)};
source1ch = source.reshape(1, 5); cv::Mat source1ch = source.reshape(1, 5);
target1ch = target.reshape(1, 5); cv::Mat target1ch = target.reshape(1, 5);
cv::Mat source_std_mat, target_std_mat; cv::Mat source_std_mat, target_std_mat;
cv::meanStdDev(source1ch, cv::noArray(), source_std_mat); cv::meanStdDev(source1ch, cv::noArray(), source_std_mat);
@ -136,21 +128,21 @@ namespace mediapipe
source /= source_std + eps; source /= source_std + eps;
target /= target_std + eps; target /= target_std + eps;
cv::Mat u, vt, rotation, w;
source1ch = source.reshape(1, 5); source1ch = source.reshape(1, 5);
target1ch = target.reshape(1, 5); target1ch = target.reshape(1, 5);
cv::Mat u, vt, w;
cv::SVD::compute(source1ch.t() * target1ch, w, u, vt); cv::SVD::compute(source1ch.t() * target1ch, w, u, vt);
rotation = (u * vt).t(); cv::Mat rotation = (u * vt).t();
float scale = target_std / (source_std + eps);
float scale = target_std / source_std + eps;
cv::Mat translation; cv::Mat translation;
cv::subtract(target_mean_mat.reshape(1, 2),
scale * rotation * source_mean_mat.reshape(1, 2), translation);
cv::subtract(target_mean_mat.reshape(1, 2), scale * rotation * source_mean_mat.reshape(1, 2), translation); return {scale, rotation, translation};
return std::make_tuple(scale, rotation, translation);
} }
std::tuple<float, float, float, float> Crop( std::tuple<float, float, float, float> Crop(
@ -210,6 +202,9 @@ namespace mediapipe
absl::Status Process(CalculatorContext *cc) override; absl::Status Process(CalculatorContext *cc) override;
absl::Status Close(CalculatorContext *cc) override; absl::Status Close(CalculatorContext *cc) override;
protected:
mediapipe::FastUtilsCalculatorOptions options_;
private: private:
absl::Status CreateRenderTargetCpu(CalculatorContext *cc, absl::Status CreateRenderTargetCpu(CalculatorContext *cc,
std::unique_ptr<cv::Mat> &image_mat, std::unique_ptr<cv::Mat> &image_mat,
@ -217,7 +212,7 @@ namespace mediapipe
absl::Status RenderToCpu( absl::Status RenderToCpu(
CalculatorContext *cc, const ImageFormat::Format &target_format, CalculatorContext *cc, const ImageFormat::Format &target_format,
uchar *data_image, std::unique_ptr<cv::Mat> &image_mat); uchar *data_image, std::unique_ptr<cv::Mat> &image_mat, std::string_view tag);
absl::Status Call(CalculatorContext *cc, absl::Status Call(CalculatorContext *cc,
std::unique_ptr<cv::Mat> &image_mat, std::unique_ptr<cv::Mat> &image_mat,
@ -231,19 +226,20 @@ namespace mediapipe
// Indicates if image frame is available as input. // Indicates if image frame is available as input.
bool image_frame_available_ = false; bool image_frame_available_ = false;
std::vector<std::pair<std::string, const std::vector<int>>> index_dict = {
const std::vector<std::pair<std::string, std::vector<int>>> index_dict = {
{"leftEye", {384, 385, 386, 387, 388, 390, 263, 362, 398, 466, 373, 374, 249, 380, 381, 382}}, {"leftEye", {384, 385, 386, 387, 388, 390, 263, 362, 398, 466, 373, 374, 249, 380, 381, 382}},
{"rightEye", {160, 33, 161, 163, 133, 7, 173, 144, 145, 246, 153, 154, 155, 157, 158, 159}}, {"rightEye", {160, 33, 161, 163, 133, 7, 173, 144, 145, 246, 153, 154, 155, 157, 158, 159}},
{"nose", {4}}, {"nose", {4}},
//{"lips", {0, 13, 14, 17, 84}},
{"leftLips", {61, 146}}, {"leftLips", {61, 146}},
{"rightLips", {291, 375}}, {"rightLips", {291, 375}},
}; };
std::unique_ptr<cv::Mat> image_mat;
cv::Mat mat_image_; cv::Mat mat_image_;
cv::Mat lm_mask;
int image_width_; int image_width_;
int image_height_; int image_height_;
bool back_to_im;
}; };
REGISTER_CALCULATOR(FastUtilsCalculator); REGISTER_CALCULATOR(FastUtilsCalculator);
@ -273,18 +269,29 @@ namespace mediapipe
{ {
cc->Inputs().Tag(kNormLandmarksTag).Set<std::vector<NormalizedLandmarkList>>(); cc->Inputs().Tag(kNormLandmarksTag).Set<std::vector<NormalizedLandmarkList>>();
} }
if (cc->Inputs().HasTag(kSizeTag))
{
cc->Inputs().Tag(kSizeTag).Set<std::pair<int, int>>();
}
if (cc->Outputs().HasTag(kImageFrameTag)) if (cc->Outputs().HasTag(kImageFrameTag))
{ {
cc->Outputs().Tag(kImageFrameTag).Set<ImageFrame>(); cc->Outputs().Tag(kImageFrameTag).Set<ImageFrame>();
} }
if (cc->Outputs().HasTag(kLmMaskTag))
{
cc->Outputs().Tag(kLmMaskTag).Set<ImageFrame>();
}
return absl::OkStatus(); return absl::OkStatus();
} }
absl::Status FastUtilsCalculator::Open(CalculatorContext *cc) absl::Status FastUtilsCalculator::Open(CalculatorContext *cc)
{ {
cc->SetOffset(TimestampDiff(0)); cc->SetOffset(TimestampDiff(0));
options_ = cc->Options<mediapipe::FastUtilsCalculatorOptions>();
back_to_im = options_.back_to_image();
if (cc->Inputs().HasTag(kImageFrameTag) || HasImageTag(cc)) if (cc->Inputs().HasTag(kImageFrameTag) || HasImageTag(cc))
{ {
@ -313,7 +320,9 @@ namespace mediapipe
} }
// Initialize render target, drawn with OpenCV. // Initialize render target, drawn with OpenCV.
std::unique_ptr<cv::Mat> image_mat;
ImageFormat::Format target_format; ImageFormat::Format target_format;
ImageFormat::Format target_format2;
std::vector<std::vector<cv::Point2f>> lms_out; std::vector<std::vector<cv::Point2f>> lms_out;
MP_RETURN_IF_ERROR(CreateRenderTargetCpu(cc, image_mat, &target_format)); MP_RETURN_IF_ERROR(CreateRenderTargetCpu(cc, image_mat, &target_format));
@ -326,14 +335,42 @@ namespace mediapipe
{ {
MP_RETURN_IF_ERROR(Call(cc, image_mat, target_format, lms_out)); MP_RETURN_IF_ERROR(Call(cc, image_mat, target_format, lms_out));
cv::Mat source_lm = cv::Mat(lms_out[0]); if (cc->Outputs().HasTag(kLmMaskTag))
{
lm_mask.convertTo(lm_mask, CV_8U);
MP_RETURN_IF_ERROR(Align(image_mat, source_lm)); std::unique_ptr<cv::Mat> lm_mask_ptr = absl::make_unique<cv::Mat>(
mat_image_.size(), lm_mask.type());
lm_mask.copyTo(*lm_mask_ptr);
target_format2 = ImageFormat::GRAY8;
uchar *lm_mask_pt = lm_mask_ptr->data;
MP_RETURN_IF_ERROR(RenderToCpu(cc, target_format2, lm_mask_pt, lm_mask_ptr, kLmMaskTag));
}
if (!back_to_im)
{
MP_RETURN_IF_ERROR(Align(image_mat, cv::Mat(lms_out[0])));
}
else
{
const auto &size =
cc->Inputs().Tag(kSizeTag).Get<std::pair<int, int>>();
cv::Mat tar = cv::Mat(FFHQ_NORM_LM) * 256;
MP_RETURN_IF_ERROR(Align(image_mat, tar,
cv::Mat(lms_out[0]), {size.first, size.second}));
}
uchar *image_mat_ptr = image_mat->data;
MP_RETURN_IF_ERROR(RenderToCpu(cc, target_format, image_mat_ptr, image_mat, kImageFrameTag));
}
else
{
uchar *image_mat_ptr = image_mat->data;
MP_RETURN_IF_ERROR(RenderToCpu(cc, target_format, image_mat_ptr, image_mat, kImageFrameTag));
} }
uchar *image_mat_ptr = image_mat->data;
MP_RETURN_IF_ERROR(RenderToCpu(cc, target_format, image_mat_ptr, image_mat));
return absl::OkStatus(); return absl::OkStatus();
} }
@ -344,7 +381,7 @@ namespace mediapipe
absl::Status FastUtilsCalculator::RenderToCpu( absl::Status FastUtilsCalculator::RenderToCpu(
CalculatorContext *cc, const ImageFormat::Format &target_format, CalculatorContext *cc, const ImageFormat::Format &target_format,
uchar *data_image, std::unique_ptr<cv::Mat> &image_mat) uchar *data_image, std::unique_ptr<cv::Mat> &image_mat, std::string_view tag)
{ {
auto output_frame = absl::make_unique<ImageFrame>( auto output_frame = absl::make_unique<ImageFrame>(
target_format, image_mat->cols, image_mat->rows); target_format, image_mat->cols, image_mat->rows);
@ -352,10 +389,10 @@ namespace mediapipe
output_frame->CopyPixelData(target_format, image_mat->cols, image_mat->rows, data_image, output_frame->CopyPixelData(target_format, image_mat->cols, image_mat->rows, data_image,
ImageFrame::kDefaultAlignmentBoundary); ImageFrame::kDefaultAlignmentBoundary);
if (cc->Outputs().HasTag(kImageFrameTag)) if (cc->Outputs().HasTag(tag))
{ {
cc->Outputs() cc->Outputs()
.Tag(kImageFrameTag) .Tag(tag)
.Add(output_frame.release(), cc->InputTimestamp()); .Add(output_frame.release(), cc->InputTimestamp());
} }
@ -410,9 +447,8 @@ namespace mediapipe
else else
{ {
image_mat = absl::make_unique<cv::Mat>( image_mat = absl::make_unique<cv::Mat>(
150, 150, CV_8UC4, 1920, 1080, CV_8UC4,
cv::Scalar(255, 255, cv::Scalar::all(255));
255));
*target_format = ImageFormat::SRGBA; *target_format = ImageFormat::SRGBA;
} }
@ -424,8 +460,6 @@ namespace mediapipe
ImageFormat::Format &target_format, ImageFormat::Format &target_format,
std::vector<std::vector<cv::Point2f>> &lms_out) std::vector<std::vector<cv::Point2f>> &lms_out)
{ {
std::vector<cv::Point2f> kps, landmarks;
if (cc->Inputs().HasTag(kNormLandmarksTag)) if (cc->Inputs().HasTag(kNormLandmarksTag))
{ {
const std::vector<NormalizedLandmarkList> &landmarkslist = const std::vector<NormalizedLandmarkList> &landmarkslist =
@ -434,11 +468,12 @@ namespace mediapipe
std::vector<cv::Point2f> point_array; std::vector<cv::Point2f> point_array;
for (const auto &face : landmarkslist) for (const auto &face : landmarkslist)
{ {
std::vector<cv::Point2f> landmarks = {};
for (const auto &[key, value] : index_dict) for (const auto &[key, value] : index_dict)
{ {
std::vector<cv::Point2f> kps = {};
for (auto order : value) for (auto order : value)
{ {
const NormalizedLandmark &landmark = face.landmark(order); const NormalizedLandmark &landmark = face.landmark(order);
if (!IsLandmarkVisibleAndPresent<NormalizedLandmark>( if (!IsLandmarkVisibleAndPresent<NormalizedLandmark>(
@ -449,11 +484,13 @@ namespace mediapipe
continue; continue;
} }
const auto &size =
cc->Inputs().Tag(kSizeTag).Get<std::pair<int, int>>();
const auto &point = landmark; const auto &point = landmark;
int x = -1; int x = -1;
int y = -1; int y = -1;
CHECK(NormalizedtoPixelCoordinates(point.x(), point.y(), image_width_, CHECK(NormalizedtoPixelCoordinates(point.x(), point.y(), size.first,
image_height_, &x, &y)); size.second, &x, &y));
kps.push_back(cv::Point2f(x, y)); kps.push_back(cv::Point2f(x, y));
} }
@ -461,12 +498,29 @@ namespace mediapipe
cv::reduce(kps, mean, 1, CV_REDUCE_AVG, CV_32F); cv::reduce(kps, mean, 1, CV_REDUCE_AVG, CV_32F);
landmarks.push_back({mean.at<float>(0, 0), mean.at<float>(0, 1)}); landmarks.push_back({mean.at<float>(0, 0), mean.at<float>(0, 1)});
kps.clear();
} }
lms_out.push_back(landmarks); lms_out.push_back(landmarks);
}
if (cc->Outputs().HasTag(kLmMaskTag))
{
std::vector<cv::Point> kpsint = {};
for (auto &ix : FACEMESH_FACE_OVAL)
{
auto i = ix.x;
landmarks.clear(); const NormalizedLandmark &landmark = landmarkslist[0].landmark(i);
const auto &point = landmark;
int x = -1;
int y = -1;
CHECK(NormalizedtoPixelCoordinates(point.x(), point.y(), image_width_,
image_height_, &x, &y));
kpsint.push_back(cv::Point(x, y));
}
std::vector<std::vector<cv::Point>> pts;
pts.push_back(kpsint);
lm_mask = cv::Mat::zeros(image_mat->size(), CV_32FC1);
cv::fillPoly(lm_mask, pts, cv::Scalar::all(1), cv::LINE_AA);
} }
} }
@ -478,6 +532,8 @@ namespace mediapipe
cv::Mat target_lm, cv::Size size, cv::Mat target_lm, cv::Size size,
float extend, std::tuple<float, float, float, float> roi) float extend, std::tuple<float, float, float, float> roi)
{ {
cv::Mat mat_image_ = *image_mat.get();
cv::Mat source, target; cv::Mat source, target;
source_lm.convertTo(source, CV_32F); source_lm.convertTo(source, CV_32F);
target_lm.convertTo(target, CV_32F); target_lm.convertTo(target, CV_32F);

View File

@ -0,0 +1,27 @@
// Copyright 2019 The MediaPipe Authors.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
syntax = "proto2";
package mediapipe;
import "mediapipe/framework/calculator.proto";
message FastUtilsCalculatorOptions {
extend CalculatorOptions {
optional FastUtilsCalculatorOptions ext = 251431399;
}
// Change color and size of rendered landmarks based on its z value.
optional bool back_to_image = 1 [default = false];
}

View File

@ -893,3 +893,34 @@ cc_library(
}), }),
alwayslink = 1, alwayslink = 1,
) )
cc_library(
name = "tensors_to_image_calculator",
srcs = ["tensors_to_image_calculator.cc"],
copts = select({
"//mediapipe:apple": [
"-x objective-c++",
"-fobjc-arc", # enable reference-counting
],
"//conditions:default": [],
}),
visibility = ["//visibility:public"],
deps = [
"@com_google_absl//absl/strings:str_format",
"@com_google_absl//absl/strings",
"@com_google_absl//absl/types:span",
"//mediapipe/framework/formats:image",
"//mediapipe/framework/formats:image_frame",
"//mediapipe/framework/formats:image_opencv",
"//mediapipe/framework/formats:tensor",
"//mediapipe/framework/port:opencv_imgproc",
"//mediapipe/framework/port:ret_check",
"//mediapipe/framework:calculator_context",
"//mediapipe/framework:calculator_framework",
"//mediapipe/framework:port",
"//mediapipe/util:resource_util",
"@org_tensorflow//tensorflow/lite:framework",
"//mediapipe/framework/port:statusor",
],
alwayslink = 1,
)

View File

@ -0,0 +1,201 @@
// Copyright 2021 The MediaPipe Authors.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#include <vector>
#include <iostream>
#include "absl/strings/str_format.h"
#include "absl/types/span.h"
#include "mediapipe/framework/calculator_context.h"
#include "mediapipe/framework/calculator_framework.h"
#include "mediapipe/framework/formats/image.h"
#include "mediapipe/framework/formats/image_opencv.h"
#include "mediapipe/framework/formats/tensor.h"
#include "mediapipe/framework/port.h"
#include "mediapipe/framework/port/opencv_imgproc_inc.h"
#include "mediapipe/framework/port/ret_check.h"
#include "mediapipe/framework/port/statusor.h"
#include "mediapipe/util/resource_util.h"
#include "tensorflow/lite/interpreter.h"
namespace
{
constexpr char kTensorsTag[] = "TENSORS";
constexpr char kOutputSizeTag[] = "OUTPUT_SIZE";
constexpr char kImageTag[] = "IMAGE";
absl::StatusOr<std::tuple<int, int, int>> GetHwcFromDims(
const std::vector<int> &dims)
{
if (dims.size() == 3)
{
return std::make_tuple(dims[0], dims[1], dims[2]);
}
else if (dims.size() == 4)
{
// BHWC format check B == 1
RET_CHECK_EQ(1, dims[0]) << "Expected batch to be 1 for BHWC heatmap";
return std::make_tuple(dims[1], dims[2], dims[3]);
}
else
{
RET_CHECK(false) << "Invalid shape for segmentation tensor " << dims.size();
}
}
} // namespace
namespace mediapipe
{
// Converts Tensors from a tflite segmentation model to an image.
//
// Performs optional upscale to OUTPUT_SIZE dimensions if provided,
// otherwise the image is the same size as input tensor.
//
//
//
// Inputs:
// One of the following TENSORS tags:
// TENSORS: Vector of Tensor,
// The tensor dimensions are specified in this calculator's options.
// OUTPUT_SIZE(optional): std::pair<int, int>,
// If provided, the size to upscale mask to.
//
// Output:
// IMAGE: An Image output, RGBA.
//
//
// Usage example:
// node {
// calculator: "TensorsToImageCalculator"
// input_stream: "TENSORS:tensors"
// input_stream: "OUTPUT_SIZE:size"
// output_stream: "IMAGE:image"
// }
//
// TODO Refactor and add support for other backends/platforms.
//
class TensorsToImageCalculator : public CalculatorBase
{
public:
static absl::Status GetContract(CalculatorContract *cc);
absl::Status Open(CalculatorContext *cc) override;
absl::Status Process(CalculatorContext *cc) override;
absl::Status Close(CalculatorContext *cc) override;
private:
absl::Status ProcessCpu(CalculatorContext *cc);
};
REGISTER_CALCULATOR(TensorsToImageCalculator);
// static
absl::Status TensorsToImageCalculator::GetContract(
CalculatorContract *cc)
{
RET_CHECK(!cc->Inputs().GetTags().empty());
RET_CHECK(!cc->Outputs().GetTags().empty());
// Inputs.
cc->Inputs().Tag(kTensorsTag).Set<std::vector<Tensor>>();
if (cc->Inputs().HasTag(kOutputSizeTag))
{
cc->Inputs().Tag(kOutputSizeTag).Set<std::pair<int, int>>();
}
// Outputs.
cc->Outputs().Tag(kImageTag).Set<Image>();
return absl::OkStatus();
}
absl::Status TensorsToImageCalculator::Open(CalculatorContext *cc)
{
cc->SetOffset(TimestampDiff(0));
return absl::OkStatus();
}
absl::Status TensorsToImageCalculator::Process(CalculatorContext *cc)
{
if (cc->Inputs().Tag(kTensorsTag).IsEmpty())
{
return absl::OkStatus();
}
const auto &input_tensors =
cc->Inputs().Tag(kTensorsTag).Get<std::vector<Tensor>>();
MP_RETURN_IF_ERROR(ProcessCpu(cc));
return absl::OkStatus();
}
absl::Status TensorsToImageCalculator::Close(CalculatorContext *cc)
{
return absl::OkStatus();
}
absl::Status TensorsToImageCalculator::ProcessCpu(
CalculatorContext *cc)
{
// Get input streams, and dimensions.
const auto &input_tensors =
cc->Inputs().Tag(kTensorsTag).Get<std::vector<Tensor>>();
ASSIGN_OR_RETURN(auto hwc, GetHwcFromDims(input_tensors[0].shape().dims));
auto [tensor_height, tensor_width, tensor_channels] = hwc;
int output_width = tensor_width, output_height = tensor_height;
if (cc->Inputs().HasTag(kOutputSizeTag))
{
const auto &size =
cc->Inputs().Tag(kOutputSizeTag).Get<std::pair<int, int>>();
output_width = size.first;
output_height = size.second;
}
cv::Mat image_mat(cv::Size(tensor_width, tensor_height), CV_32FC1);
// Wrap input tensor.
auto raw_input_tensor = &input_tensors[0];
auto raw_input_view = raw_input_tensor->GetCpuReadView();
const float *raw_input_data = raw_input_view.buffer<float>();
cv::Mat tensor_mat(cv::Size(tensor_width, tensor_height),
CV_MAKETYPE(CV_32F, tensor_channels),
const_cast<float *>(raw_input_data));
std::vector<cv::Mat> channels(4);
cv::split(tensor_mat, channels);
for (auto ch : channels)
ch = (ch + 1) * 127.5;
cv::merge(channels, tensor_mat);
cv::convertScaleAbs(tensor_mat, tensor_mat);
// Send out image as CPU packet.
std::shared_ptr<ImageFrame> image_frame = std::make_shared<ImageFrame>(
ImageFormat::SRGB, output_width, output_height);
std::unique_ptr<Image> output_image = absl::make_unique<Image>(image_frame);
auto output_mat = formats::MatView(output_image.get());
// Upsample image into output.
cv::resize(tensor_mat, *output_mat,
cv::Size(output_width, output_height));
cc->Outputs().Tag(kImageTag).Add(output_image.release(), cc->InputTimestamp());
return absl::OkStatus();
}
} // namespace mediapipe

View File

@ -14,7 +14,7 @@ profiler_config {
trace_enabled: true trace_enabled: true
enable_profiler: true enable_profiler: true
trace_log_interval_count: 200 trace_log_interval_count: 200
trace_log_path: "/Users/alena/Workdir/mediapipe/logs/beauty/" trace_log_path: "/home/mslight/Work/clone/mediapipe/mediapipe/logs/beauty/"
} }
# Throttles the images flowing downstream for flow control. It passes through # Throttles the images flowing downstream for flow control. It passes through

View File

@ -19,7 +19,6 @@
#include <string> #include <string>
#include <iostream> #include <iostream>
#include <fstream> #include <fstream>
//#include <android/log.h>
#include <memory> #include <memory>
#include "Tensor.h" #include "Tensor.h"

View File

@ -16,18 +16,6 @@ licenses(["notice"])
package(default_visibility = ["//visibility:public"]) package(default_visibility = ["//visibility:public"])
load("//mediapipe/framework:encode_binary_proto.bzl", "encode_binary_proto")
encode_binary_proto(
name = "triangles",
input = "triangles.pbtxt",
message_type = "mediapipe.face_geometry.Mesh3d",
output = "triangles.binarypb",
deps = [
"//mediapipe/modules/face_geometry/protos:mesh_3d_proto",
],
)
exports_files( exports_files(
srcs = glob(["**"]), srcs = glob(["**"]),
) )

View File

@ -29,14 +29,16 @@ cc_library(
"//mediapipe/calculators/tensor:tensor_converter_calculator", "//mediapipe/calculators/tensor:tensor_converter_calculator",
"//mediapipe/calculators/tensor:inference_calculator", "//mediapipe/calculators/tensor:inference_calculator",
"//mediapipe/calculators/image:image_transformation_calculator", "//mediapipe/calculators/image:image_transformation_calculator",
"//mediapipe/calculators/tensor:tensors_to_segmentation_calculator",
"//mediapipe/calculators/util:to_image_calculator", "//mediapipe/calculators/util:to_image_calculator",
"//mediapipe/calculators/util:from_image_calculator", "//mediapipe/calculators/util:from_image_calculator",
"//mediapipe/calculators/image:image_properties_calculator", "//mediapipe/calculators/image:image_properties_calculator",
"//mediapipe/modules/face_landmark:face_landmark_front_gpu", "//mediapipe/modules/face_landmark:face_landmark_front_gpu",
"//mediapipe/calculators/image_style:apply_mask_calculator",
"//mediapipe/calculators/image_style:fast_utils_calculator", "//mediapipe/calculators/image_style:fast_utils_calculator",
"//mediapipe/gpu:gpu_buffer_to_image_frame_calculator", "//mediapipe/gpu:gpu_buffer_to_image_frame_calculator",
"//mediapipe/gpu:image_frame_to_gpu_buffer_calculator",
"//mediapipe/calculators/core:constant_side_packet_calculator", "//mediapipe/calculators/core:constant_side_packet_calculator",
"//mediapipe/calculators/tensor:tensors_to_image_calculator",
], ],
) )
@ -48,11 +50,12 @@ cc_library(
"//mediapipe/calculators/image:image_transformation_calculator", "//mediapipe/calculators/image:image_transformation_calculator",
"//mediapipe/calculators/tensor:inference_calculator", "//mediapipe/calculators/tensor:inference_calculator",
"//mediapipe/calculators/tensor:tensor_converter_calculator", "//mediapipe/calculators/tensor:tensor_converter_calculator",
"//mediapipe/calculators/tensor:tensors_to_segmentation_calculator", "//mediapipe/calculators/tensor:tensors_to_image_calculator",
"//mediapipe/calculators/util:to_image_calculator", "//mediapipe/calculators/util:to_image_calculator",
"//mediapipe/calculators/util:from_image_calculator", "//mediapipe/calculators/util:from_image_calculator",
"//mediapipe/modules/face_landmark:face_landmark_front_cpu", "//mediapipe/modules/face_landmark:face_landmark_front_cpu",
"//mediapipe/calculators/image_style:fast_utils_calculator", "//mediapipe/calculators/image_style:fast_utils_calculator",
"//mediapipe/calculators/image_style:apply_mask_calculator",
"//mediapipe/calculators/image:image_properties_calculator", "//mediapipe/calculators/image:image_properties_calculator",
"//mediapipe/calculators/core:constant_side_packet_calculator", "//mediapipe/calculators/core:constant_side_packet_calculator",
], ],

View File

@ -17,6 +17,12 @@ node {
output_stream: "throttled_input_video" output_stream: "throttled_input_video"
} }
node {
calculator: "ImagePropertiesCalculator"
input_stream: "IMAGE:input_video"
output_stream: "SIZE:original_size"
}
# Defines side packets for further use in the graph. # Defines side packets for further use in the graph.
node { node {
@ -44,7 +50,14 @@ node {
calculator: "FastUtilsCalculator" calculator: "FastUtilsCalculator"
input_stream: "NORM_LANDMARKS:multi_face_landmarks" input_stream: "NORM_LANDMARKS:multi_face_landmarks"
input_stream: "IMAGE:throttled_input_video" input_stream: "IMAGE:throttled_input_video"
input_stream: "SIZE:original_size"
output_stream: "IMAGE:out_image_frame" output_stream: "IMAGE:out_image_frame"
output_stream: "LM_MASK:lm_mask"
options {
[mediapipe.FastUtilsCalculatorOptions.ext] {
back_to_image: false
}
}
} }
node: { node: {
@ -83,26 +96,91 @@ node {
} }
node { node {
calculator: "ImagePropertiesCalculator" calculator: "TensorsToImageCalculator"
input_stream: "IMAGE:transformed_input_video" input_stream: "TENSORS:output_tensor"
output_stream: "SIZE:input_size" output_stream: "IMAGE:fake_image"
}
node{
calculator: "FromImageCalculator"
input_stream: "IMAGE:fake_image"
output_stream: "IMAGE_CPU:fake_image2"
}
node: {
calculator: "ImageTransformationCalculator"
input_stream: "IMAGE:input_video"
output_stream: "IMAGE:transformed_input_img"
node_options: {
[type.googleapis.com/mediapipe.ImageTransformationCalculatorOptions] {
output_width: 256
output_height: 256
}
}
} }
node { node {
calculator: "TensorsToSegmentationCalculator" calculator: "TensorConverterCalculator"
input_stream: "TENSORS:output_tensor" input_stream: "IMAGE:transformed_input_img"
input_stream: "OUTPUT_SIZE:input_size" output_stream: "TENSORS:input_tensor_img"
output_stream: "MASK:output" options: {
[mediapipe.TensorConverterCalculatorOptions.ext] {
zero_center: true
}
}
}
node {
calculator: "InferenceCalculator"
input_stream: "TENSORS:input_tensor_img"
output_stream: "TENSORS:output_tensor_img"
options: { options: {
[mediapipe.TensorsToSegmentationCalculatorOptions.ext] { [mediapipe.InferenceCalculatorOptions.ext] {
activation: NONE model_path: "mediapipe/models/model_float32.tflite"
delegate { xnnpack {} }
} }
} }
}
node {
calculator: "ImagePropertiesCalculator"
input_stream: "IMAGE:transformed_input_img"
output_stream: "SIZE:input_size_img"
}
node {
calculator: "TensorsToImageCalculator"
input_stream: "TENSORS:output_tensor_img"
input_stream: "OUTPUT_SIZE:input_size_img"
output_stream: "IMAGE:fake_bg2"
} }
node{ node{
calculator: "FromImageCalculator" calculator: "FromImageCalculator"
input_stream: "IMAGE:output" input_stream: "IMAGE:fake_bg2"
output_stream: "IMAGE_CPU:output_video" output_stream: "IMAGE_CPU:fake_bg"
} }
node {
calculator: "FastUtilsCalculator"
input_stream: "NORM_LANDMARKS:multi_face_landmarks"
input_stream: "IMAGE:fake_image2"
input_stream: "SIZE:original_size"
output_stream: "IMAGE:back_image"
options {
[mediapipe.FastUtilsCalculatorOptions.ext] {
back_to_image: true
}
}
}
node {
calculator: "ApplyMaskCalculator"
input_stream: "IMAGE:back_image"
input_stream: "FAKE_BG:fake_bg"
input_stream: "LM_MASK:lm_mask"
output_stream: "IMAGE:output_video"
}

View File

@ -36,6 +36,12 @@ node {
output_stream: "throttled_input_video_cpu" output_stream: "throttled_input_video_cpu"
} }
node {
calculator: "ImagePropertiesCalculator"
input_stream: "IMAGE_CPU:throttled_input_video_cpu"
output_stream: "SIZE:original_size"
}
# Subgraph that detects faces and corresponding landmarks. # Subgraph that detects faces and corresponding landmarks.
node { node {
calculator: "FaceLandmarkFrontGpu" calculator: "FaceLandmarkFrontGpu"
@ -48,26 +54,33 @@ node {
calculator: "FastUtilsCalculator" calculator: "FastUtilsCalculator"
input_stream: "NORM_LANDMARKS:multi_face_landmarks" input_stream: "NORM_LANDMARKS:multi_face_landmarks"
input_stream: "IMAGE:throttled_input_video_cpu" input_stream: "IMAGE:throttled_input_video_cpu"
input_stream: "SIZE:original_size"
output_stream: "IMAGE:out_image_frame" output_stream: "IMAGE:out_image_frame"
output_stream: "LM_MASK:lm_mask"
options {
[mediapipe.FastUtilsCalculatorOptions.ext] {
back_to_image: false
}
}
} }
#node: { node: {
# calculator: "ImageTransformationCalculator" calculator: "ImageTransformationCalculator"
# input_stream: "IMAGE:out_image_frame" input_stream: "IMAGE:out_image_frame"
# output_stream: "IMAGE:out_image_frame1" output_stream: "IMAGE:image_frame"
# node_options: { node_options: {
# [type.googleapis.com/mediapipe.ImageTransformationCalculatorOptions] { [type.googleapis.com/mediapipe.ImageTransformationCalculatorOptions] {
# output_width: 256 output_width: 256
# output_height: 256 output_height: 256
# } }
# } }
#} }
node { node {
calculator: "TensorConverterCalculator" calculator: "TensorConverterCalculator"
input_stream: "IMAGE:out_image_frame" input_stream: "IMAGE:image_frame"
output_stream: "TENSORS:input_tensors" output_stream: "TENSORS:input_tensor"
options: { options: {
[mediapipe.TensorConverterCalculatorOptions.ext] { [mediapipe.TensorConverterCalculatorOptions.ext] {
zero_center: true zero_center: true
@ -75,35 +88,109 @@ node {
} }
} }
#node {
# calculator: "InferenceCalculator"
# input_stream: "TENSORS:input_tensors"
# output_stream: "TENSORS:output_tensors"
# options: {
# [mediapipe.InferenceCalculatorOptions.ext] {
# model_path:"mediapipe/models/model_float32.tflite"
# delegate { gpu {} }
# }
# }
#}
# Processes the output tensors into a segmentation mask that has the same size
# as the input image into the graph.
node { node {
calculator: "TensorsToSegmentationCalculator" calculator: "InferenceCalculator"
input_stream: "TENSORS:input_tensors" input_stream: "TENSORS:input_tensor"
#input_stream: "OUTPUT_SIZE:input_size" output_stream: "TENSORS:output_tensor"
output_stream: "MASK:mask_image"
options: { options: {
[mediapipe.TensorsToSegmentationCalculatorOptions.ext] { [mediapipe.InferenceCalculatorOptions.ext] {
activation: NONE model_path:"mediapipe/models/model_float32.tflite"
gpu_origin: TOP_LEFT delegate { gpu {} }
} }
} }
}
node {
calculator: "TensorsToImageCalculator"
input_stream: "TENSORS:output_tensor"
output_stream: "IMAGE:fake_image"
} }
node: { node: {
calculator: "FromImageCalculator" calculator: "FromImageCalculator"
input_stream: "IMAGE:mask_image" input_stream: "IMAGE:fake_image"
output_stream: "IMAGE_GPU:output_video" output_stream: "IMAGE_CPU:cpu_fake_image"
}
node: {
calculator: "ImageTransformationCalculator"
input_stream: "IMAGE:throttled_input_video_cpu"
output_stream: "IMAGE:transformed_input_img"
node_options: {
[type.googleapis.com/mediapipe.ImageTransformationCalculatorOptions] {
output_width: 256
output_height: 256
}
}
}
node {
calculator: "TensorConverterCalculator"
input_stream: "IMAGE:transformed_input_img"
output_stream: "TENSORS:input_tensor_img"
options: {
[mediapipe.TensorConverterCalculatorOptions.ext] {
zero_center: true
}
}
}
node {
calculator: "InferenceCalculator"
input_stream: "TENSORS:input_tensor_img"
output_stream: "TENSORS:output_tensor_img"
options: {
[mediapipe.InferenceCalculatorOptions.ext] {
model_path: "mediapipe/models/model_float32.tflite"
delegate { xnnpack {} }
}
}
}
node {
calculator: "ImagePropertiesCalculator"
input_stream: "IMAGE_CPU:transformed_input_img"
output_stream: "SIZE:input_size_img"
}
node {
calculator: "TensorsToImageCalculator"
input_stream: "TENSORS:output_tensor_img"
input_stream: "OUTPUT_SIZE:input_size_img"
output_stream: "IMAGE:fake_bg2"
}
node{
calculator: "FromImageCalculator"
input_stream: "IMAGE:fake_bg2"
output_stream: "IMAGE_CPU:fake_bg"
}
node {
calculator: "FastUtilsCalculator"
input_stream: "NORM_LANDMARKS:multi_face_landmarks"
input_stream: "IMAGE:cpu_fake_image"
input_stream: "SIZE:original_size"
output_stream: "IMAGE:back_image"
options {
[mediapipe.FastUtilsCalculatorOptions.ext] {
back_to_image: true
}
}
}
node {
calculator: "ApplyMaskCalculator"
input_stream: "IMAGE:back_image"
input_stream: "FAKE_BG:fake_bg"
input_stream: "LM_MASK:lm_mask"
output_stream: "IMAGE:out_image"
}
# Defines side packets for further use in the graph.
node {
calculator: "ImageFrameToGpuBufferCalculator"
input_stream: "out_image"
output_stream: "output_video"
} }