48bcbb115f
GitOrigin-RevId: 50714fe28298d7b707eff7304547d89d6ec34a54
834 lines
30 KiB
C++
834 lines
30 KiB
C++
// Copyright 2019 The MediaPipe Authors.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
#include "mediapipe/util/audio_decoder.h"
|
|
|
|
#include <algorithm>
|
|
#include <cstdint> // required by avutil.h
|
|
#include <cstdlib>
|
|
#include <memory>
|
|
#include <string>
|
|
|
|
#include "Eigen/Core"
|
|
#include "absl/base/internal/endian.h"
|
|
#include "absl/strings/numbers.h"
|
|
#include "absl/strings/str_cat.h"
|
|
#include "absl/strings/substitute.h"
|
|
#include "absl/time/time.h"
|
|
#include "mediapipe/framework/deps/cleanup.h"
|
|
#include "mediapipe/framework/formats/matrix.h"
|
|
#include "mediapipe/framework/port/map_util.h"
|
|
#include "mediapipe/framework/port/ret_check.h"
|
|
#include "mediapipe/framework/port/status.h"
|
|
#include "mediapipe/framework/tool/status_util.h"
|
|
|
|
extern "C" {
|
|
#include "libavcodec/avcodec.h"
|
|
#include "libavformat/avformat.h"
|
|
#include "libavutil/avutil.h"
|
|
#include "libavutil/mem.h"
|
|
#include "libavutil/samplefmt.h"
|
|
}
|
|
|
|
DEFINE_int64(media_decoder_allowed_audio_gap_merge, 5,
|
|
"The time gap forwards or backwards in the audio to ignore. "
|
|
"Timestamps in media files are restricted by the container format "
|
|
"and stream codec and are invariably not accurate to exact sample "
|
|
"numbers. If the discrepency between time based on counting "
|
|
"samples and based on the container timestamps grows beyond this "
|
|
"value it will be reset to the value in the audio stream and "
|
|
"counting based on samples will resume.");
|
|
|
|
namespace mediapipe {
|
|
|
|
// MPEG PTS max value + 1, used to correct for PTS rollover. Unit is PTS ticks.
|
|
const int64 kMpegPtsEpoch = 1LL << 33;
|
|
// Maximum PTS change between frames. Larger changes are considered to indicate
|
|
// the MPEG PTS has rolled over. Unit is PTS ticks.
|
|
const int64 kMpegPtsMaxDelta = kMpegPtsEpoch / 2;
|
|
|
|
// BasePacketProcessor
|
|
namespace {
|
|
|
|
inline std::string TimestampToString(int64 timestamp) {
|
|
if (timestamp == AV_NOPTS_VALUE) {
|
|
return "NOPTS";
|
|
}
|
|
return absl::StrCat(timestamp);
|
|
}
|
|
|
|
float Uint32ToFloat(uint32 raw_value) {
|
|
float value;
|
|
memcpy(&value, &raw_value, 4);
|
|
return value;
|
|
}
|
|
|
|
std::string AvErrorToString(int error) {
|
|
if (error >= 0) {
|
|
return absl::StrCat("Not an error (", error, ")");
|
|
}
|
|
|
|
switch (error) {
|
|
case AVERROR(EINVAL):
|
|
return "AVERROR(EINVAL) - unknown error or invalid data";
|
|
case AVERROR(EIO):
|
|
return "AVERROR(EIO) - I/O error";
|
|
case AVERROR(EDOM):
|
|
return "AVERROR(EDOM) - Number syntax expected in filename.";
|
|
case AVERROR(ENOMEM):
|
|
return "AVERROR(ENOMEM) - not enough memory";
|
|
case AVERROR(EILSEQ):
|
|
return "AVERROR(EILSEQ) - unknown format";
|
|
case AVERROR(ENOSYS):
|
|
return "AVERROR(ENOSYS) - Operation not supported.";
|
|
case AVERROR(ENOENT):
|
|
return "AVERROR(ENOENT) - No such file or directory.";
|
|
case AVERROR(EPIPE):
|
|
return "AVERROR(EPIPE) - End of file.";
|
|
case AVERROR_BSF_NOT_FOUND:
|
|
return "AVERROR_BSF_NOT_FOUND - Bitstream filter not found.";
|
|
case AVERROR_BUG:
|
|
return "AVERROR_BUG - Internal bug, should not have happened.";
|
|
case AVERROR_BUG2:
|
|
return "AVERROR_BUG2 - Internal bug, should not have happened.";
|
|
case AVERROR_BUFFER_TOO_SMALL:
|
|
return "AVERROR_BUFFER_TOO_SMALL - Buffer too small.";
|
|
case AVERROR_DECODER_NOT_FOUND:
|
|
return "AVERROR_DECODER_NOT_FOUND - Decoder not found.";
|
|
case AVERROR_DEMUXER_NOT_FOUND:
|
|
return "AVERROR_DEMUXER_NOT_FOUND - Demuxer not found.";
|
|
case AVERROR_ENCODER_NOT_FOUND:
|
|
return "AVERROR_ENCODER_NOT_FOUND - Encoder not found.";
|
|
case AVERROR_EOF:
|
|
return "AVERROR_EOF - End of file.";
|
|
case AVERROR_EXIT:
|
|
return "AVERROR_EXIT - Immediate exit was requested.";
|
|
case AVERROR_EXTERNAL:
|
|
return "AVERROR_EXTERNAL - Generic error in an external library.";
|
|
case AVERROR_FILTER_NOT_FOUND:
|
|
return "AVERROR_FILTER_NOT_FOUND - Filter not found.";
|
|
case AVERROR_INVALIDDATA:
|
|
return "AVERROR_INVALIDDATA - Invalid data found when processing input.";
|
|
case AVERROR_MUXER_NOT_FOUND:
|
|
return "AVERROR_MUXER_NOT_FOUND - Muxer not found.";
|
|
case AVERROR_OPTION_NOT_FOUND:
|
|
return "AVERROR_OPTION_NOT_FOUND - Option not found.";
|
|
case AVERROR_PATCHWELCOME:
|
|
return "AVERROR_PATCHWELCOME - Not yet implemented in FFmpeg, "
|
|
"patches welcome.";
|
|
case AVERROR_PROTOCOL_NOT_FOUND:
|
|
return "AVERROR_PROTOCOL_NOT_FOUND - Protocol not found.";
|
|
case AVERROR_STREAM_NOT_FOUND:
|
|
return "AVERROR_STREAM_NOT_FOUND - Stream not found.";
|
|
case AVERROR_EXPERIMENTAL:
|
|
return "AVERROR_EXPERIMENTAL - Requested feature is flagged "
|
|
"experimental.";
|
|
case AVERROR_INPUT_CHANGED:
|
|
return "AVERROR_INPUT_CHANGED - Input changed between calls.";
|
|
case AVERROR_OUTPUT_CHANGED:
|
|
return "AVERROR_OUTPUT_CHANGED - Output changed between calls.";
|
|
default:
|
|
// FALLTHRU
|
|
{}
|
|
}
|
|
|
|
char buf[AV_ERROR_MAX_STRING_SIZE];
|
|
if (av_strerror(error, buf, sizeof(buf)) == 0) {
|
|
return absl::StrCat("AVERROR(", error, ") - ", buf);
|
|
}
|
|
|
|
return absl::StrCat("Unknown AVERROR number ", error);
|
|
}
|
|
|
|
// Send a packet to the decoder.
|
|
mediapipe::Status SendPacket(const AVPacket& packet,
|
|
AVCodecContext* avcodec_ctx) {
|
|
const int error = avcodec_send_packet(avcodec_ctx, &packet);
|
|
if (error != 0 && error != AVERROR_EOF) {
|
|
// Not consider AVERROR_EOF as an error because it can happen when more
|
|
// than 1 flush packet is sent.
|
|
return UnknownError(absl::StrCat("Failed to send packet: error=", error,
|
|
" (", AvErrorToString(error),
|
|
"). Packet size: ", packet.size));
|
|
}
|
|
return mediapipe::OkStatus();
|
|
}
|
|
|
|
// Receive a decoded frame from the decoder.
|
|
mediapipe::Status ReceiveFrame(AVCodecContext* avcodec_ctx, AVFrame* frame,
|
|
bool* received) {
|
|
const int error = avcodec_receive_frame(avcodec_ctx, frame);
|
|
*received = error == 0;
|
|
if (error != 0 && error != AVERROR_EOF && error != AVERROR(EAGAIN)) {
|
|
// Not consider AVERROR_EOF as an error because it can happen after a
|
|
// flush, and AVERROR(EAGAIN) because it happens when there's no (more)
|
|
// frame to be received from this packet.
|
|
return UnknownError(absl::StrCat(" Failed to receive frame: error=", error,
|
|
" (", AvErrorToString(error), ")."));
|
|
}
|
|
return mediapipe::OkStatus();
|
|
}
|
|
|
|
mediapipe::Status LogStatus(const mediapipe::Status& status,
|
|
const AVCodecContext& avcodec_ctx,
|
|
const AVPacket& packet,
|
|
bool always_return_ok_status) {
|
|
if (status.ok()) {
|
|
return status;
|
|
}
|
|
|
|
VLOG(3) << "Failed to process packet:"
|
|
<< " media_type:"
|
|
<< (avcodec_ctx.codec_type == AVMEDIA_TYPE_VIDEO ? "video" : "audio")
|
|
<< " codec_id:" << avcodec_ctx.codec_id
|
|
<< " frame_number:" << avcodec_ctx.frame_number
|
|
<< " pts:" << TimestampToString(packet.pts)
|
|
<< " dts:" << TimestampToString(packet.dts) << " size:" << packet.size
|
|
<< (packet.flags & AV_PKT_FLAG_KEY ? " Key Frame." : "");
|
|
|
|
if (always_return_ok_status) {
|
|
LOG(WARNING) << status.message();
|
|
return mediapipe::OkStatus();
|
|
} else {
|
|
return status;
|
|
}
|
|
}
|
|
|
|
class AVPacketDeleter {
|
|
public:
|
|
void operator()(void* x) const {
|
|
AVPacket* packet = static_cast<AVPacket*>(x);
|
|
if (packet) {
|
|
av_free_packet(packet);
|
|
delete packet;
|
|
}
|
|
}
|
|
};
|
|
|
|
} // namespace
|
|
|
|
BasePacketProcessor::BasePacketProcessor()
|
|
: decoded_frame_(av_frame_alloc()),
|
|
source_time_base_{0, 0},
|
|
output_time_base_{1, 1000000},
|
|
source_frame_rate_{0, 0} {}
|
|
|
|
BasePacketProcessor::~BasePacketProcessor() { Close(); }
|
|
|
|
bool BasePacketProcessor::HasData() { return !buffer_.empty(); }
|
|
|
|
mediapipe::Status BasePacketProcessor::GetData(Packet* packet) {
|
|
CHECK(packet);
|
|
CHECK(!buffer_.empty());
|
|
*packet = buffer_.front();
|
|
buffer_.pop_front();
|
|
|
|
return mediapipe::OkStatus();
|
|
}
|
|
|
|
mediapipe::Status BasePacketProcessor::Flush() {
|
|
int64 last_num_frames_processed;
|
|
do {
|
|
std::unique_ptr<AVPacket, AVPacketDeleter> av_packet(new AVPacket());
|
|
av_init_packet(av_packet.get());
|
|
av_packet->size = 0;
|
|
av_packet->data = nullptr;
|
|
av_packet->stream_index = id_;
|
|
|
|
last_num_frames_processed = num_frames_processed_;
|
|
// ProcessPacket increments num_frames_processed_ if it is able to
|
|
// decode a frame. Not being able to decode a frame while being
|
|
// flushed signals that the codec is completely done.
|
|
MP_RETURN_IF_ERROR(ProcessPacket(av_packet.get()));
|
|
} while (last_num_frames_processed != num_frames_processed_);
|
|
|
|
flushed_ = true;
|
|
return mediapipe::OkStatus();
|
|
}
|
|
|
|
void BasePacketProcessor::Close() {
|
|
if (avcodec_ctx_) {
|
|
if (avcodec_ctx_->codec) {
|
|
avcodec_close(avcodec_ctx_);
|
|
av_free(avcodec_ctx_);
|
|
}
|
|
avcodec_ctx_ = nullptr;
|
|
}
|
|
if (avcodec_opts_) {
|
|
av_dict_free(&avcodec_opts_);
|
|
}
|
|
if (decoded_frame_) {
|
|
av_frame_free(&decoded_frame_);
|
|
}
|
|
}
|
|
|
|
mediapipe::Status BasePacketProcessor::Decode(const AVPacket& packet,
|
|
bool ignore_decode_failures) {
|
|
MP_RETURN_IF_ERROR(LogStatus(SendPacket(packet, avcodec_ctx_), *avcodec_ctx_,
|
|
packet, ignore_decode_failures));
|
|
while (true) {
|
|
bool received;
|
|
MP_RETURN_IF_ERROR(
|
|
LogStatus(ReceiveFrame(avcodec_ctx_, decoded_frame_, &received),
|
|
*avcodec_ctx_, packet, ignore_decode_failures));
|
|
if (received) {
|
|
// Successfully decoded a frame (i.e., received it from the decoder). Now
|
|
// further process it.
|
|
MP_RETURN_IF_ERROR(ProcessDecodedFrame(packet));
|
|
} else {
|
|
break;
|
|
}
|
|
}
|
|
return mediapipe::OkStatus();
|
|
}
|
|
|
|
int64 BasePacketProcessor::CorrectPtsForRollover(int64 media_pts) {
|
|
const int64 rollover_pts_media_bits = kMpegPtsEpoch - 1;
|
|
// Ensure PTS in range 0 ... kMpegPtsEpoch. This avoids errors from post
|
|
// decode PTS corrections that overflow the epoch range (while still yielding
|
|
// the correct result as long as the corrections do not exceed
|
|
// kMpegPtsMaxDelta).
|
|
media_pts &= rollover_pts_media_bits;
|
|
if (rollover_corrected_last_pts_ == AV_NOPTS_VALUE) {
|
|
// First seen PTS.
|
|
rollover_corrected_last_pts_ = media_pts;
|
|
} else {
|
|
int64 prev_media_pts =
|
|
rollover_corrected_last_pts_ & rollover_pts_media_bits;
|
|
int64 pts_step = media_pts - prev_media_pts;
|
|
if (pts_step > kMpegPtsMaxDelta) {
|
|
pts_step = pts_step - kMpegPtsEpoch;
|
|
} else if (pts_step < -kMpegPtsMaxDelta) {
|
|
pts_step = kMpegPtsEpoch + pts_step;
|
|
}
|
|
rollover_corrected_last_pts_ =
|
|
std::max((int64)0, rollover_corrected_last_pts_ + pts_step);
|
|
}
|
|
return rollover_corrected_last_pts_;
|
|
}
|
|
|
|
// AudioPacketProcessor
|
|
namespace {
|
|
|
|
// Converts a PCM_S16LE-encoded input sample to float between -1 and 1.
|
|
inline float PcmEncodedSampleToFloat(const char* data) {
|
|
static const float kMultiplier = 1.f / (1 << 15);
|
|
return absl::little_endian::Load16(data) * kMultiplier;
|
|
}
|
|
|
|
// Converts a PCM_S32LE-encoded input sample to float between -1 and 1.
|
|
inline float PcmEncodedSampleInt32ToFloat(const char* data) {
|
|
static constexpr float kMultiplier = 1.f / (1u << 31);
|
|
return absl::little_endian::Load32(data) * kMultiplier;
|
|
}
|
|
|
|
} // namespace
|
|
|
|
AudioPacketProcessor::AudioPacketProcessor(const AudioStreamOptions& options)
|
|
: sample_time_base_{0, 0}, options_(options) {
|
|
DCHECK(absl::little_endian::IsLittleEndian());
|
|
}
|
|
|
|
mediapipe::Status AudioPacketProcessor::Open(int id, AVStream* stream) {
|
|
id_ = id;
|
|
avcodec_ = avcodec_find_decoder(stream->codecpar->codec_id);
|
|
if (!avcodec_) {
|
|
return ::mediapipe::InvalidArgumentError("Failed to find codec");
|
|
}
|
|
avcodec_ctx_ = avcodec_alloc_context3(avcodec_);
|
|
avcodec_parameters_to_context(avcodec_ctx_, stream->codecpar);
|
|
if (avcodec_open2(avcodec_ctx_, avcodec_, &avcodec_opts_) < 0) {
|
|
return UnknownError("avcodec_open() failed.");
|
|
}
|
|
CHECK(avcodec_ctx_->codec);
|
|
|
|
source_time_base_ = stream->time_base;
|
|
source_frame_rate_ = stream->r_frame_rate;
|
|
last_frame_time_regression_detected_ = false;
|
|
|
|
MP_RETURN_IF_ERROR(ValidateSampleFormat());
|
|
bytes_per_sample_ = av_get_bytes_per_sample(avcodec_ctx_->sample_fmt);
|
|
num_channels_ = avcodec_ctx_->channels;
|
|
sample_rate_ = avcodec_ctx_->sample_rate;
|
|
|
|
if (num_channels_ <= 0) {
|
|
return UnknownError("num_channels must be strictly positive.");
|
|
}
|
|
if (sample_rate_ <= 0) {
|
|
return UnknownError("sample_rate must be strictly positive.");
|
|
}
|
|
|
|
sample_time_base_ = {1, static_cast<int>(sample_rate_)};
|
|
|
|
VLOG(0) << absl::Substitute(
|
|
"Opened audio stream (id: $0, channels: $1, sample rate: $2, time base: "
|
|
"$3/$4).",
|
|
id_, num_channels_, sample_rate_, source_time_base_.num,
|
|
source_time_base_.den);
|
|
|
|
return mediapipe::OkStatus();
|
|
}
|
|
|
|
mediapipe::Status AudioPacketProcessor::ValidateSampleFormat() {
|
|
switch (avcodec_ctx_->sample_fmt) {
|
|
case AV_SAMPLE_FMT_S16:
|
|
case AV_SAMPLE_FMT_S16P:
|
|
case AV_SAMPLE_FMT_S32:
|
|
case AV_SAMPLE_FMT_FLT:
|
|
case AV_SAMPLE_FMT_FLTP:
|
|
return mediapipe::OkStatus();
|
|
default:
|
|
return mediapipe::UnimplementedErrorBuilder(MEDIAPIPE_LOC)
|
|
<< "sample_fmt = " << avcodec_ctx_->sample_fmt;
|
|
}
|
|
}
|
|
|
|
int64 AudioPacketProcessor::SampleNumberToTimestamp(const int64 sample_number) {
|
|
return av_rescale_q(sample_number, sample_time_base_, source_time_base_);
|
|
}
|
|
|
|
int64 AudioPacketProcessor::TimestampToSampleNumber(const int64 timestamp) {
|
|
return av_rescale_q(timestamp, source_time_base_, sample_time_base_);
|
|
}
|
|
|
|
int64 AudioPacketProcessor::TimestampToMicroseconds(const int64 timestamp) {
|
|
return av_rescale_q(timestamp, source_time_base_, {1, 1000000});
|
|
}
|
|
|
|
int64 AudioPacketProcessor::SampleNumberToMicroseconds(
|
|
const int64 sample_number) {
|
|
return av_rescale_q(sample_number, sample_time_base_, {1, 1000000});
|
|
}
|
|
|
|
mediapipe::Status AudioPacketProcessor::ProcessPacket(AVPacket* packet) {
|
|
CHECK(packet);
|
|
if (flushed_) {
|
|
return UnknownError(
|
|
"ProcessPacket was called, but AudioPacketProcessor is already "
|
|
"finished.");
|
|
}
|
|
RET_CHECK_EQ(packet->stream_index, id_);
|
|
|
|
decoded_frame_->nb_samples = 0;
|
|
return Decode(*packet, options_.ignore_decode_failures());
|
|
}
|
|
|
|
mediapipe::Status AudioPacketProcessor::ProcessDecodedFrame(
|
|
const AVPacket& packet) {
|
|
RET_CHECK_EQ(decoded_frame_->channels, num_channels_);
|
|
int buf_size_bytes = av_samples_get_buffer_size(nullptr, num_channels_,
|
|
decoded_frame_->nb_samples,
|
|
avcodec_ctx_->sample_fmt, 1);
|
|
VLOG(3) << "Audio packet " << avcodec_ctx_->frame_number
|
|
<< " pts: " << TimestampToString(packet.pts)
|
|
<< " frame.pts:" << TimestampToString(decoded_frame_->pts)
|
|
<< " pkt_dts:" << TimestampToString(decoded_frame_->pkt_dts)
|
|
<< " dts:" << TimestampToString(packet.dts) << " size:" << packet.size
|
|
<< " decoded:" << buf_size_bytes;
|
|
uint8* const* data_ptr = decoded_frame_->data;
|
|
if (!data_ptr[0]) {
|
|
return UnknownError("No data in audio frame.");
|
|
}
|
|
if (decoded_frame_->pts != AV_NOPTS_VALUE) {
|
|
int64 pts = MaybeCorrectPtsForRollover(decoded_frame_->pts);
|
|
if (num_frames_processed_ == 0) {
|
|
expected_sample_number_ = TimestampToSampleNumber(pts);
|
|
}
|
|
|
|
const int64 expected_us =
|
|
SampleNumberToMicroseconds(expected_sample_number_);
|
|
const int64 actual_us = TimestampToMicroseconds(pts);
|
|
if (absl::Microseconds(std::abs(expected_us - actual_us)) >
|
|
absl::Seconds(FLAGS_media_decoder_allowed_audio_gap_merge)) {
|
|
LOG(ERROR) << "The expected time based on how many samples we have seen ("
|
|
<< expected_us
|
|
<< " microseconds) no longer matches the time based "
|
|
"on what the audio stream is telling us ("
|
|
<< actual_us
|
|
<< " microseconds). The difference is more than "
|
|
"--media_decoder_allowed_audio_gap_merge ("
|
|
<< absl::FormatDuration(absl::Seconds(
|
|
FLAGS_media_decoder_allowed_audio_gap_merge))
|
|
<< " microseconds). Resetting the timestamps to track what "
|
|
"the audio stream is telling us.";
|
|
expected_sample_number_ = TimestampToSampleNumber(pts);
|
|
}
|
|
}
|
|
|
|
MP_RETURN_IF_ERROR(AddAudioDataToBuffer(
|
|
Timestamp(av_rescale_q(expected_sample_number_, sample_time_base_,
|
|
output_time_base_)),
|
|
data_ptr, buf_size_bytes));
|
|
|
|
++num_frames_processed_;
|
|
return mediapipe::OkStatus();
|
|
}
|
|
|
|
mediapipe::Status AudioPacketProcessor::AddAudioDataToBuffer(
|
|
const Timestamp output_timestamp, uint8* const* raw_audio,
|
|
int buf_size_bytes) {
|
|
if (buf_size_bytes == 0) {
|
|
return mediapipe::OkStatus();
|
|
}
|
|
|
|
if (buf_size_bytes % (num_channels_ * bytes_per_sample_) != 0) {
|
|
return UnknownError("Buffer is not an integral number of samples.");
|
|
}
|
|
|
|
const int64 num_samples = buf_size_bytes / bytes_per_sample_ / num_channels_;
|
|
VLOG(3) << "Adding " << num_samples << " audio samples in " << num_channels_
|
|
<< " channels to output.";
|
|
auto current_frame = absl::make_unique<Matrix>(num_channels_, num_samples);
|
|
|
|
const char* sample_ptr = nullptr;
|
|
switch (avcodec_ctx_->sample_fmt) {
|
|
case AV_SAMPLE_FMT_S16:
|
|
sample_ptr = reinterpret_cast<const char*>(raw_audio[0]);
|
|
for (int64 sample_index = 0; sample_index < num_samples; ++sample_index) {
|
|
for (int channel = 0; channel < num_channels_; ++channel) {
|
|
(*current_frame)(channel, sample_index) =
|
|
PcmEncodedSampleToFloat(sample_ptr);
|
|
sample_ptr += bytes_per_sample_;
|
|
}
|
|
}
|
|
break;
|
|
case AV_SAMPLE_FMT_S32:
|
|
sample_ptr = reinterpret_cast<const char*>(raw_audio[0]);
|
|
for (int64 sample_index = 0; sample_index < num_samples; ++sample_index) {
|
|
for (int channel = 0; channel < num_channels_; ++channel) {
|
|
(*current_frame)(channel, sample_index) =
|
|
PcmEncodedSampleInt32ToFloat(sample_ptr);
|
|
sample_ptr += bytes_per_sample_;
|
|
}
|
|
}
|
|
break;
|
|
case AV_SAMPLE_FMT_FLT:
|
|
sample_ptr = reinterpret_cast<const char*>(raw_audio[0]);
|
|
for (int64 sample_index = 0; sample_index < num_samples; ++sample_index) {
|
|
for (int channel = 0; channel < num_channels_; ++channel) {
|
|
(*current_frame)(channel, sample_index) =
|
|
Uint32ToFloat(absl::little_endian::Load32(sample_ptr));
|
|
sample_ptr += bytes_per_sample_;
|
|
}
|
|
}
|
|
break;
|
|
case AV_SAMPLE_FMT_S16P:
|
|
for (int channel = 0; channel < num_channels_; ++channel) {
|
|
sample_ptr = reinterpret_cast<const char*>(raw_audio[channel]);
|
|
for (int64 sample_index = 0; sample_index < num_samples;
|
|
++sample_index) {
|
|
(*current_frame)(channel, sample_index) =
|
|
PcmEncodedSampleToFloat(sample_ptr);
|
|
sample_ptr += bytes_per_sample_;
|
|
}
|
|
}
|
|
break;
|
|
case AV_SAMPLE_FMT_FLTP:
|
|
for (int channel = 0; channel < num_channels_; ++channel) {
|
|
sample_ptr = reinterpret_cast<const char*>(raw_audio[channel]);
|
|
for (int64 sample_index = 0; sample_index < num_samples;
|
|
++sample_index) {
|
|
(*current_frame)(channel, sample_index) =
|
|
Uint32ToFloat(absl::little_endian::Load32(sample_ptr));
|
|
sample_ptr += bytes_per_sample_;
|
|
}
|
|
}
|
|
break;
|
|
default:
|
|
return mediapipe::UnimplementedErrorBuilder(MEDIAPIPE_LOC)
|
|
<< "sample_fmt = " << avcodec_ctx_->sample_fmt;
|
|
}
|
|
|
|
if (options_.output_regressing_timestamps() ||
|
|
last_timestamp_ == Timestamp::Unset() ||
|
|
output_timestamp > last_timestamp_) {
|
|
buffer_.push_back(Adopt(current_frame.release()).At(output_timestamp));
|
|
last_timestamp_ = output_timestamp;
|
|
if (last_frame_time_regression_detected_) {
|
|
last_frame_time_regression_detected_ = false;
|
|
LOG(INFO) << "Processor " << this << " resumed audio packet processing.";
|
|
}
|
|
} else if (!last_frame_time_regression_detected_) {
|
|
last_frame_time_regression_detected_ = true;
|
|
LOG(ERROR) << "Processor " << this
|
|
<< " is dropping an audio packet because the timestamps "
|
|
"regressed. Was "
|
|
<< last_timestamp_ << " but got " << output_timestamp;
|
|
}
|
|
expected_sample_number_ += num_samples;
|
|
|
|
return mediapipe::OkStatus();
|
|
}
|
|
|
|
mediapipe::Status AudioPacketProcessor::FillHeader(
|
|
TimeSeriesHeader* header) const {
|
|
CHECK(header);
|
|
header->set_sample_rate(sample_rate_);
|
|
header->set_num_channels(num_channels_);
|
|
return mediapipe::OkStatus();
|
|
}
|
|
|
|
int64 AudioPacketProcessor::MaybeCorrectPtsForRollover(int64 media_pts) {
|
|
return options_.correct_pts_for_rollover() ? CorrectPtsForRollover(media_pts)
|
|
: media_pts;
|
|
}
|
|
|
|
// AudioDecoder
|
|
AudioDecoder::AudioDecoder() { av_register_all(); }
|
|
|
|
AudioDecoder::~AudioDecoder() {
|
|
::mediapipe::Status status = Close();
|
|
if (!status.ok()) {
|
|
LOG(ERROR) << "Encountered error while closing media file: "
|
|
<< status.message();
|
|
}
|
|
}
|
|
|
|
::mediapipe::Status AudioDecoder::Initialize(
|
|
const std::string& input_file,
|
|
const mediapipe::AudioDecoderOptions options) {
|
|
if (options.audio_stream().empty()) {
|
|
return ::mediapipe::InvalidArgumentError(
|
|
"At least one audio_stream must be defined in AudioDecoderOptions");
|
|
}
|
|
std::map<int, int> stream_index_to_audio_options_index;
|
|
int options_index = 0;
|
|
for (const auto& audio_stream : options.audio_stream()) {
|
|
InsertIfNotPresent(&stream_index_to_audio_options_index,
|
|
audio_stream.stream_index(), options_index);
|
|
++options_index;
|
|
}
|
|
|
|
Cleanup<std::function<void()>> decoder_closer([this]() {
|
|
::mediapipe::Status status = Close();
|
|
if (!status.ok()) {
|
|
LOG(ERROR) << "Encountered error while closing media file: "
|
|
<< status.message();
|
|
}
|
|
});
|
|
|
|
avformat_ctx_ = avformat_alloc_context();
|
|
if (avformat_open_input(&avformat_ctx_, input_file.c_str(), NULL, NULL) < 0) {
|
|
return ::mediapipe::InvalidArgumentError(
|
|
absl::StrCat("Could not open file: ", input_file));
|
|
}
|
|
|
|
if (avformat_find_stream_info(avformat_ctx_, NULL) < 0) {
|
|
return ::mediapipe::InvalidArgumentError(absl::StrCat(
|
|
"Could not find stream information of file: ", input_file));
|
|
}
|
|
|
|
std::map<int, int> audio_options_index_to_stream_id;
|
|
for (int current_audio_index = 0, stream_id = 0;
|
|
stream_id < avformat_ctx_->nb_streams; ++stream_id) {
|
|
AVStream* stream = avformat_ctx_->streams[stream_id];
|
|
AVCodecParameters* dec_param = stream->codecpar;
|
|
switch (dec_param->codec_type) {
|
|
case AVMEDIA_TYPE_AUDIO: {
|
|
const int* options_index_ptr = FindOrNull(
|
|
stream_index_to_audio_options_index, current_audio_index);
|
|
if (options_index_ptr) {
|
|
std::unique_ptr<AudioPacketProcessor> processor =
|
|
absl::make_unique<AudioPacketProcessor>(
|
|
options.audio_stream(*options_index_ptr));
|
|
if (!ContainsKey(audio_processor_, stream_id)) {
|
|
LOG(INFO) << "Created audio processor " << processor.get()
|
|
<< " for file \"" << input_file << "\"";
|
|
} else {
|
|
LOG(ERROR) << "Stream " << stream_id
|
|
<< " already mapped to audio processor "
|
|
<< audio_processor_[stream_id].get();
|
|
}
|
|
|
|
MP_RETURN_IF_ERROR(processor->Open(stream_id, stream));
|
|
audio_processor_.emplace(stream_id, std::move(processor));
|
|
CHECK(InsertIfNotPresent(
|
|
&stream_index_to_stream_id_,
|
|
options.audio_stream(*options_index_ptr).stream_index(),
|
|
stream_id));
|
|
CHECK(InsertIfNotPresent(&stream_id_to_audio_options_index_,
|
|
stream_id, *options_index_ptr));
|
|
CHECK(InsertIfNotPresent(&audio_options_index_to_stream_id,
|
|
*options_index_ptr, stream_id));
|
|
}
|
|
++current_audio_index;
|
|
break;
|
|
}
|
|
default: {
|
|
// Ignore other stream types.
|
|
}
|
|
}
|
|
}
|
|
for (int i = 0; i < options.audio_stream_size(); ++i) {
|
|
RET_CHECK(ContainsKey(audio_options_index_to_stream_id, i) ||
|
|
options.audio_stream(i).allow_missing())
|
|
<< absl::StrCat("Could not find audio stream with index ", i,
|
|
" in file ", input_file);
|
|
}
|
|
|
|
if (options.has_start_time()) {
|
|
start_time_ = Timestamp::FromSeconds(options.start_time());
|
|
}
|
|
if (options.has_end_time()) {
|
|
end_time_ = Timestamp::FromSeconds(options.end_time());
|
|
}
|
|
is_first_packet_.resize(avformat_ctx_->nb_streams, true);
|
|
|
|
decoder_closer.release();
|
|
return ::mediapipe::OkStatus();
|
|
}
|
|
|
|
::mediapipe::Status AudioDecoder::GetData(int* options_index, Packet* data) {
|
|
while (true) {
|
|
for (auto& item : audio_processor_) {
|
|
while (item.second && item.second->HasData()) {
|
|
bool is_first_packet = is_first_packet_[item.first];
|
|
is_first_packet_[item.first] = false;
|
|
*options_index =
|
|
FindOrDie(stream_id_to_audio_options_index_, item.first);
|
|
::mediapipe::Status status = item.second->GetData(data);
|
|
// Ignore packets which are out of the requested timestamp range.
|
|
if (start_time_ != Timestamp::Unset()) {
|
|
if (is_first_packet && data->Timestamp() > start_time_) {
|
|
LOG(ERROR) << "First packet in audio stream " << *options_index
|
|
<< " has timestamp " << data->Timestamp()
|
|
<< " which is after start time of " << start_time_
|
|
<< ".";
|
|
}
|
|
if (data->Timestamp() < start_time_) {
|
|
VLOG(1) << "Skipping audio frame with timestamp "
|
|
<< data->Timestamp() << " before start time "
|
|
<< start_time_;
|
|
*data = Packet();
|
|
continue;
|
|
}
|
|
}
|
|
if (end_time_ != Timestamp::Unset() && data->Timestamp() > end_time_) {
|
|
VLOG(1) << "Skipping audio frame with timestamp " << data->Timestamp()
|
|
<< " after end time " << end_time_;
|
|
// We are past the last timestamp we care about, close the
|
|
// packet processor. We cannot remove the element from
|
|
// audio_processor_ right now, because we need to continue
|
|
// iterating through it.
|
|
item.second->Close();
|
|
item.second.reset(nullptr);
|
|
*data = Packet();
|
|
continue;
|
|
}
|
|
return status;
|
|
}
|
|
}
|
|
if (flushed_) {
|
|
MP_RETURN_IF_ERROR(Close());
|
|
return tool::StatusStop();
|
|
}
|
|
MP_RETURN_IF_ERROR(ProcessPacket());
|
|
}
|
|
return ::mediapipe::OkStatus();
|
|
}
|
|
|
|
::mediapipe::Status AudioDecoder::Close() {
|
|
for (auto& item : audio_processor_) {
|
|
if (item.second) {
|
|
item.second->Close();
|
|
item.second.reset(nullptr);
|
|
}
|
|
}
|
|
// Free the context.
|
|
if (avformat_ctx_) {
|
|
avformat_close_input(&avformat_ctx_);
|
|
}
|
|
return ::mediapipe::OkStatus();
|
|
}
|
|
|
|
::mediapipe::Status AudioDecoder::FillAudioHeader(
|
|
const AudioStreamOptions& stream_option, TimeSeriesHeader* header) const {
|
|
const std::unique_ptr<AudioPacketProcessor>* processor_ptr_ = FindOrNull(
|
|
audio_processor_,
|
|
FindOrDie(stream_index_to_stream_id_, stream_option.stream_index()));
|
|
|
|
RET_CHECK(processor_ptr_ && *processor_ptr_) << "audio stream is not open.";
|
|
MP_RETURN_IF_ERROR((*processor_ptr_)->FillHeader(header));
|
|
return ::mediapipe::OkStatus();
|
|
}
|
|
|
|
::mediapipe::Status AudioDecoder::ProcessPacket() {
|
|
std::unique_ptr<AVPacket, AVPacketDeleter> av_packet(new AVPacket());
|
|
av_init_packet(av_packet.get());
|
|
av_packet->size = 0;
|
|
av_packet->data = nullptr;
|
|
int ret = av_read_frame(avformat_ctx_, av_packet.get());
|
|
if (ret >= 0) {
|
|
CHECK(av_packet->data) << "AVPacket does not include any data but "
|
|
"av_read_frame was successful.";
|
|
const int stream_id = av_packet->stream_index;
|
|
auto audio_iterator = audio_processor_.find(stream_id);
|
|
if (audio_iterator != audio_processor_.end()) {
|
|
// This stream_id is belongs to an audio stream we care about.
|
|
if (audio_iterator->second) {
|
|
MP_RETURN_IF_ERROR(
|
|
audio_iterator->second->ProcessPacket(av_packet.get()));
|
|
} else {
|
|
VLOG(3) << "processor for stream " << stream_id << " is nullptr.";
|
|
}
|
|
} else {
|
|
VLOG(3) << "Ignoring packet for stream " << stream_id;
|
|
}
|
|
return ::mediapipe::OkStatus();
|
|
}
|
|
VLOG(1) << "Demuxing returned error (or EOF): " << AvErrorToString(ret);
|
|
if (ret == AVERROR(EAGAIN)) {
|
|
// EAGAIN is used to signify that the av_packet should be skipped
|
|
// (maybe the demuxer is trying to re-sync). This definitely
|
|
// occurs in the FLV and MpegT demuxers.
|
|
return ::mediapipe::OkStatus();
|
|
}
|
|
|
|
// Unrecoverable demuxing error with details in avformat_ctx_->pb->error.
|
|
int demuxing_error =
|
|
avformat_ctx_->pb ? avformat_ctx_->pb->error : 0 /* no error */;
|
|
if (ret == AVERROR_EOF && !demuxing_error) {
|
|
VLOG(1) << "Reached EOF.";
|
|
return Flush();
|
|
}
|
|
|
|
RET_CHECK(!demuxing_error) << absl::Substitute(
|
|
"Failed to read a frame: retval = $0 ($1), avformat_ctx_->pb->error = "
|
|
"$2 ($3)",
|
|
ret, AvErrorToString(ret), demuxing_error,
|
|
AvErrorToString(demuxing_error));
|
|
|
|
if (is_first_packet_[av_packet->stream_index]) {
|
|
RET_CHECK_FAIL() << "Couldn't even read the first frame; maybe a partial "
|
|
"file with only metadata?";
|
|
}
|
|
|
|
// Unrecoverable demuxing error without details.
|
|
RET_CHECK_FAIL() << absl::Substitute(
|
|
"Failed to read a frame: retval = $0 ($1)", ret, AvErrorToString(ret));
|
|
}
|
|
|
|
::mediapipe::Status AudioDecoder::Flush() {
|
|
std::vector<::mediapipe::Status> statuses;
|
|
for (auto& item : audio_processor_) {
|
|
if (item.second) {
|
|
statuses.push_back(item.second->Flush());
|
|
}
|
|
}
|
|
flushed_ = true;
|
|
return tool::CombinedStatus("Error while flushing codecs: ", statuses);
|
|
}
|
|
|
|
} // namespace mediapipe
|