Skip one data copy on dav1d decoding

This CL wraps the |Dav1dPicture| data directly for |VideoFrame| using
instead of copy data out to new buffer.

Bug: None
Change-Id: I21ceffb5cac7dda4a44eafbd0ed221974b8d45ca
Reviewed-on: https://webrtc-review.googlesource.com/c/src/+/276526
Commit-Queue: Zhaoliang Ma <zhaoliang.ma@intel.com>
Reviewed-by: Ilya Nikolaevskiy <ilnik@webrtc.org>
Cr-Commit-Position: refs/heads/main@{#38194}
This commit is contained in:
Zhaoliang Ma 2022-09-26 10:26:42 +08:00 committed by WebRTC LUCI CQ
parent 9f33aea093
commit dd65499002

View File

@ -15,7 +15,7 @@
#include "api/scoped_refptr.h" #include "api/scoped_refptr.h"
#include "api/video/encoded_image.h" #include "api/video/encoded_image.h"
#include "api/video/video_frame_buffer.h" #include "api/video/video_frame_buffer.h"
#include "common_video/include/video_frame_buffer_pool.h" #include "common_video/include/video_frame_buffer.h"
#include "modules/video_coding/include/video_error_codes.h" #include "modules/video_coding/include/video_error_codes.h"
#include "rtc_base/logging.h" #include "rtc_base/logging.h"
#include "third_party/dav1d/libdav1d/include/dav1d/dav1d.h" #include "third_party/dav1d/libdav1d/include/dav1d/dav1d.h"
@ -44,7 +44,6 @@ class Dav1dDecoder : public VideoDecoder {
const char* ImplementationName() const override; const char* ImplementationName() const override;
private: private:
VideoFrameBufferPool buffer_pool_;
Dav1dContext* context_ = nullptr; Dav1dContext* context_ = nullptr;
DecodedImageCallback* decode_complete_callback_ = nullptr; DecodedImageCallback* decode_complete_callback_ = nullptr;
}; };
@ -59,11 +58,13 @@ class ScopedDav1dData {
Dav1dData data_ = {}; Dav1dData data_ = {};
}; };
class ScopedDav1dPicture { class ScopedDav1dPicture
: public rtc::RefCountedNonVirtual<ScopedDav1dPicture> {
public: public:
~ScopedDav1dPicture() { dav1d_picture_unref(&picture_); } ~ScopedDav1dPicture() { dav1d_picture_unref(&picture_); }
Dav1dPicture& Picture() { return picture_; } Dav1dPicture& Picture() { return picture_; }
using rtc::RefCountedNonVirtual<ScopedDav1dPicture>::HasOneRef;
private: private:
Dav1dPicture picture_ = {}; Dav1dPicture picture_ = {};
@ -71,61 +72,10 @@ class ScopedDav1dPicture {
constexpr char kDav1dName[] = "dav1d"; constexpr char kDav1dName[] = "dav1d";
rtc::scoped_refptr<VideoFrameBuffer> CopyFrameData(
VideoFrameBufferPool& buffer_pool,
const Dav1dPicture& dav1d_picture) {
// PlanarYuv8Buffer can represent both I420 and I444 frames.
rtc::scoped_refptr<PlanarYuv8Buffer> buffer;
if (dav1d_picture.p.layout == DAV1D_PIXEL_LAYOUT_I420) {
buffer = buffer_pool.CreateI420Buffer(dav1d_picture.p.w, dav1d_picture.p.h);
} else if (dav1d_picture.p.layout == DAV1D_PIXEL_LAYOUT_I444) {
buffer = buffer_pool.CreateI444Buffer(dav1d_picture.p.w, dav1d_picture.p.h);
} else {
RTC_DCHECK_NOTREACHED() << "Unsupported layout: " << dav1d_picture.p.layout;
}
if (!buffer.get()) {
RTC_LOG(LS_ERROR) << "Dav1dDecoder: failed to allocate a video frame.";
return nullptr;
}
uint8_t* src_y_data = static_cast<uint8_t*>(dav1d_picture.data[0]);
uint8_t* src_u_data = static_cast<uint8_t*>(dav1d_picture.data[1]);
uint8_t* src_v_data = static_cast<uint8_t*>(dav1d_picture.data[2]);
int src_y_stride = dav1d_picture.stride[0];
int src_uv_stride = dav1d_picture.stride[1];
uint8_t* dst_y_data = const_cast<uint8_t*>(buffer->DataY());
uint8_t* dst_u_data = const_cast<uint8_t*>(buffer->DataU());
uint8_t* dst_v_data = const_cast<uint8_t*>(buffer->DataV());
if (dav1d_picture.p.layout == DAV1D_PIXEL_LAYOUT_I420) {
libyuv::I420Copy(src_y_data, src_y_stride, //
src_u_data, src_uv_stride, //
src_v_data, src_uv_stride, //
dst_y_data, buffer->StrideY(), //
dst_u_data, buffer->StrideU(), //
dst_v_data, buffer->StrideV(), //
dav1d_picture.p.w, //
dav1d_picture.p.h); //
} else {
RTC_DCHECK_EQ(dav1d_picture.p.layout, DAV1D_PIXEL_LAYOUT_I444);
libyuv::I444Copy(src_y_data, src_y_stride, //
src_u_data, src_uv_stride, //
src_v_data, src_uv_stride, //
dst_y_data, buffer->StrideY(), //
dst_u_data, buffer->StrideU(), //
dst_v_data, buffer->StrideV(), //
dav1d_picture.p.w, //
dav1d_picture.p.h); //
}
return buffer;
}
// Calling `dav1d_data_wrap` requires a `free_callback` to be registered. // Calling `dav1d_data_wrap` requires a `free_callback` to be registered.
void NullFreeCallback(const uint8_t* buffer, void* opaque) {} void NullFreeCallback(const uint8_t* buffer, void* opaque) {}
Dav1dDecoder::Dav1dDecoder() Dav1dDecoder::Dav1dDecoder() = default;
: buffer_pool_(/*zero_initialize=*/false, /*max_number_of_buffers=*/150) {}
Dav1dDecoder::~Dav1dDecoder() { Dav1dDecoder::~Dav1dDecoder() {
Release(); Release();
@ -154,7 +104,6 @@ int32_t Dav1dDecoder::Release() {
if (context_ != nullptr) { if (context_ != nullptr) {
return WEBRTC_VIDEO_CODEC_MEMORY; return WEBRTC_VIDEO_CODEC_MEMORY;
} }
buffer_pool_.Release();
return WEBRTC_VIDEO_CODEC_OK; return WEBRTC_VIDEO_CODEC_OK;
} }
@ -189,8 +138,9 @@ int32_t Dav1dDecoder::Decode(const EncodedImage& encoded_image,
return WEBRTC_VIDEO_CODEC_ERROR; return WEBRTC_VIDEO_CODEC_ERROR;
} }
ScopedDav1dPicture scoped_dav1d_picture; rtc::scoped_refptr<ScopedDav1dPicture> scoped_dav1d_picture(
Dav1dPicture& dav1d_picture = scoped_dav1d_picture.Picture(); new ScopedDav1dPicture{});
Dav1dPicture& dav1d_picture = scoped_dav1d_picture->Picture();
if (int get_picture_res = dav1d_get_picture(context_, &dav1d_picture)) { if (int get_picture_res = dav1d_get_picture(context_, &dav1d_picture)) {
RTC_LOG(LS_WARNING) RTC_LOG(LS_WARNING)
<< "Dav1dDecoder::Decode getting picture failed with error code " << "Dav1dDecoder::Decode getting picture failed with error code "
@ -205,13 +155,23 @@ int32_t Dav1dDecoder::Decode(const EncodedImage& encoded_image,
return WEBRTC_VIDEO_CODEC_ERROR; return WEBRTC_VIDEO_CODEC_ERROR;
} }
rtc::scoped_refptr<VideoFrameBuffer> buffer; rtc::scoped_refptr<VideoFrameBuffer> wrapped_buffer;
if (dav1d_picture.p.layout == DAV1D_PIXEL_LAYOUT_I420 || if (dav1d_picture.p.layout == DAV1D_PIXEL_LAYOUT_I420) {
dav1d_picture.p.layout == DAV1D_PIXEL_LAYOUT_I444) { wrapped_buffer = WrapI420Buffer(
buffer = CopyFrameData(buffer_pool_, dav1d_picture); dav1d_picture.p.w, dav1d_picture.p.h,
if (!buffer.get()) { static_cast<uint8_t*>(dav1d_picture.data[0]), dav1d_picture.stride[0],
return WEBRTC_VIDEO_CODEC_ERROR; static_cast<uint8_t*>(dav1d_picture.data[1]), dav1d_picture.stride[1],
} static_cast<uint8_t*>(dav1d_picture.data[2]), dav1d_picture.stride[1],
// To keep |scoped_dav1d_picture.Picture()| alive
[scoped_dav1d_picture] {});
} else if (dav1d_picture.p.layout == DAV1D_PIXEL_LAYOUT_I444) {
wrapped_buffer = WrapI444Buffer(
dav1d_picture.p.w, dav1d_picture.p.h,
static_cast<uint8_t*>(dav1d_picture.data[0]), dav1d_picture.stride[0],
static_cast<uint8_t*>(dav1d_picture.data[1]), dav1d_picture.stride[1],
static_cast<uint8_t*>(dav1d_picture.data[2]), dav1d_picture.stride[1],
// To keep |scoped_dav1d_picture.Picture()| alive
[scoped_dav1d_picture] {});
} else { } else {
// Only accept I420 or I444 pixel format. // Only accept I420 or I444 pixel format.
RTC_LOG(LS_ERROR) << "Dav1dDecoder::Decode unhandled pixel layout: " RTC_LOG(LS_ERROR) << "Dav1dDecoder::Decode unhandled pixel layout: "
@ -219,8 +179,12 @@ int32_t Dav1dDecoder::Decode(const EncodedImage& encoded_image,
return WEBRTC_VIDEO_CODEC_ERROR; return WEBRTC_VIDEO_CODEC_ERROR;
} }
if (!wrapped_buffer.get()) {
return WEBRTC_VIDEO_CODEC_ERROR;
}
VideoFrame decoded_frame = VideoFrame::Builder() VideoFrame decoded_frame = VideoFrame::Builder()
.set_video_frame_buffer(buffer) .set_video_frame_buffer(wrapped_buffer)
.set_timestamp_rtp(encoded_image.Timestamp()) .set_timestamp_rtp(encoded_image.Timestamp())
.set_ntp_time_ms(encoded_image.ntp_time_ms_) .set_ntp_time_ms(encoded_image.ntp_time_ms_)
.set_color_space(encoded_image.ColorSpace()) .set_color_space(encoded_image.ColorSpace())