| // Copyright 2015 The Chromium Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| #include "media/gpu/vaapi/vaapi_jpeg_decode_accelerator.h" |
| |
| #include <stddef.h> |
| #include <string.h> |
| |
| #include <memory> |
| #include <utility> |
| |
| #include <va/va.h> |
| |
| #include "base/bind.h" |
| #include "base/logging.h" |
| #include "base/metrics/histogram_macros.h" |
| #include "base/threading/thread_task_runner_handle.h" |
| #include "base/trace_event/trace_event.h" |
| #include "media/base/bitstream_buffer.h" |
| #include "media/base/unaligned_shared_memory.h" |
| #include "media/base/video_frame.h" |
| #include "media/filters/jpeg_parser.h" |
| #include "media/gpu/vaapi/vaapi_picture.h" |
| #include "media/gpu/vaapi/vaapi_wrapper.h" |
| #include "third_party/libyuv/include/libyuv.h" |
| |
| #define VLOGF(level) VLOG(level) << __func__ << "(): " |
| #define DVLOGF(level) DVLOG(level) << __func__ << "(): " |
| |
| namespace media { |
| |
| namespace { |
| // UMA errors that the VaapiJpegDecodeAccelerator class reports. |
| enum VAJDADecoderFailure { |
| VAAPI_ERROR = 0, |
| VAJDA_DECODER_FAILURES_MAX, |
| }; |
| |
| static void ReportToUMA(VAJDADecoderFailure failure) { |
| UMA_HISTOGRAM_ENUMERATION("Media.VAJDA.DecoderFailure", failure, |
| VAJDA_DECODER_FAILURES_MAX + 1); |
| } |
| |
| static unsigned int VaSurfaceFormatForJpeg( |
| const JpegFrameHeader& frame_header) { |
| // The range of sampling factor is [1, 4]. Pack them into integer to make the |
| // matching code simpler. For example, 0x211 means the sampling factor are 2, |
| // 1, 1 for 3 components. |
| unsigned int h = 0, v = 0; |
| for (int i = 0; i < frame_header.num_components; i++) { |
| DCHECK_LE(frame_header.components[i].horizontal_sampling_factor, 4); |
| DCHECK_LE(frame_header.components[i].vertical_sampling_factor, 4); |
| h = h << 4 | frame_header.components[i].horizontal_sampling_factor; |
| v = v << 4 | frame_header.components[i].vertical_sampling_factor; |
| } |
| |
| switch (frame_header.num_components) { |
| case 1: // Grey image |
| return VA_RT_FORMAT_YUV400; |
| |
| case 3: // Y Cb Cr color image |
| // See https://en.wikipedia.org/wiki/Chroma_subsampling for the |
| // definition of these numbers. |
| if (h == 0x211 && v == 0x211) |
| return VA_RT_FORMAT_YUV420; |
| |
| if (h == 0x211 && v == 0x111) |
| return VA_RT_FORMAT_YUV422; |
| |
| if (h == 0x111 && v == 0x111) |
| return VA_RT_FORMAT_YUV444; |
| |
| if (h == 0x411 && v == 0x111) |
| return VA_RT_FORMAT_YUV411; |
| } |
| VLOGF(1) << "Unsupported sampling factor: num_components=" |
| << frame_header.num_components << ", h=" << std::hex << h |
| << ", v=" << v; |
| |
| return 0; |
| } |
| |
| // VAAPI only supports a subset of JPEG profiles. This function determines |
| // whether a given parsed JPEG result is supported or not. |
| static bool IsVaapiSupportedJpeg(const JpegParseResult& jpeg) { |
| if (jpeg.frame_header.visible_width < 1 || |
| jpeg.frame_header.visible_height < 1) { |
| DLOG(ERROR) << "width(" << jpeg.frame_header.visible_width |
| << ") and height(" << jpeg.frame_header.visible_height |
| << ") should be at least 1"; |
| return false; |
| } |
| |
| // Size 64k*64k is the maximum in the JPEG standard. VAAPI doesn't support |
| // resolutions larger than 16k*16k. |
| const int kMaxDimension = 16384; |
| if (jpeg.frame_header.coded_width > kMaxDimension || |
| jpeg.frame_header.coded_height > kMaxDimension) { |
| DLOG(ERROR) << "VAAPI doesn't support size(" |
| << jpeg.frame_header.coded_width << "*" |
| << jpeg.frame_header.coded_height << ") larger than " |
| << kMaxDimension << "*" << kMaxDimension; |
| return false; |
| } |
| |
| if (jpeg.frame_header.num_components != 3) { |
| DLOG(ERROR) << "VAAPI doesn't support num_components(" |
| << static_cast<int>(jpeg.frame_header.num_components) |
| << ") != 3"; |
| return false; |
| } |
| |
| if (jpeg.frame_header.components[0].horizontal_sampling_factor < |
| jpeg.frame_header.components[1].horizontal_sampling_factor || |
| jpeg.frame_header.components[0].horizontal_sampling_factor < |
| jpeg.frame_header.components[2].horizontal_sampling_factor) { |
| DLOG(ERROR) << "VAAPI doesn't supports horizontal sampling factor of Y" |
| << " smaller than Cb and Cr"; |
| return false; |
| } |
| |
| if (jpeg.frame_header.components[0].vertical_sampling_factor < |
| jpeg.frame_header.components[1].vertical_sampling_factor || |
| jpeg.frame_header.components[0].vertical_sampling_factor < |
| jpeg.frame_header.components[2].vertical_sampling_factor) { |
| DLOG(ERROR) << "VAAPI doesn't supports vertical sampling factor of Y" |
| << " smaller than Cb and Cr"; |
| return false; |
| } |
| |
| return true; |
| } |
| |
| static void FillPictureParameters( |
| const JpegFrameHeader& frame_header, |
| VAPictureParameterBufferJPEGBaseline* pic_param) { |
| memset(pic_param, 0, sizeof(*pic_param)); |
| pic_param->picture_width = frame_header.coded_width; |
| pic_param->picture_height = frame_header.coded_height; |
| pic_param->num_components = frame_header.num_components; |
| |
| for (int i = 0; i < pic_param->num_components; i++) { |
| pic_param->components[i].component_id = frame_header.components[i].id; |
| pic_param->components[i].h_sampling_factor = |
| frame_header.components[i].horizontal_sampling_factor; |
| pic_param->components[i].v_sampling_factor = |
| frame_header.components[i].vertical_sampling_factor; |
| pic_param->components[i].quantiser_table_selector = |
| frame_header.components[i].quantization_table_selector; |
| } |
| } |
| |
| static void FillIQMatrix(const JpegQuantizationTable* q_table, |
| VAIQMatrixBufferJPEGBaseline* iq_matrix) { |
| memset(iq_matrix, 0, sizeof(*iq_matrix)); |
| static_assert(kJpegMaxQuantizationTableNum == |
| base::size(decltype(iq_matrix->load_quantiser_table){}), |
| "max number of quantization table mismatched"); |
| static_assert( |
| sizeof(iq_matrix->quantiser_table[0]) == sizeof(q_table[0].value), |
| "number of quantization entries mismatched"); |
| for (size_t i = 0; i < kJpegMaxQuantizationTableNum; i++) { |
| if (!q_table[i].valid) |
| continue; |
| iq_matrix->load_quantiser_table[i] = 1; |
| for (size_t j = 0; j < base::size(q_table[i].value); j++) |
| iq_matrix->quantiser_table[i][j] = q_table[i].value[j]; |
| } |
| } |
| |
| static void FillHuffmanTable(const JpegHuffmanTable* dc_table, |
| const JpegHuffmanTable* ac_table, |
| VAHuffmanTableBufferJPEGBaseline* huffman_table) { |
| memset(huffman_table, 0, sizeof(*huffman_table)); |
| // Use default huffman tables if not specified in header. |
| bool has_huffman_table = false; |
| for (size_t i = 0; i < kJpegMaxHuffmanTableNumBaseline; i++) { |
| if (dc_table[i].valid || ac_table[i].valid) { |
| has_huffman_table = true; |
| break; |
| } |
| } |
| if (!has_huffman_table) { |
| dc_table = kDefaultDcTable; |
| ac_table = kDefaultAcTable; |
| } |
| |
| static_assert(kJpegMaxHuffmanTableNumBaseline == |
| base::size(decltype(huffman_table->load_huffman_table){}), |
| "max number of huffman table mismatched"); |
| static_assert(sizeof(huffman_table->huffman_table[0].num_dc_codes) == |
| sizeof(dc_table[0].code_length), |
| "size of huffman table code length mismatch"); |
| static_assert(sizeof(huffman_table->huffman_table[0].dc_values[0]) == |
| sizeof(dc_table[0].code_value[0]), |
| "size of huffman table code value mismatch"); |
| for (size_t i = 0; i < kJpegMaxHuffmanTableNumBaseline; i++) { |
| if (!dc_table[i].valid || !ac_table[i].valid) |
| continue; |
| huffman_table->load_huffman_table[i] = 1; |
| |
| memcpy(huffman_table->huffman_table[i].num_dc_codes, |
| dc_table[i].code_length, |
| sizeof(huffman_table->huffman_table[i].num_dc_codes)); |
| memcpy(huffman_table->huffman_table[i].dc_values, dc_table[i].code_value, |
| sizeof(huffman_table->huffman_table[i].dc_values)); |
| memcpy(huffman_table->huffman_table[i].num_ac_codes, |
| ac_table[i].code_length, |
| sizeof(huffman_table->huffman_table[i].num_ac_codes)); |
| memcpy(huffman_table->huffman_table[i].ac_values, ac_table[i].code_value, |
| sizeof(huffman_table->huffman_table[i].ac_values)); |
| } |
| } |
| |
| static void FillSliceParameters( |
| const JpegParseResult& parse_result, |
| VASliceParameterBufferJPEGBaseline* slice_param) { |
| memset(slice_param, 0, sizeof(*slice_param)); |
| slice_param->slice_data_size = parse_result.data_size; |
| slice_param->slice_data_offset = 0; |
| slice_param->slice_data_flag = VA_SLICE_DATA_FLAG_ALL; |
| slice_param->slice_horizontal_position = 0; |
| slice_param->slice_vertical_position = 0; |
| slice_param->num_components = parse_result.scan.num_components; |
| for (int i = 0; i < slice_param->num_components; i++) { |
| slice_param->components[i].component_selector = |
| parse_result.scan.components[i].component_selector; |
| slice_param->components[i].dc_table_selector = |
| parse_result.scan.components[i].dc_selector; |
| slice_param->components[i].ac_table_selector = |
| parse_result.scan.components[i].ac_selector; |
| } |
| slice_param->restart_interval = parse_result.restart_interval; |
| |
| // Cast to int to prevent overflow. |
| int max_h_factor = |
| parse_result.frame_header.components[0].horizontal_sampling_factor; |
| int max_v_factor = |
| parse_result.frame_header.components[0].vertical_sampling_factor; |
| int mcu_cols = parse_result.frame_header.coded_width / (max_h_factor * 8); |
| DCHECK_GT(mcu_cols, 0); |
| int mcu_rows = parse_result.frame_header.coded_height / (max_v_factor * 8); |
| DCHECK_GT(mcu_rows, 0); |
| slice_param->num_mcus = mcu_rows * mcu_cols; |
| } |
| |
| } // namespace |
| |
| void VaapiJpegDecodeAccelerator::NotifyError(int32_t bitstream_buffer_id, |
| Error error) { |
| if (!task_runner_->BelongsToCurrentThread()) { |
| task_runner_->PostTask( |
| FROM_HERE, base::BindOnce(&VaapiJpegDecodeAccelerator::NotifyError, |
| weak_this_factory_.GetWeakPtr(), |
| bitstream_buffer_id, error)); |
| return; |
| } |
| VLOGF(1) << "Notifying of error " << error; |
| DCHECK(client_); |
| client_->NotifyError(bitstream_buffer_id, error); |
| } |
| |
| void VaapiJpegDecodeAccelerator::VideoFrameReady(int32_t bitstream_buffer_id) { |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| client_->VideoFrameReady(bitstream_buffer_id); |
| } |
| |
| VaapiJpegDecodeAccelerator::VaapiJpegDecodeAccelerator( |
| const scoped_refptr<base::SingleThreadTaskRunner>& io_task_runner) |
| : task_runner_(base::ThreadTaskRunnerHandle::Get()), |
| io_task_runner_(io_task_runner), |
| client_(nullptr), |
| decoder_thread_("VaapiJpegDecoderThread"), |
| va_surface_id_(VA_INVALID_SURFACE), |
| va_rt_format_(0), |
| weak_this_factory_(this) {} |
| |
| VaapiJpegDecodeAccelerator::~VaapiJpegDecodeAccelerator() { |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| VLOGF(2) << "Destroying VaapiJpegDecodeAccelerator"; |
| |
| weak_this_factory_.InvalidateWeakPtrs(); |
| decoder_thread_.Stop(); |
| } |
| |
| bool VaapiJpegDecodeAccelerator::Initialize(Client* client) { |
| VLOGF(2); |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| |
| client_ = client; |
| |
| // Set the image format that will be requested from the VA API. Currently we |
| // always use I420, as this is the expected output format. |
| // TODO(crbug.com/828119): Try a list of possible supported formats rather |
| // than hardcoding the format to I420 here. |
| va_image_format_ = base::WrapUnique(new VAImageFormat{}); |
| va_image_format_->fourcc = VA_FOURCC_I420; |
| va_image_format_->byte_order = VA_LSB_FIRST; |
| va_image_format_->bits_per_pixel = 12; |
| |
| if (!VaapiWrapper::IsImageFormatSupported(*va_image_format_)) { |
| VLOGF(1) << "I420 image format not supported"; |
| va_image_format_.reset(); |
| return false; |
| } |
| |
| vaapi_wrapper_ = |
| VaapiWrapper::Create(VaapiWrapper::kDecode, VAProfileJPEGBaseline, |
| base::Bind(&ReportToUMA, VAAPI_ERROR)); |
| |
| if (!vaapi_wrapper_.get()) { |
| VLOGF(1) << "Failed initializing VAAPI"; |
| return false; |
| } |
| |
| if (!decoder_thread_.Start()) { |
| VLOGF(1) << "Failed to start decoding thread."; |
| return false; |
| } |
| decoder_task_runner_ = decoder_thread_.task_runner(); |
| |
| return true; |
| } |
| |
| bool VaapiJpegDecodeAccelerator::OutputPicture( |
| VASurfaceID va_surface_id, |
| int32_t input_buffer_id, |
| const scoped_refptr<VideoFrame>& video_frame) { |
| DCHECK(decoder_task_runner_->BelongsToCurrentThread()); |
| |
| TRACE_EVENT1("jpeg", "VaapiJpegDecodeAccelerator::OutputPicture", |
| "input_buffer_id", input_buffer_id); |
| |
| DVLOGF(4) << "Outputting VASurface " << va_surface_id |
| << " into video_frame associated with input buffer id " |
| << input_buffer_id; |
| |
| VAImage image = {}; |
| uint8_t* mem = nullptr; |
| gfx::Size coded_size = video_frame->coded_size(); |
| DCHECK(va_image_format_); |
| if (!vaapi_wrapper_->GetVaImage(va_surface_id, va_image_format_.get(), |
| coded_size, &image, |
| reinterpret_cast<void**>(&mem))) { |
| VLOGF(1) << "Cannot get VAImage"; |
| return false; |
| } |
| |
| // Copy image content from VAImage to VideoFrame. |
| // The component order of VAImage I420 are Y, U, and V. |
| DCHECK_EQ(image.num_planes, 3u); |
| DCHECK_GE(image.width, coded_size.width()); |
| DCHECK_GE(image.height, coded_size.height()); |
| const uint8_t* src_y = mem + image.offsets[0]; |
| const uint8_t* src_u = mem + image.offsets[1]; |
| const uint8_t* src_v = mem + image.offsets[2]; |
| size_t src_y_stride = image.pitches[0]; |
| size_t src_u_stride = image.pitches[1]; |
| size_t src_v_stride = image.pitches[2]; |
| uint8_t* dst_y = video_frame->data(VideoFrame::kYPlane); |
| uint8_t* dst_u = video_frame->data(VideoFrame::kUPlane); |
| uint8_t* dst_v = video_frame->data(VideoFrame::kVPlane); |
| size_t dst_y_stride = video_frame->stride(VideoFrame::kYPlane); |
| size_t dst_u_stride = video_frame->stride(VideoFrame::kUPlane); |
| size_t dst_v_stride = video_frame->stride(VideoFrame::kVPlane); |
| |
| if (libyuv::I420Copy(src_y, src_y_stride, // Y |
| src_u, src_u_stride, // U |
| src_v, src_v_stride, // V |
| dst_y, dst_y_stride, // Y |
| dst_u, dst_u_stride, // U |
| dst_v, dst_v_stride, // V |
| coded_size.width(), coded_size.height())) { |
| VLOGF(1) << "I420Copy failed"; |
| return false; |
| } |
| |
| vaapi_wrapper_->ReturnVaImage(&image); |
| |
| task_runner_->PostTask( |
| FROM_HERE, |
| base::BindOnce(&VaapiJpegDecodeAccelerator::VideoFrameReady, |
| weak_this_factory_.GetWeakPtr(), input_buffer_id)); |
| |
| return true; |
| } |
| |
| void VaapiJpegDecodeAccelerator::DecodeTask( |
| int32_t bitstream_buffer_id, |
| std::unique_ptr<UnalignedSharedMemory> shm, |
| scoped_refptr<VideoFrame> video_frame) { |
| DVLOGF(4); |
| DCHECK(decoder_task_runner_->BelongsToCurrentThread()); |
| TRACE_EVENT0("jpeg", "DecodeTask"); |
| |
| JpegParseResult parse_result; |
| if (!ParseJpegPicture(static_cast<const uint8_t*>(shm->memory()), shm->size(), |
| &parse_result)) { |
| VLOGF(1) << "ParseJpegPicture failed"; |
| NotifyError(bitstream_buffer_id, PARSE_JPEG_FAILED); |
| return; |
| } |
| |
| unsigned int new_va_rt_format = |
| VaSurfaceFormatForJpeg(parse_result.frame_header); |
| if (!new_va_rt_format) { |
| VLOGF(1) << "Unsupported subsampling"; |
| NotifyError(bitstream_buffer_id, UNSUPPORTED_JPEG); |
| return; |
| } |
| |
| // Reuse VASurface if size doesn't change. |
| gfx::Size new_coded_size(parse_result.frame_header.coded_width, |
| parse_result.frame_header.coded_height); |
| if (new_coded_size != coded_size_ || va_surface_id_ == VA_INVALID_SURFACE || |
| new_va_rt_format != va_rt_format_) { |
| vaapi_wrapper_->DestroySurfaces(); |
| va_surface_id_ = VA_INVALID_SURFACE; |
| va_rt_format_ = new_va_rt_format; |
| |
| std::vector<VASurfaceID> va_surfaces; |
| if (!vaapi_wrapper_->CreateSurfaces(va_rt_format_, new_coded_size, 1, |
| &va_surfaces)) { |
| VLOGF(1) << "Create VA surface failed"; |
| NotifyError(bitstream_buffer_id, PLATFORM_FAILURE); |
| return; |
| } |
| va_surface_id_ = va_surfaces[0]; |
| coded_size_ = new_coded_size; |
| } |
| |
| if (!DoDecode(vaapi_wrapper_.get(), parse_result, va_surface_id_)) { |
| VLOGF(1) << "Decode JPEG failed"; |
| NotifyError(bitstream_buffer_id, PLATFORM_FAILURE); |
| return; |
| } |
| |
| if (!OutputPicture(va_surface_id_, bitstream_buffer_id, video_frame)) { |
| VLOGF(1) << "Output picture failed"; |
| NotifyError(bitstream_buffer_id, PLATFORM_FAILURE); |
| return; |
| } |
| } |
| |
| void VaapiJpegDecodeAccelerator::Decode( |
| const BitstreamBuffer& bitstream_buffer, |
| const scoped_refptr<VideoFrame>& video_frame) { |
| DCHECK(io_task_runner_->BelongsToCurrentThread()); |
| TRACE_EVENT1("jpeg", "Decode", "input_id", bitstream_buffer.id()); |
| |
| DVLOGF(4) << "Mapping new input buffer id: " << bitstream_buffer.id() |
| << " size: " << bitstream_buffer.size(); |
| |
| // UnalignedSharedMemory will take over the |bitstream_buffer.handle()|. |
| auto shm = std::make_unique<UnalignedSharedMemory>( |
| bitstream_buffer.handle(), bitstream_buffer.size(), true); |
| |
| if (bitstream_buffer.id() < 0) { |
| VLOGF(1) << "Invalid bitstream_buffer, id: " << bitstream_buffer.id(); |
| NotifyError(bitstream_buffer.id(), INVALID_ARGUMENT); |
| return; |
| } |
| |
| if (!shm->MapAt(bitstream_buffer.offset(), bitstream_buffer.size())) { |
| VLOGF(1) << "Failed to map input buffer"; |
| NotifyError(bitstream_buffer.id(), UNREADABLE_INPUT); |
| return; |
| } |
| |
| // It's safe to use base::Unretained(this) because |decoder_task_runner_| runs |
| // tasks on |decoder_thread_| which is stopped in the destructor of |this|. |
| decoder_task_runner_->PostTask( |
| FROM_HERE, base::BindOnce(&VaapiJpegDecodeAccelerator::DecodeTask, |
| base::Unretained(this), bitstream_buffer.id(), |
| std::move(shm), std::move(video_frame))); |
| } |
| |
| bool VaapiJpegDecodeAccelerator::IsSupported() { |
| return VaapiWrapper::IsJpegDecodeSupported(); |
| } |
| |
| // static |
| bool VaapiJpegDecodeAccelerator::DoDecode(VaapiWrapper* vaapi_wrapper, |
| const JpegParseResult& parse_result, |
| VASurfaceID va_surface) { |
| DCHECK_NE(va_surface, VA_INVALID_SURFACE); |
| if (!IsVaapiSupportedJpeg(parse_result)) |
| return false; |
| |
| // Set picture parameters. |
| VAPictureParameterBufferJPEGBaseline pic_param; |
| FillPictureParameters(parse_result.frame_header, &pic_param); |
| if (!vaapi_wrapper->SubmitBuffer(VAPictureParameterBufferType, |
| sizeof(pic_param), &pic_param)) { |
| return false; |
| } |
| |
| // Set quantization table. |
| VAIQMatrixBufferJPEGBaseline iq_matrix; |
| FillIQMatrix(parse_result.q_table, &iq_matrix); |
| if (!vaapi_wrapper->SubmitBuffer(VAIQMatrixBufferType, sizeof(iq_matrix), |
| &iq_matrix)) { |
| return false; |
| } |
| |
| // Set huffman table. |
| VAHuffmanTableBufferJPEGBaseline huffman_table; |
| FillHuffmanTable(parse_result.dc_table, parse_result.ac_table, |
| &huffman_table); |
| if (!vaapi_wrapper->SubmitBuffer(VAHuffmanTableBufferType, |
| sizeof(huffman_table), &huffman_table)) { |
| return false; |
| } |
| |
| // Set slice parameters. |
| VASliceParameterBufferJPEGBaseline slice_param; |
| FillSliceParameters(parse_result, &slice_param); |
| if (!vaapi_wrapper->SubmitBuffer(VASliceParameterBufferType, |
| sizeof(slice_param), &slice_param)) { |
| return false; |
| } |
| |
| // Set scan data. |
| if (!vaapi_wrapper->SubmitBuffer(VASliceDataBufferType, |
| parse_result.data_size, |
| const_cast<char*>(parse_result.data))) { |
| return false; |
| } |
| |
| return vaapi_wrapper->ExecuteAndDestroyPendingBuffers(va_surface); |
| } |
| |
| } // namespace media |