Merge Chromium + Blink git repositories
[chromium-blink-merge.git] / content / renderer / media / rtc_video_decoder.cc
blobf2d4872399a80c09cb677195f08da68c0f7bf4ef
1 // Copyright 2013 The Chromium Authors. All rights reserved.
2 // Use of this source code is governed by a BSD-style license that can be
3 // found in the LICENSE file.
5 #include "content/renderer/media/rtc_video_decoder.h"
7 #include "base/bind.h"
8 #include "base/logging.h"
9 #include "base/memory/ref_counted.h"
10 #include "base/metrics/histogram.h"
11 #include "base/numerics/safe_conversions.h"
12 #include "base/stl_util.h"
13 #include "base/synchronization/waitable_event.h"
14 #include "base/task_runner_util.h"
15 #include "content/renderer/media/webrtc/webrtc_video_frame_adapter.h"
16 #include "gpu/command_buffer/common/mailbox_holder.h"
17 #include "media/base/bind_to_current_loop.h"
18 #include "media/renderers/gpu_video_accelerator_factories.h"
19 #include "third_party/skia/include/core/SkBitmap.h"
20 #include "third_party/webrtc/base/bind.h"
21 #include "third_party/webrtc/system_wrappers/interface/ref_count.h"
22 #include "third_party/webrtc/video_frame.h"
24 namespace content {
26 const int32 RTCVideoDecoder::ID_LAST = 0x3FFFFFFF;
27 const int32 RTCVideoDecoder::ID_HALF = 0x20000000;
28 const int32 RTCVideoDecoder::ID_INVALID = -1;
30 // Maximum number of concurrent VDA::Decode() operations RVD will maintain.
31 // Higher values allow better pipelining in the GPU, but also require more
32 // resources.
33 static const size_t kMaxInFlightDecodes = 8;
35 // Number of allocated shared memory segments.
36 static const size_t kNumSharedMemorySegments = 16;
38 // Maximum number of pending WebRTC buffers that are waiting for shared memory.
39 static const size_t kMaxNumOfPendingBuffers = 8;
41 // A shared memory segment and its allocated size. This class has the ownership
42 // of |shm|.
43 class RTCVideoDecoder::SHMBuffer {
44 public:
45 SHMBuffer(scoped_ptr<base::SharedMemory> shm, size_t size);
46 ~SHMBuffer();
47 scoped_ptr<base::SharedMemory> const shm;
48 const size_t size;
51 RTCVideoDecoder::SHMBuffer::SHMBuffer(scoped_ptr<base::SharedMemory> shm,
52 size_t size)
53 : shm(shm.Pass()), size(size) {
56 RTCVideoDecoder::SHMBuffer::~SHMBuffer() {
59 RTCVideoDecoder::BufferData::BufferData(int32 bitstream_buffer_id,
60 uint32_t timestamp,
61 size_t size)
62 : bitstream_buffer_id(bitstream_buffer_id),
63 timestamp(timestamp),
64 size(size) {}
66 RTCVideoDecoder::BufferData::BufferData() {}
68 RTCVideoDecoder::BufferData::~BufferData() {}
70 RTCVideoDecoder::RTCVideoDecoder(
71 webrtc::VideoCodecType type,
72 const scoped_refptr<media::GpuVideoAcceleratorFactories>& factories)
73 : video_codec_type_(type),
74 factories_(factories),
75 decoder_texture_target_(0),
76 next_picture_buffer_id_(0),
77 state_(UNINITIALIZED),
78 decode_complete_callback_(NULL),
79 num_shm_buffers_(0),
80 next_bitstream_buffer_id_(0),
81 reset_bitstream_buffer_id_(ID_INVALID),
82 weak_factory_(this) {
83 DCHECK(!factories_->GetTaskRunner()->BelongsToCurrentThread());
86 RTCVideoDecoder::~RTCVideoDecoder() {
87 DVLOG(2) << "~RTCVideoDecoder";
88 DCheckGpuVideoAcceleratorFactoriesTaskRunnerIsCurrent();
89 DestroyVDA();
91 // Delete all shared memories.
92 STLDeleteElements(&available_shm_segments_);
93 STLDeleteValues(&bitstream_buffers_in_decoder_);
94 STLDeleteContainerPairFirstPointers(decode_buffers_.begin(),
95 decode_buffers_.end());
96 decode_buffers_.clear();
97 ClearPendingBuffers();
100 // static
101 scoped_ptr<RTCVideoDecoder> RTCVideoDecoder::Create(
102 webrtc::VideoCodecType type,
103 const scoped_refptr<media::GpuVideoAcceleratorFactories>& factories) {
104 scoped_ptr<RTCVideoDecoder> decoder;
105 // Convert WebRTC codec type to media codec profile.
106 media::VideoCodecProfile profile;
107 switch (type) {
108 case webrtc::kVideoCodecVP8:
109 profile = media::VP8PROFILE_ANY;
110 break;
111 case webrtc::kVideoCodecH264:
112 profile = media::H264PROFILE_MAIN;
113 break;
114 default:
115 DVLOG(2) << "Video codec not supported:" << type;
116 return decoder.Pass();
119 base::WaitableEvent waiter(true, false);
120 decoder.reset(new RTCVideoDecoder(type, factories));
121 decoder->factories_->GetTaskRunner()->PostTask(
122 FROM_HERE,
123 base::Bind(&RTCVideoDecoder::CreateVDA,
124 base::Unretained(decoder.get()),
125 profile,
126 &waiter));
127 waiter.Wait();
128 // vda can be NULL if the codec is not supported.
129 if (decoder->vda_ != NULL) {
130 decoder->state_ = INITIALIZED;
131 } else {
132 factories->GetTaskRunner()->DeleteSoon(FROM_HERE, decoder.release());
134 return decoder.Pass();
137 int32_t RTCVideoDecoder::InitDecode(const webrtc::VideoCodec* codecSettings,
138 int32_t /*numberOfCores*/) {
139 DVLOG(2) << "InitDecode";
140 DCHECK_EQ(video_codec_type_, codecSettings->codecType);
141 if (codecSettings->codecType == webrtc::kVideoCodecVP8 &&
142 codecSettings->codecSpecific.VP8.feedbackModeOn) {
143 LOG(ERROR) << "Feedback mode not supported";
144 return RecordInitDecodeUMA(WEBRTC_VIDEO_CODEC_ERROR);
147 base::AutoLock auto_lock(lock_);
148 if (state_ == UNINITIALIZED || state_ == DECODE_ERROR) {
149 LOG(ERROR) << "VDA is not initialized. state=" << state_;
150 return RecordInitDecodeUMA(WEBRTC_VIDEO_CODEC_UNINITIALIZED);
153 return RecordInitDecodeUMA(WEBRTC_VIDEO_CODEC_OK);
156 int32_t RTCVideoDecoder::Decode(
157 const webrtc::EncodedImage& inputImage,
158 bool missingFrames,
159 const webrtc::RTPFragmentationHeader* /*fragmentation*/,
160 const webrtc::CodecSpecificInfo* /*codecSpecificInfo*/,
161 int64_t /*renderTimeMs*/) {
162 DVLOG(3) << "Decode";
164 base::AutoLock auto_lock(lock_);
166 if (state_ == UNINITIALIZED || decode_complete_callback_ == NULL) {
167 LOG(ERROR) << "The decoder has not initialized.";
168 return WEBRTC_VIDEO_CODEC_UNINITIALIZED;
171 if (state_ == DECODE_ERROR) {
172 LOG(ERROR) << "Decoding error occurred.";
173 return WEBRTC_VIDEO_CODEC_ERROR;
176 if (missingFrames || !inputImage._completeFrame) {
177 DLOG(ERROR) << "Missing or incomplete frames.";
178 // Unlike the SW decoder in libvpx, hw decoder cannot handle broken frames.
179 // Return an error to request a key frame.
180 return WEBRTC_VIDEO_CODEC_ERROR;
183 // Most platforms' VDA implementations support mid-stream resolution change
184 // internally. Platforms whose VDAs fail to support mid-stream resolution
185 // change gracefully need to have their clients cover for them, and we do that
186 // here.
187 #ifdef ANDROID
188 const bool kVDACanHandleMidstreamResize = false;
189 #else
190 const bool kVDACanHandleMidstreamResize = true;
191 #endif
193 bool need_to_reset_for_midstream_resize = false;
194 if (inputImage._frameType == webrtc::kKeyFrame) {
195 gfx::Size new_frame_size(inputImage._encodedWidth,
196 inputImage._encodedHeight);
197 DVLOG(2) << "Got key frame. size=" << new_frame_size.ToString();
199 if (new_frame_size.width() > max_resolution_.width() ||
200 new_frame_size.width() < min_resolution_.width() ||
201 new_frame_size.height() > max_resolution_.height() ||
202 new_frame_size.height() < min_resolution_.height()) {
203 DVLOG(1) << "Resolution unsupported, falling back to software decode";
204 return WEBRTC_VIDEO_CODEC_FALLBACK_SOFTWARE;
207 gfx::Size prev_frame_size = frame_size_;
208 frame_size_ = new_frame_size;
209 if (!kVDACanHandleMidstreamResize && !prev_frame_size.IsEmpty() &&
210 prev_frame_size != frame_size_) {
211 need_to_reset_for_midstream_resize = true;
213 } else if (IsFirstBufferAfterReset(next_bitstream_buffer_id_,
214 reset_bitstream_buffer_id_)) {
215 // TODO(wuchengli): VDA should handle it. Remove this when
216 // http://crosbug.com/p/21913 is fixed.
217 DVLOG(1) << "The first frame should be a key frame. Drop this.";
218 return WEBRTC_VIDEO_CODEC_ERROR;
221 // Create buffer metadata.
222 BufferData buffer_data(next_bitstream_buffer_id_,
223 inputImage._timeStamp,
224 inputImage._length);
225 // Mask against 30 bits, to avoid (undefined) wraparound on signed integer.
226 next_bitstream_buffer_id_ = (next_bitstream_buffer_id_ + 1) & ID_LAST;
228 // If a shared memory segment is available, there are no pending buffers, and
229 // this isn't a mid-stream resolution change, then send the buffer for decode
230 // immediately. Otherwise, save the buffer in the queue for later decode.
231 scoped_ptr<SHMBuffer> shm_buffer;
232 if (!need_to_reset_for_midstream_resize && pending_buffers_.empty())
233 shm_buffer = GetSHM_Locked(inputImage._length);
234 if (!shm_buffer) {
235 if (!SaveToPendingBuffers_Locked(inputImage, buffer_data)) {
236 // We have exceeded the pending buffers count, we are severely behind.
237 // Since we are returning ERROR, WebRTC will not be interested in the
238 // remaining buffers, and will provide us with a new keyframe instead.
239 // Better to drop any pending buffers and start afresh to catch up faster.
240 DVLOG(1) << "Exceeded maximum pending buffer count, dropping";
241 ClearPendingBuffers();
242 return WEBRTC_VIDEO_CODEC_ERROR;
245 if (need_to_reset_for_midstream_resize) {
246 base::AutoUnlock auto_unlock(lock_);
247 Reset();
250 return WEBRTC_VIDEO_CODEC_OK;
253 SaveToDecodeBuffers_Locked(inputImage, shm_buffer.Pass(), buffer_data);
254 factories_->GetTaskRunner()->PostTask(
255 FROM_HERE,
256 base::Bind(&RTCVideoDecoder::RequestBufferDecode,
257 weak_factory_.GetWeakPtr()));
258 return WEBRTC_VIDEO_CODEC_OK;
261 int32_t RTCVideoDecoder::RegisterDecodeCompleteCallback(
262 webrtc::DecodedImageCallback* callback) {
263 DVLOG(2) << "RegisterDecodeCompleteCallback";
264 base::AutoLock auto_lock(lock_);
265 decode_complete_callback_ = callback;
266 return WEBRTC_VIDEO_CODEC_OK;
269 int32_t RTCVideoDecoder::Release() {
270 DVLOG(2) << "Release";
271 // Do not destroy VDA because WebRTC can call InitDecode and start decoding
272 // again.
273 return Reset();
276 int32_t RTCVideoDecoder::Reset() {
277 DVLOG(2) << "Reset";
278 base::AutoLock auto_lock(lock_);
279 if (state_ == UNINITIALIZED) {
280 LOG(ERROR) << "Decoder not initialized.";
281 return WEBRTC_VIDEO_CODEC_UNINITIALIZED;
283 if (next_bitstream_buffer_id_ != 0)
284 reset_bitstream_buffer_id_ = next_bitstream_buffer_id_ - 1;
285 else
286 reset_bitstream_buffer_id_ = ID_LAST;
287 // If VDA is already resetting, no need to request the reset again.
288 if (state_ != RESETTING) {
289 state_ = RESETTING;
290 factories_->GetTaskRunner()->PostTask(
291 FROM_HERE,
292 base::Bind(&RTCVideoDecoder::ResetInternal,
293 weak_factory_.GetWeakPtr()));
295 return WEBRTC_VIDEO_CODEC_OK;
298 void RTCVideoDecoder::ProvidePictureBuffers(uint32 count,
299 const gfx::Size& size,
300 uint32 texture_target) {
301 DCheckGpuVideoAcceleratorFactoriesTaskRunnerIsCurrent();
302 DVLOG(3) << "ProvidePictureBuffers. texture_target=" << texture_target;
304 if (!vda_)
305 return;
307 std::vector<uint32> texture_ids;
308 std::vector<gpu::Mailbox> texture_mailboxes;
309 decoder_texture_target_ = texture_target;
310 if (!factories_->CreateTextures(count,
311 size,
312 &texture_ids,
313 &texture_mailboxes,
314 decoder_texture_target_)) {
315 NotifyError(media::VideoDecodeAccelerator::PLATFORM_FAILURE);
316 return;
318 DCHECK_EQ(count, texture_ids.size());
319 DCHECK_EQ(count, texture_mailboxes.size());
321 std::vector<media::PictureBuffer> picture_buffers;
322 for (size_t i = 0; i < texture_ids.size(); ++i) {
323 picture_buffers.push_back(media::PictureBuffer(
324 next_picture_buffer_id_++, size, texture_ids[i], texture_mailboxes[i]));
325 bool inserted = assigned_picture_buffers_.insert(std::make_pair(
326 picture_buffers.back().id(), picture_buffers.back())).second;
327 DCHECK(inserted);
329 vda_->AssignPictureBuffers(picture_buffers);
332 void RTCVideoDecoder::DismissPictureBuffer(int32 id) {
333 DVLOG(3) << "DismissPictureBuffer. id=" << id;
334 DCheckGpuVideoAcceleratorFactoriesTaskRunnerIsCurrent();
336 std::map<int32, media::PictureBuffer>::iterator it =
337 assigned_picture_buffers_.find(id);
338 if (it == assigned_picture_buffers_.end()) {
339 NOTREACHED() << "Missing picture buffer: " << id;
340 return;
343 media::PictureBuffer buffer_to_dismiss = it->second;
344 assigned_picture_buffers_.erase(it);
346 if (!picture_buffers_at_display_.count(id)) {
347 // We can delete the texture immediately as it's not being displayed.
348 factories_->DeleteTexture(buffer_to_dismiss.texture_id());
349 return;
351 // Not destroying a texture in display in |picture_buffers_at_display_|.
352 // Postpone deletion until after it's returned to us.
355 void RTCVideoDecoder::PictureReady(const media::Picture& picture) {
356 DVLOG(3) << "PictureReady";
357 DCheckGpuVideoAcceleratorFactoriesTaskRunnerIsCurrent();
359 std::map<int32, media::PictureBuffer>::iterator it =
360 assigned_picture_buffers_.find(picture.picture_buffer_id());
361 if (it == assigned_picture_buffers_.end()) {
362 NOTREACHED() << "Missing picture buffer: " << picture.picture_buffer_id();
363 NotifyError(media::VideoDecodeAccelerator::PLATFORM_FAILURE);
364 return;
366 const media::PictureBuffer& pb = it->second;
368 // Validate picture rectangle from GPU.
369 if (picture.visible_rect().IsEmpty() ||
370 !gfx::Rect(pb.size()).Contains(picture.visible_rect())) {
371 NOTREACHED() << "Invalid picture size from VDA: "
372 << picture.visible_rect().ToString() << " should fit in "
373 << pb.size().ToString();
374 NotifyError(media::VideoDecodeAccelerator::PLATFORM_FAILURE);
375 return;
378 // Create a media::VideoFrame.
379 uint32_t timestamp = 0;
380 GetBufferData(picture.bitstream_buffer_id(), &timestamp);
381 scoped_refptr<media::VideoFrame> frame =
382 CreateVideoFrame(picture, pb, timestamp);
383 bool inserted =
384 picture_buffers_at_display_.insert(std::make_pair(
385 picture.picture_buffer_id(),
386 pb.texture_id())).second;
387 DCHECK(inserted);
389 // Create a WebRTC video frame.
390 webrtc::VideoFrame decoded_image(
391 new rtc::RefCountedObject<WebRtcVideoFrameAdapter>(frame), timestamp, 0,
392 webrtc::kVideoRotation_0);
394 // Invoke decode callback. WebRTC expects no callback after Reset or Release.
396 base::AutoLock auto_lock(lock_);
397 DCHECK(decode_complete_callback_ != NULL);
398 if (IsBufferAfterReset(picture.bitstream_buffer_id(),
399 reset_bitstream_buffer_id_)) {
400 decode_complete_callback_->Decoded(decoded_image);
405 scoped_refptr<media::VideoFrame> RTCVideoDecoder::CreateVideoFrame(
406 const media::Picture& picture,
407 const media::PictureBuffer& pb,
408 uint32_t timestamp) {
409 gfx::Rect visible_rect(picture.visible_rect());
410 DCHECK(decoder_texture_target_);
411 // Convert timestamp from 90KHz to ms.
412 base::TimeDelta timestamp_ms = base::TimeDelta::FromInternalValue(
413 base::checked_cast<uint64_t>(timestamp) * 1000 / 90);
414 // TODO(mcasas): The incoming data is actually a YUV format, but is labelled
415 // as ARGB. This prevents the compositor from messing with it, since the
416 // underlying platform can handle the former format natively. Make sure the
417 // correct format is used and everyone down the line understands it.
418 scoped_refptr<media::VideoFrame> frame(media::VideoFrame::WrapNativeTexture(
419 media::PIXEL_FORMAT_ARGB,
420 gpu::MailboxHolder(pb.texture_mailbox(), decoder_texture_target_, 0),
421 media::BindToCurrentLoop(base::Bind(
422 &RTCVideoDecoder::ReleaseMailbox, weak_factory_.GetWeakPtr(),
423 factories_, picture.picture_buffer_id(), pb.texture_id())),
424 pb.size(), visible_rect, visible_rect.size(), timestamp_ms));
425 if (picture.allow_overlay()) {
426 frame->metadata()->SetBoolean(media::VideoFrameMetadata::ALLOW_OVERLAY,
427 true);
429 return frame;
432 void RTCVideoDecoder::NotifyEndOfBitstreamBuffer(int32 id) {
433 DVLOG(3) << "NotifyEndOfBitstreamBuffer. id=" << id;
434 DCheckGpuVideoAcceleratorFactoriesTaskRunnerIsCurrent();
436 std::map<int32, SHMBuffer*>::iterator it =
437 bitstream_buffers_in_decoder_.find(id);
438 if (it == bitstream_buffers_in_decoder_.end()) {
439 NotifyError(media::VideoDecodeAccelerator::PLATFORM_FAILURE);
440 NOTREACHED() << "Missing bitstream buffer: " << id;
441 return;
445 base::AutoLock auto_lock(lock_);
446 PutSHM_Locked(scoped_ptr<SHMBuffer>(it->second));
448 bitstream_buffers_in_decoder_.erase(it);
450 RequestBufferDecode();
453 void RTCVideoDecoder::NotifyFlushDone() {
454 DVLOG(3) << "NotifyFlushDone";
455 NOTREACHED() << "Unexpected flush done notification.";
458 void RTCVideoDecoder::NotifyResetDone() {
459 DCheckGpuVideoAcceleratorFactoriesTaskRunnerIsCurrent();
460 DVLOG(3) << "NotifyResetDone";
462 if (!vda_)
463 return;
465 input_buffer_data_.clear();
467 base::AutoLock auto_lock(lock_);
468 state_ = INITIALIZED;
470 // Send the pending buffers for decoding.
471 RequestBufferDecode();
474 void RTCVideoDecoder::NotifyError(media::VideoDecodeAccelerator::Error error) {
475 DCheckGpuVideoAcceleratorFactoriesTaskRunnerIsCurrent();
476 if (!vda_)
477 return;
479 LOG(ERROR) << "VDA Error:" << error;
480 UMA_HISTOGRAM_ENUMERATION("Media.RTCVideoDecoderError",
481 error,
482 media::VideoDecodeAccelerator::LARGEST_ERROR_ENUM);
483 DestroyVDA();
485 base::AutoLock auto_lock(lock_);
486 state_ = DECODE_ERROR;
489 void RTCVideoDecoder::RequestBufferDecode() {
490 DCheckGpuVideoAcceleratorFactoriesTaskRunnerIsCurrent();
491 if (!vda_)
492 return;
494 MovePendingBuffersToDecodeBuffers();
496 while (CanMoreDecodeWorkBeDone()) {
497 // Get a buffer and data from the queue.
498 SHMBuffer* shm_buffer = NULL;
499 BufferData buffer_data;
501 base::AutoLock auto_lock(lock_);
502 // Do not request decode if VDA is resetting.
503 if (decode_buffers_.empty() || state_ == RESETTING)
504 return;
505 shm_buffer = decode_buffers_.front().first;
506 buffer_data = decode_buffers_.front().second;
507 decode_buffers_.pop_front();
508 // Drop the buffers before Reset or Release is called.
509 if (!IsBufferAfterReset(buffer_data.bitstream_buffer_id,
510 reset_bitstream_buffer_id_)) {
511 PutSHM_Locked(scoped_ptr<SHMBuffer>(shm_buffer));
512 continue;
516 // Create a BitstreamBuffer and send to VDA to decode.
517 media::BitstreamBuffer bitstream_buffer(
518 buffer_data.bitstream_buffer_id, shm_buffer->shm->handle(),
519 buffer_data.size,
520 base::TimeDelta::FromInternalValue(buffer_data.timestamp));
521 bool inserted = bitstream_buffers_in_decoder_
522 .insert(std::make_pair(bitstream_buffer.id(), shm_buffer)).second;
523 DCHECK(inserted);
524 RecordBufferData(buffer_data);
525 vda_->Decode(bitstream_buffer);
529 bool RTCVideoDecoder::CanMoreDecodeWorkBeDone() {
530 return bitstream_buffers_in_decoder_.size() < kMaxInFlightDecodes;
533 bool RTCVideoDecoder::IsBufferAfterReset(int32 id_buffer, int32 id_reset) {
534 if (id_reset == ID_INVALID)
535 return true;
536 int32 diff = id_buffer - id_reset;
537 if (diff <= 0)
538 diff += ID_LAST + 1;
539 return diff < ID_HALF;
542 bool RTCVideoDecoder::IsFirstBufferAfterReset(int32 id_buffer, int32 id_reset) {
543 if (id_reset == ID_INVALID)
544 return id_buffer == 0;
545 return id_buffer == ((id_reset + 1) & ID_LAST);
548 void RTCVideoDecoder::SaveToDecodeBuffers_Locked(
549 const webrtc::EncodedImage& input_image,
550 scoped_ptr<SHMBuffer> shm_buffer,
551 const BufferData& buffer_data) {
552 memcpy(shm_buffer->shm->memory(), input_image._buffer, input_image._length);
553 std::pair<SHMBuffer*, BufferData> buffer_pair =
554 std::make_pair(shm_buffer.release(), buffer_data);
556 // Store the buffer and the metadata to the queue.
557 decode_buffers_.push_back(buffer_pair);
560 bool RTCVideoDecoder::SaveToPendingBuffers_Locked(
561 const webrtc::EncodedImage& input_image,
562 const BufferData& buffer_data) {
563 DVLOG(2) << "SaveToPendingBuffers_Locked"
564 << ". pending_buffers size=" << pending_buffers_.size()
565 << ". decode_buffers_ size=" << decode_buffers_.size()
566 << ". available_shm size=" << available_shm_segments_.size();
567 // Queued too many buffers. Something goes wrong.
568 if (pending_buffers_.size() >= kMaxNumOfPendingBuffers) {
569 LOG(WARNING) << "Too many pending buffers!";
570 return false;
573 // Clone the input image and save it to the queue.
574 uint8_t* buffer = new uint8_t[input_image._length];
575 // TODO(wuchengli): avoid memcpy. Extend webrtc::VideoDecoder::Decode()
576 // interface to take a non-const ptr to the frame and add a method to the
577 // frame that will swap buffers with another.
578 memcpy(buffer, input_image._buffer, input_image._length);
579 webrtc::EncodedImage encoded_image(
580 buffer, input_image._length, input_image._length);
581 std::pair<webrtc::EncodedImage, BufferData> buffer_pair =
582 std::make_pair(encoded_image, buffer_data);
584 pending_buffers_.push_back(buffer_pair);
585 return true;
588 void RTCVideoDecoder::MovePendingBuffersToDecodeBuffers() {
589 base::AutoLock auto_lock(lock_);
590 while (pending_buffers_.size() > 0) {
591 // Get a pending buffer from the queue.
592 const webrtc::EncodedImage& input_image = pending_buffers_.front().first;
593 const BufferData& buffer_data = pending_buffers_.front().second;
595 // Drop the frame if it comes before Reset or Release.
596 if (!IsBufferAfterReset(buffer_data.bitstream_buffer_id,
597 reset_bitstream_buffer_id_)) {
598 delete[] input_image._buffer;
599 pending_buffers_.pop_front();
600 continue;
602 // Get shared memory and save it to decode buffers.
603 scoped_ptr<SHMBuffer> shm_buffer = GetSHM_Locked(input_image._length);
604 if (!shm_buffer)
605 return;
606 SaveToDecodeBuffers_Locked(input_image, shm_buffer.Pass(), buffer_data);
607 delete[] input_image._buffer;
608 pending_buffers_.pop_front();
612 void RTCVideoDecoder::ResetInternal() {
613 DCheckGpuVideoAcceleratorFactoriesTaskRunnerIsCurrent();
614 DVLOG(2) << "ResetInternal";
615 if (vda_)
616 vda_->Reset();
619 // static
620 void RTCVideoDecoder::ReleaseMailbox(
621 base::WeakPtr<RTCVideoDecoder> decoder,
622 const scoped_refptr<media::GpuVideoAcceleratorFactories>& factories,
623 int64 picture_buffer_id,
624 uint32 texture_id,
625 uint32 release_sync_point) {
626 DCHECK(factories->GetTaskRunner()->BelongsToCurrentThread());
627 factories->WaitSyncPoint(release_sync_point);
629 if (decoder) {
630 decoder->ReusePictureBuffer(picture_buffer_id);
631 return;
633 // It's the last chance to delete the texture after display,
634 // because RTCVideoDecoder was destructed.
635 factories->DeleteTexture(texture_id);
638 void RTCVideoDecoder::ReusePictureBuffer(int64 picture_buffer_id) {
639 DCheckGpuVideoAcceleratorFactoriesTaskRunnerIsCurrent();
640 DVLOG(3) << "ReusePictureBuffer. id=" << picture_buffer_id;
642 DCHECK(!picture_buffers_at_display_.empty());
643 PictureBufferTextureMap::iterator display_iterator =
644 picture_buffers_at_display_.find(picture_buffer_id);
645 uint32 texture_id = display_iterator->second;
646 DCHECK(display_iterator != picture_buffers_at_display_.end());
647 picture_buffers_at_display_.erase(display_iterator);
649 if (!assigned_picture_buffers_.count(picture_buffer_id)) {
650 // This picture was dismissed while in display, so we postponed deletion.
651 factories_->DeleteTexture(texture_id);
652 return;
655 // DestroyVDA() might already have been called.
656 if (vda_)
657 vda_->ReusePictureBuffer(picture_buffer_id);
660 bool RTCVideoDecoder::IsProfileSupported(media::VideoCodecProfile profile) {
661 DCheckGpuVideoAcceleratorFactoriesTaskRunnerIsCurrent();
662 media::VideoDecodeAccelerator::SupportedProfiles supported_profiles =
663 factories_->GetVideoDecodeAcceleratorSupportedProfiles();
665 for (const auto& supported_profile : supported_profiles) {
666 if (profile == supported_profile.profile) {
667 min_resolution_ = supported_profile.min_resolution;
668 max_resolution_ = supported_profile.max_resolution;
669 return true;
673 return false;
676 void RTCVideoDecoder::CreateVDA(media::VideoCodecProfile profile,
677 base::WaitableEvent* waiter) {
678 DCheckGpuVideoAcceleratorFactoriesTaskRunnerIsCurrent();
680 if (!IsProfileSupported(profile)) {
681 DVLOG(1) << "Unsupported profile " << profile;
682 } else {
683 vda_ = factories_->CreateVideoDecodeAccelerator();
684 if (vda_ && !vda_->Initialize(profile, this))
685 vda_.release()->Destroy();
688 waiter->Signal();
691 void RTCVideoDecoder::DestroyTextures() {
692 DCheckGpuVideoAcceleratorFactoriesTaskRunnerIsCurrent();
694 // Not destroying PictureBuffers in |picture_buffers_at_display_| yet, since
695 // their textures may still be in use by the user of this RTCVideoDecoder.
696 for (PictureBufferTextureMap::iterator it =
697 picture_buffers_at_display_.begin();
698 it != picture_buffers_at_display_.end();
699 ++it) {
700 assigned_picture_buffers_.erase(it->first);
703 for (std::map<int32, media::PictureBuffer>::iterator it =
704 assigned_picture_buffers_.begin();
705 it != assigned_picture_buffers_.end();
706 ++it) {
707 factories_->DeleteTexture(it->second.texture_id());
709 assigned_picture_buffers_.clear();
712 void RTCVideoDecoder::DestroyVDA() {
713 DVLOG(2) << "DestroyVDA";
714 DCheckGpuVideoAcceleratorFactoriesTaskRunnerIsCurrent();
715 if (vda_)
716 vda_.release()->Destroy();
717 DestroyTextures();
718 base::AutoLock auto_lock(lock_);
719 state_ = UNINITIALIZED;
722 scoped_ptr<RTCVideoDecoder::SHMBuffer> RTCVideoDecoder::GetSHM_Locked(
723 size_t min_size) {
724 // Reuse a SHM if possible.
725 if (!available_shm_segments_.empty() &&
726 available_shm_segments_.back()->size >= min_size) {
727 scoped_ptr<SHMBuffer> buffer(available_shm_segments_.back());
728 available_shm_segments_.pop_back();
729 return buffer;
732 if (available_shm_segments_.size() != num_shm_buffers_) {
733 // Either available_shm_segments_ is empty (and we already have some SHM
734 // buffers allocated), or the size of available segments is not large
735 // enough. In the former case we need to wait for buffers to be returned,
736 // in the latter we need to wait for all buffers to be returned to drop
737 // them and reallocate with a new size.
738 return NULL;
741 if (num_shm_buffers_ != 0) {
742 STLDeleteElements(&available_shm_segments_);
743 num_shm_buffers_ = 0;
746 // Create twice as large buffers as required, to avoid frequent reallocation.
747 factories_->GetTaskRunner()->PostTask(
748 FROM_HERE,
749 base::Bind(&RTCVideoDecoder::CreateSHM, weak_factory_.GetWeakPtr(),
750 kNumSharedMemorySegments, min_size * 2));
752 // We'll be called again after the shared memory is created.
753 return NULL;
756 void RTCVideoDecoder::PutSHM_Locked(scoped_ptr<SHMBuffer> shm_buffer) {
757 available_shm_segments_.push_back(shm_buffer.release());
760 void RTCVideoDecoder::CreateSHM(size_t count, size_t size) {
761 DCheckGpuVideoAcceleratorFactoriesTaskRunnerIsCurrent();
762 DVLOG(2) << "CreateSHM. count=" << count << ", size=" << size;
764 for (size_t i = 0; i < count; i++) {
765 scoped_ptr<base::SharedMemory> shm = factories_->CreateSharedMemory(size);
766 if (!shm) {
767 LOG(ERROR) << "Failed allocating shared memory of size=" << size;
768 NotifyError(media::VideoDecodeAccelerator::PLATFORM_FAILURE);
769 return;
772 base::AutoLock auto_lock(lock_);
773 PutSHM_Locked(scoped_ptr<SHMBuffer>(new SHMBuffer(shm.Pass(), size)));
774 ++num_shm_buffers_;
777 // Kick off the decoding.
778 RequestBufferDecode();
781 void RTCVideoDecoder::RecordBufferData(const BufferData& buffer_data) {
782 input_buffer_data_.push_front(buffer_data);
783 // Why this value? Because why not. avformat.h:MAX_REORDER_DELAY is 16, but
784 // that's too small for some pathological B-frame test videos. The cost of
785 // using too-high a value is low (192 bits per extra slot).
786 static const size_t kMaxInputBufferDataSize = 128;
787 // Pop from the back of the list, because that's the oldest and least likely
788 // to be useful in the future data.
789 if (input_buffer_data_.size() > kMaxInputBufferDataSize)
790 input_buffer_data_.pop_back();
793 void RTCVideoDecoder::GetBufferData(int32 bitstream_buffer_id,
794 uint32_t* timestamp) {
795 for (std::list<BufferData>::iterator it = input_buffer_data_.begin();
796 it != input_buffer_data_.end();
797 ++it) {
798 if (it->bitstream_buffer_id != bitstream_buffer_id)
799 continue;
800 *timestamp = it->timestamp;
801 return;
803 NOTREACHED() << "Missing bitstream buffer id: " << bitstream_buffer_id;
806 int32_t RTCVideoDecoder::RecordInitDecodeUMA(int32_t status) {
807 // Logging boolean is enough to know if HW decoding has been used. Also,
808 // InitDecode is less likely to return an error so enum is not used here.
809 bool sample = (status == WEBRTC_VIDEO_CODEC_OK) ? true : false;
810 UMA_HISTOGRAM_BOOLEAN("Media.RTCVideoDecoderInitDecodeSuccess", sample);
811 return status;
814 void RTCVideoDecoder::DCheckGpuVideoAcceleratorFactoriesTaskRunnerIsCurrent()
815 const {
816 DCHECK(factories_->GetTaskRunner()->BelongsToCurrentThread());
819 void RTCVideoDecoder::ClearPendingBuffers() {
820 // Delete WebRTC input buffers.
821 for (std::deque<std::pair<webrtc::EncodedImage, BufferData>>::iterator it =
822 pending_buffers_.begin();
823 it != pending_buffers_.end(); ++it) {
824 delete[] it->first._buffer;
827 pending_buffers_.clear();
830 } // namespace content