| /* |
| * Copyright (c) 2013 The WebRTC project authors. All Rights Reserved. |
| * |
| * Use of this source code is governed by a BSD-style license |
| * that can be found in the LICENSE file in the root of the source |
| * tree. An additional intellectual property rights grant can be found |
| * in the file PATENTS. All contributing project authors may |
| * be found in the AUTHORS file in the root of the source tree. |
| */ |
| |
| #include "video/video_receive_stream.h" |
| |
| #include <stdlib.h> |
| |
| #include <set> |
| #include <string> |
| #include <utility> |
| |
| #include "api/optional.h" |
| #include "call/rtp_stream_receiver_controller_interface.h" |
| #include "call/rtx_receive_stream.h" |
| #include "common_types.h" // NOLINT(build/include) |
| #include "common_video/h264/profile_level_id.h" |
| #include "common_video/libyuv/include/webrtc_libyuv.h" |
| #include "common_video/include/incoming_video_stream.h" |
| #include "modules/rtp_rtcp/include/rtp_receiver.h" |
| #include "modules/rtp_rtcp/include/rtp_rtcp.h" |
| #include "modules/utility/include/process_thread.h" |
| #include "modules/video_coding/frame_object.h" |
| #include "modules/video_coding/include/video_coding.h" |
| #include "modules/video_coding/jitter_estimator.h" |
| #include "modules/video_coding/timing.h" |
| #include "modules/video_coding/utility/ivf_file_writer.h" |
| #include "rtc_base/checks.h" |
| #include "rtc_base/location.h" |
| #include "rtc_base/logging.h" |
| #include "rtc_base/ptr_util.h" |
| #include "rtc_base/trace_event.h" |
| #include "system_wrappers/include/clock.h" |
| #include "system_wrappers/include/field_trial.h" |
| #include "video/call_stats.h" |
| #include "video/receive_statistics_proxy.h" |
| |
| namespace webrtc { |
| |
| namespace { |
| VideoCodec CreateDecoderVideoCodec(const VideoReceiveStream::Decoder& decoder) { |
| VideoCodec codec; |
| memset(&codec, 0, sizeof(codec)); |
| |
| codec.plType = decoder.payload_type; |
| codec.codecType = PayloadStringToCodecType(decoder.payload_name); |
| |
| if (codec.codecType == kVideoCodecVP8) { |
| *(codec.VP8()) = VideoEncoder::GetDefaultVp8Settings(); |
| } else if (codec.codecType == kVideoCodecVP9) { |
| *(codec.VP9()) = VideoEncoder::GetDefaultVp9Settings(); |
| } else if (codec.codecType == kVideoCodecH264) { |
| *(codec.H264()) = VideoEncoder::GetDefaultH264Settings(); |
| codec.H264()->profile = |
| H264::ParseSdpProfileLevelId(decoder.codec_params)->profile; |
| } else if (codec.codecType == kVideoCodecMultiplex) { |
| VideoReceiveStream::Decoder associated_decoder = decoder; |
| associated_decoder.payload_name = CodecTypeToPayloadString(kVideoCodecVP9); |
| VideoCodec associated_codec = CreateDecoderVideoCodec(associated_decoder); |
| associated_codec.codecType = kVideoCodecMultiplex; |
| return associated_codec; |
| } |
| |
| codec.width = 320; |
| codec.height = 180; |
| const int kDefaultStartBitrate = 300; |
| codec.startBitrate = codec.minBitrate = codec.maxBitrate = |
| kDefaultStartBitrate; |
| |
| return codec; |
| } |
| } // namespace |
| |
| namespace internal { |
| |
| VideoReceiveStream::VideoReceiveStream( |
| RtpStreamReceiverControllerInterface* receiver_controller, |
| int num_cpu_cores, |
| PacketRouter* packet_router, |
| VideoReceiveStream::Config config, |
| ProcessThread* process_thread, |
| CallStats* call_stats) |
| : transport_adapter_(config.rtcp_send_transport), |
| config_(std::move(config)), |
| num_cpu_cores_(num_cpu_cores), |
| process_thread_(process_thread), |
| clock_(Clock::GetRealTimeClock()), |
| decode_thread_(&DecodeThreadFunction, |
| this, |
| "DecodingThread", |
| rtc::kHighestPriority), |
| call_stats_(call_stats), |
| rtp_receive_statistics_(ReceiveStatistics::Create(clock_)), |
| timing_(new VCMTiming(clock_)), |
| video_receiver_(clock_, nullptr, this, timing_.get(), this, this), |
| stats_proxy_(&config_, clock_), |
| rtp_video_stream_receiver_(&transport_adapter_, |
| call_stats, |
| packet_router, |
| &config_, |
| rtp_receive_statistics_.get(), |
| &stats_proxy_, |
| process_thread_, |
| this, // NackSender |
| this, // KeyFrameRequestSender |
| this), // OnCompleteFrameCallback |
| rtp_stream_sync_(this) { |
| RTC_LOG(LS_INFO) << "VideoReceiveStream: " << config_.ToString(); |
| |
| RTC_DCHECK(process_thread_); |
| RTC_DCHECK(call_stats_); |
| |
| module_process_sequence_checker_.Detach(); |
| |
| RTC_DCHECK(!config_.decoders.empty()); |
| std::set<int> decoder_payload_types; |
| for (const Decoder& decoder : config_.decoders) { |
| RTC_CHECK(decoder.decoder); |
| RTC_CHECK(decoder_payload_types.find(decoder.payload_type) == |
| decoder_payload_types.end()) |
| << "Duplicate payload type (" << decoder.payload_type |
| << ") for different decoders."; |
| decoder_payload_types.insert(decoder.payload_type); |
| } |
| |
| video_receiver_.SetRenderDelay(config_.render_delay_ms); |
| |
| jitter_estimator_.reset(new VCMJitterEstimator(clock_)); |
| frame_buffer_.reset(new video_coding::FrameBuffer( |
| clock_, jitter_estimator_.get(), timing_.get(), &stats_proxy_)); |
| |
| process_thread_->RegisterModule(&rtp_stream_sync_, RTC_FROM_HERE); |
| |
| // Register with RtpStreamReceiverController. |
| media_receiver_ = receiver_controller->CreateReceiver( |
| config_.rtp.remote_ssrc, &rtp_video_stream_receiver_); |
| if (config_.rtp.rtx_ssrc) { |
| rtx_receive_stream_ = rtc::MakeUnique<RtxReceiveStream>( |
| &rtp_video_stream_receiver_, config.rtp.rtx_associated_payload_types, |
| config_.rtp.remote_ssrc, rtp_receive_statistics_.get()); |
| rtx_receiver_ = receiver_controller->CreateReceiver( |
| config_.rtp.rtx_ssrc, rtx_receive_stream_.get()); |
| } |
| } |
| |
| VideoReceiveStream::~VideoReceiveStream() { |
| RTC_DCHECK_CALLED_SEQUENTIALLY(&worker_sequence_checker_); |
| RTC_LOG(LS_INFO) << "~VideoReceiveStream: " << config_.ToString(); |
| Stop(); |
| |
| process_thread_->DeRegisterModule(&rtp_stream_sync_); |
| } |
| |
| void VideoReceiveStream::SignalNetworkState(NetworkState state) { |
| RTC_DCHECK_CALLED_SEQUENTIALLY(&worker_sequence_checker_); |
| rtp_video_stream_receiver_.SignalNetworkState(state); |
| } |
| |
| bool VideoReceiveStream::DeliverRtcp(const uint8_t* packet, size_t length) { |
| return rtp_video_stream_receiver_.DeliverRtcp(packet, length); |
| } |
| |
| void VideoReceiveStream::SetSync(Syncable* audio_syncable) { |
| RTC_DCHECK_CALLED_SEQUENTIALLY(&worker_sequence_checker_); |
| rtp_stream_sync_.ConfigureSync(audio_syncable); |
| } |
| |
| void VideoReceiveStream::Start() { |
| RTC_DCHECK_CALLED_SEQUENTIALLY(&worker_sequence_checker_); |
| if (decode_thread_.IsRunning()) |
| return; |
| |
| bool protected_by_fec = config_.rtp.protected_by_flexfec || |
| rtp_video_stream_receiver_.IsUlpfecEnabled(); |
| |
| frame_buffer_->Start(); |
| |
| if (rtp_video_stream_receiver_.IsRetransmissionsEnabled() && |
| protected_by_fec) { |
| frame_buffer_->SetProtectionMode(kProtectionNackFEC); |
| } |
| |
| transport_adapter_.Enable(); |
| rtc::VideoSinkInterface<VideoFrame>* renderer = nullptr; |
| if (config_.renderer) { |
| if (config_.disable_prerenderer_smoothing) { |
| renderer = this; |
| } else { |
| incoming_video_stream_.reset( |
| new IncomingVideoStream(config_.render_delay_ms, this)); |
| renderer = incoming_video_stream_.get(); |
| } |
| } |
| RTC_DCHECK(renderer != nullptr); |
| |
| for (const Decoder& decoder : config_.decoders) { |
| video_receiver_.RegisterExternalDecoder(decoder.decoder, |
| decoder.payload_type); |
| VideoCodec codec = CreateDecoderVideoCodec(decoder); |
| RTC_CHECK(rtp_video_stream_receiver_.AddReceiveCodec(codec, |
| decoder.codec_params)); |
| RTC_CHECK_EQ(VCM_OK, video_receiver_.RegisterReceiveCodec( |
| &codec, num_cpu_cores_, false)); |
| } |
| |
| video_stream_decoder_.reset(new VideoStreamDecoder( |
| &video_receiver_, &rtp_video_stream_receiver_, |
| &rtp_video_stream_receiver_, |
| rtp_video_stream_receiver_.IsRetransmissionsEnabled(), protected_by_fec, |
| &stats_proxy_, renderer)); |
| |
| // Make sure we register as a stats observer *after* we've prepared the |
| // |video_stream_decoder_|. |
| call_stats_->RegisterStatsObserver(this); |
| |
| process_thread_->RegisterModule(&video_receiver_, RTC_FROM_HERE); |
| |
| // Start the decode thread |
| video_receiver_.DecoderThreadStarting(); |
| stats_proxy_.DecoderThreadStarting(); |
| decode_thread_.Start(); |
| rtp_video_stream_receiver_.StartReceive(); |
| } |
| |
| void VideoReceiveStream::Stop() { |
| RTC_DCHECK_CALLED_SEQUENTIALLY(&worker_sequence_checker_); |
| rtp_video_stream_receiver_.StopReceive(); |
| |
| stats_proxy_.OnUniqueFramesCounted( |
| rtp_video_stream_receiver_.GetUniqueFramesSeen()); |
| |
| frame_buffer_->Stop(); |
| call_stats_->DeregisterStatsObserver(this); |
| process_thread_->DeRegisterModule(&video_receiver_); |
| |
| if (decode_thread_.IsRunning()) { |
| // TriggerDecoderShutdown will release any waiting decoder thread and make |
| // it stop immediately, instead of waiting for a timeout. Needs to be called |
| // before joining the decoder thread. |
| video_receiver_.TriggerDecoderShutdown(); |
| |
| decode_thread_.Stop(); |
| video_receiver_.DecoderThreadStopped(); |
| stats_proxy_.DecoderThreadStopped(); |
| // Deregister external decoders so they are no longer running during |
| // destruction. This effectively stops the VCM since the decoder thread is |
| // stopped, the VCM is deregistered and no asynchronous decoder threads are |
| // running. |
| for (const Decoder& decoder : config_.decoders) |
| video_receiver_.RegisterExternalDecoder(nullptr, decoder.payload_type); |
| } |
| |
| video_stream_decoder_.reset(); |
| incoming_video_stream_.reset(); |
| transport_adapter_.Disable(); |
| } |
| |
| VideoReceiveStream::Stats VideoReceiveStream::GetStats() const { |
| return stats_proxy_.GetStats(); |
| } |
| |
| void VideoReceiveStream::EnableEncodedFrameRecording(rtc::PlatformFile file, |
| size_t byte_limit) { |
| { |
| rtc::CritScope lock(&ivf_writer_lock_); |
| if (file == rtc::kInvalidPlatformFileValue) { |
| ivf_writer_.reset(); |
| } else { |
| ivf_writer_ = IvfFileWriter::Wrap(rtc::File(file), byte_limit); |
| } |
| } |
| |
| if (file != rtc::kInvalidPlatformFileValue) { |
| // Make a keyframe appear as early as possible in the logs, to give actually |
| // decodable output. |
| RequestKeyFrame(); |
| } |
| } |
| |
| void VideoReceiveStream::AddSecondarySink(RtpPacketSinkInterface* sink) { |
| rtp_video_stream_receiver_.AddSecondarySink(sink); |
| } |
| |
| void VideoReceiveStream::RemoveSecondarySink( |
| const RtpPacketSinkInterface* sink) { |
| rtp_video_stream_receiver_.RemoveSecondarySink(sink); |
| } |
| |
| // TODO(tommi): This method grabs a lock 6 times. |
| void VideoReceiveStream::OnFrame(const VideoFrame& video_frame) { |
| int64_t sync_offset_ms; |
| double estimated_freq_khz; |
| // TODO(tommi): GetStreamSyncOffsetInMs grabs three locks. One inside the |
| // function itself, another in GetChannel() and a third in |
| // GetPlayoutTimestamp. Seems excessive. Anyhow, I'm assuming the function |
| // succeeds most of the time, which leads to grabbing a fourth lock. |
| if (rtp_stream_sync_.GetStreamSyncOffsetInMs(video_frame.timestamp(), |
| video_frame.render_time_ms(), |
| &sync_offset_ms, |
| &estimated_freq_khz)) { |
| // TODO(tommi): OnSyncOffsetUpdated grabs a lock. |
| stats_proxy_.OnSyncOffsetUpdated(sync_offset_ms, estimated_freq_khz); |
| } |
| // config_.renderer must never be null if we're getting this callback. |
| config_.renderer->OnFrame(video_frame); |
| |
| // TODO(tommi): OnRenderFrame grabs a lock too. |
| stats_proxy_.OnRenderedFrame(video_frame); |
| } |
| |
| // TODO(asapersson): Consider moving callback from video_encoder.h or |
| // creating a different callback. |
| EncodedImageCallback::Result VideoReceiveStream::OnEncodedImage( |
| const EncodedImage& encoded_image, |
| const CodecSpecificInfo* codec_specific_info, |
| const RTPFragmentationHeader* fragmentation) { |
| stats_proxy_.OnPreDecode(encoded_image, codec_specific_info); |
| size_t simulcast_idx = 0; |
| if (codec_specific_info->codecType == kVideoCodecVP8) { |
| simulcast_idx = codec_specific_info->codecSpecific.VP8.simulcastIdx; |
| } |
| if (config_.pre_decode_callback) { |
| config_.pre_decode_callback->EncodedFrameCallback(EncodedFrame( |
| encoded_image._buffer, encoded_image._length, encoded_image._frameType, |
| simulcast_idx, encoded_image._timeStamp)); |
| } |
| { |
| rtc::CritScope lock(&ivf_writer_lock_); |
| if (ivf_writer_.get()) { |
| RTC_DCHECK(codec_specific_info); |
| bool ok = ivf_writer_->WriteFrame(encoded_image, |
| codec_specific_info->codecType); |
| RTC_DCHECK(ok); |
| } |
| } |
| |
| return Result(Result::OK, encoded_image._timeStamp); |
| } |
| |
| void VideoReceiveStream::SendNack( |
| const std::vector<uint16_t>& sequence_numbers) { |
| rtp_video_stream_receiver_.RequestPacketRetransmit(sequence_numbers); |
| } |
| |
| void VideoReceiveStream::RequestKeyFrame() { |
| rtp_video_stream_receiver_.RequestKeyFrame(); |
| } |
| |
| void VideoReceiveStream::OnCompleteFrame( |
| std::unique_ptr<video_coding::EncodedFrame> frame) { |
| // TODO(webrtc:9249): Workaround to allow decoding of VP9 SVC stream with |
| // partially enabled inter-layer prediction. |
| frame->id.spatial_layer = 0; |
| |
| int64_t last_continuous_pid = frame_buffer_->InsertFrame(std::move(frame)); |
| if (last_continuous_pid != -1) |
| rtp_video_stream_receiver_.FrameContinuous(last_continuous_pid); |
| } |
| |
| void VideoReceiveStream::OnRttUpdate(int64_t avg_rtt_ms, int64_t max_rtt_ms) { |
| RTC_DCHECK_CALLED_SEQUENTIALLY(&module_process_sequence_checker_); |
| frame_buffer_->UpdateRtt(max_rtt_ms); |
| rtp_video_stream_receiver_.UpdateRtt(max_rtt_ms); |
| video_stream_decoder_->UpdateRtt(max_rtt_ms); |
| } |
| |
| int VideoReceiveStream::id() const { |
| RTC_DCHECK_CALLED_SEQUENTIALLY(&worker_sequence_checker_); |
| return config_.rtp.remote_ssrc; |
| } |
| |
| rtc::Optional<Syncable::Info> VideoReceiveStream::GetInfo() const { |
| RTC_DCHECK_CALLED_SEQUENTIALLY(&module_process_sequence_checker_); |
| Syncable::Info info; |
| |
| RtpReceiver* rtp_receiver = rtp_video_stream_receiver_.GetRtpReceiver(); |
| RTC_DCHECK(rtp_receiver); |
| if (!rtp_receiver->GetLatestTimestamps( |
| &info.latest_received_capture_timestamp, |
| &info.latest_receive_time_ms)) |
| return rtc::nullopt; |
| |
| RtpRtcp* rtp_rtcp = rtp_video_stream_receiver_.rtp_rtcp(); |
| RTC_DCHECK(rtp_rtcp); |
| if (rtp_rtcp->RemoteNTP(&info.capture_time_ntp_secs, |
| &info.capture_time_ntp_frac, |
| nullptr, |
| nullptr, |
| &info.capture_time_source_clock) != 0) { |
| return rtc::nullopt; |
| } |
| |
| info.current_delay_ms = video_receiver_.Delay(); |
| return info; |
| } |
| |
| uint32_t VideoReceiveStream::GetPlayoutTimestamp() const { |
| RTC_NOTREACHED(); |
| return 0; |
| } |
| |
| void VideoReceiveStream::SetMinimumPlayoutDelay(int delay_ms) { |
| RTC_DCHECK_CALLED_SEQUENTIALLY(&module_process_sequence_checker_); |
| video_receiver_.SetMinimumPlayoutDelay(delay_ms); |
| } |
| |
| void VideoReceiveStream::DecodeThreadFunction(void* ptr) { |
| while (static_cast<VideoReceiveStream*>(ptr)->Decode()) { |
| } |
| } |
| |
| bool VideoReceiveStream::Decode() { |
| TRACE_EVENT0("webrtc", "VideoReceiveStream::Decode"); |
| static const int kMaxWaitForFrameMs = 3000; |
| static const int kMaxWaitForKeyFrameMs = 200; |
| |
| int wait_ms = keyframe_required_ ? kMaxWaitForKeyFrameMs : kMaxWaitForFrameMs; |
| std::unique_ptr<video_coding::EncodedFrame> frame; |
| // TODO(philipel): Call NextFrame with |keyframe_required| argument when |
| // downstream project has been fixed. |
| video_coding::FrameBuffer::ReturnReason res = |
| frame_buffer_->NextFrame(wait_ms, &frame); |
| |
| if (res == video_coding::FrameBuffer::ReturnReason::kStopped) { |
| return false; |
| } |
| |
| if (frame) { |
| int64_t now_ms = clock_->TimeInMilliseconds(); |
| RTC_DCHECK_EQ(res, video_coding::FrameBuffer::ReturnReason::kFrameFound); |
| int decode_result = video_receiver_.Decode(frame.get()); |
| if (decode_result == WEBRTC_VIDEO_CODEC_OK || |
| decode_result == WEBRTC_VIDEO_CODEC_OK_REQUEST_KEYFRAME) { |
| keyframe_required_ = false; |
| frame_decoded_ = true; |
| rtp_video_stream_receiver_.FrameDecoded(frame->id.picture_id); |
| |
| if (decode_result == WEBRTC_VIDEO_CODEC_OK_REQUEST_KEYFRAME) |
| RequestKeyFrame(); |
| } else if (!frame_decoded_ || !keyframe_required_ || |
| (last_keyframe_request_ms_ + kMaxWaitForKeyFrameMs < now_ms)) { |
| keyframe_required_ = true; |
| // TODO(philipel): Remove this keyframe request when downstream project |
| // has been fixed. |
| RequestKeyFrame(); |
| last_keyframe_request_ms_ = now_ms; |
| } |
| } else { |
| RTC_DCHECK_EQ(res, video_coding::FrameBuffer::ReturnReason::kTimeout); |
| int64_t now_ms = clock_->TimeInMilliseconds(); |
| rtc::Optional<int64_t> last_packet_ms = |
| rtp_video_stream_receiver_.LastReceivedPacketMs(); |
| rtc::Optional<int64_t> last_keyframe_packet_ms = |
| rtp_video_stream_receiver_.LastReceivedKeyframePacketMs(); |
| |
| // To avoid spamming keyframe requests for a stream that is not active we |
| // check if we have received a packet within the last 5 seconds. |
| bool stream_is_active = last_packet_ms && now_ms - *last_packet_ms < 5000; |
| if (!stream_is_active) |
| stats_proxy_.OnStreamInactive(); |
| |
| // If we recently have been receiving packets belonging to a keyframe then |
| // we assume a keyframe is currently being received. |
| bool receiving_keyframe = |
| last_keyframe_packet_ms && |
| now_ms - *last_keyframe_packet_ms < kMaxWaitForKeyFrameMs; |
| |
| if (stream_is_active && !receiving_keyframe) { |
| RTC_LOG(LS_WARNING) << "No decodable frame in " << wait_ms |
| << " ms, requesting keyframe."; |
| RequestKeyFrame(); |
| } |
| } |
| return true; |
| } |
| } // namespace internal |
| } // namespace webrtc |