| /* |
| * Copyright (c) 2012 The WebRTC project authors. All Rights Reserved. |
| * |
| * Use of this source code is governed by a BSD-style license |
| * that can be found in the LICENSE file in the root of the source |
| * tree. An additional intellectual property rights grant can be found |
| * in the file PATENTS. All contributing project authors may |
| * be found in the AUTHORS file in the root of the source tree. |
| */ |
| |
| #include "video/rtp_video_stream_receiver.h" |
| |
| #include <algorithm> |
| #include <utility> |
| #include <vector> |
| |
| #include "call/video_config.h" |
| #include "common_types.h" // NOLINT(build/include) |
| #include "media/base/mediaconstants.h" |
| #include "modules/pacing/packet_router.h" |
| #include "modules/remote_bitrate_estimator/include/remote_bitrate_estimator.h" |
| #include "modules/rtp_rtcp/include/receive_statistics.h" |
| #include "modules/rtp_rtcp/include/rtp_cvo.h" |
| #include "modules/rtp_rtcp/include/rtp_receiver.h" |
| #include "modules/rtp_rtcp/include/rtp_rtcp.h" |
| #include "modules/rtp_rtcp/include/ulpfec_receiver.h" |
| #include "modules/rtp_rtcp/source/rtp_header_extensions.h" |
| #include "modules/rtp_rtcp/source/rtp_packet_received.h" |
| #include "modules/video_coding/frame_object.h" |
| #include "modules/video_coding/h264_sprop_parameter_sets.h" |
| #include "modules/video_coding/h264_sps_pps_tracker.h" |
| #include "modules/video_coding/packet_buffer.h" |
| #include "modules/video_coding/video_coding_impl.h" |
| #include "rtc_base/checks.h" |
| #include "rtc_base/location.h" |
| #include "rtc_base/logging.h" |
| #include "rtc_base/system/fallthrough.h" |
| #include "system_wrappers/include/field_trial.h" |
| #include "system_wrappers/include/metrics.h" |
| #include "system_wrappers/include/timestamp_extrapolator.h" |
| #include "video/receive_statistics_proxy.h" |
| |
| namespace webrtc { |
| |
| namespace { |
| // TODO(philipel): Change kPacketBufferStartSize back to 32 in M63 see: |
| // crbug.com/752886 |
| constexpr int kPacketBufferStartSize = 512; |
| constexpr int kPacketBufferMaxSixe = 2048; |
| } |
| |
| std::unique_ptr<RtpRtcp> CreateRtpRtcpModule( |
| ReceiveStatistics* receive_statistics, |
| Transport* outgoing_transport, |
| RtcpRttStats* rtt_stats, |
| RtcpPacketTypeCounterObserver* rtcp_packet_type_counter_observer, |
| TransportSequenceNumberAllocator* transport_sequence_number_allocator) { |
| RtpRtcp::Configuration configuration; |
| configuration.audio = false; |
| configuration.receiver_only = true; |
| configuration.receive_statistics = receive_statistics; |
| configuration.outgoing_transport = outgoing_transport; |
| configuration.intra_frame_callback = nullptr; |
| configuration.rtt_stats = rtt_stats; |
| configuration.rtcp_packet_type_counter_observer = |
| rtcp_packet_type_counter_observer; |
| configuration.transport_sequence_number_allocator = |
| transport_sequence_number_allocator; |
| configuration.send_bitrate_observer = nullptr; |
| configuration.send_frame_count_observer = nullptr; |
| configuration.send_side_delay_observer = nullptr; |
| configuration.send_packet_observer = nullptr; |
| configuration.bandwidth_callback = nullptr; |
| configuration.transport_feedback_callback = nullptr; |
| |
| std::unique_ptr<RtpRtcp> rtp_rtcp(RtpRtcp::CreateRtpRtcp(configuration)); |
| rtp_rtcp->SetRTCPStatus(RtcpMode::kCompound); |
| |
| return rtp_rtcp; |
| } |
| |
| static const int kPacketLogIntervalMs = 10000; |
| |
| RtpVideoStreamReceiver::RtpVideoStreamReceiver( |
| Transport* transport, |
| RtcpRttStats* rtt_stats, |
| PacketRouter* packet_router, |
| const VideoReceiveStream::Config* config, |
| ReceiveStatistics* rtp_receive_statistics, |
| ReceiveStatisticsProxy* receive_stats_proxy, |
| ProcessThread* process_thread, |
| NackSender* nack_sender, |
| KeyFrameRequestSender* keyframe_request_sender, |
| video_coding::OnCompleteFrameCallback* complete_frame_callback) |
| : clock_(Clock::GetRealTimeClock()), |
| config_(*config), |
| packet_router_(packet_router), |
| process_thread_(process_thread), |
| ntp_estimator_(clock_), |
| rtp_header_extensions_(config_.rtp.extensions), |
| rtp_receiver_(RtpReceiver::CreateVideoReceiver(clock_, |
| this, |
| this, |
| &rtp_payload_registry_)), |
| rtp_receive_statistics_(rtp_receive_statistics), |
| ulpfec_receiver_(UlpfecReceiver::Create(config->rtp.remote_ssrc, this)), |
| receiving_(false), |
| last_packet_log_ms_(-1), |
| rtp_rtcp_(CreateRtpRtcpModule(rtp_receive_statistics_, |
| transport, |
| rtt_stats, |
| receive_stats_proxy, |
| packet_router)), |
| complete_frame_callback_(complete_frame_callback), |
| keyframe_request_sender_(keyframe_request_sender), |
| has_received_frame_(false) { |
| constexpr bool remb_candidate = true; |
| packet_router_->AddReceiveRtpModule(rtp_rtcp_.get(), remb_candidate); |
| rtp_receive_statistics_->RegisterRtpStatisticsCallback(receive_stats_proxy); |
| rtp_receive_statistics_->RegisterRtcpStatisticsCallback(receive_stats_proxy); |
| |
| RTC_DCHECK(config_.rtp.rtcp_mode != RtcpMode::kOff) |
| << "A stream should not be configured with RTCP disabled. This value is " |
| "reserved for internal usage."; |
| RTC_DCHECK(config_.rtp.remote_ssrc != 0); |
| // TODO(pbos): What's an appropriate local_ssrc for receive-only streams? |
| RTC_DCHECK(config_.rtp.local_ssrc != 0); |
| RTC_DCHECK(config_.rtp.remote_ssrc != config_.rtp.local_ssrc); |
| |
| rtp_rtcp_->SetRTCPStatus(config_.rtp.rtcp_mode); |
| rtp_rtcp_->SetSSRC(config_.rtp.local_ssrc); |
| rtp_rtcp_->SetRemoteSSRC(config_.rtp.remote_ssrc); |
| rtp_rtcp_->SetKeyFrameRequestMethod(kKeyFrameReqPliRtcp); |
| |
| static const int kMaxPacketAgeToNack = 450; |
| const int max_reordering_threshold = (config_.rtp.nack.rtp_history_ms > 0) |
| ? kMaxPacketAgeToNack |
| : kDefaultMaxReorderingThreshold; |
| rtp_receive_statistics_->SetMaxReorderingThreshold(max_reordering_threshold); |
| |
| if (config_.rtp.rtx_ssrc) { |
| // Needed for rtp_payload_registry_.RtxEnabled(). |
| rtp_payload_registry_.SetRtxSsrc(config_.rtp.rtx_ssrc); |
| } |
| |
| if (IsUlpfecEnabled()) { |
| VideoCodec ulpfec_codec = {}; |
| ulpfec_codec.codecType = kVideoCodecULPFEC; |
| ulpfec_codec.plType = config_.rtp.ulpfec_payload_type; |
| RTC_CHECK(AddReceiveCodec(ulpfec_codec)); |
| } |
| |
| if (IsRedEnabled()) { |
| VideoCodec red_codec = {}; |
| red_codec.codecType = kVideoCodecRED; |
| red_codec.plType = config_.rtp.red_payload_type; |
| RTC_CHECK(AddReceiveCodec(red_codec)); |
| } |
| |
| if (config_.rtp.rtcp_xr.receiver_reference_time_report) |
| rtp_rtcp_->SetRtcpXrRrtrStatus(true); |
| |
| // Stats callback for CNAME changes. |
| rtp_rtcp_->RegisterRtcpStatisticsCallback(receive_stats_proxy); |
| |
| process_thread_->RegisterModule(rtp_rtcp_.get(), RTC_FROM_HERE); |
| |
| if (config_.rtp.nack.rtp_history_ms != 0) { |
| nack_module_.reset( |
| new NackModule(clock_, nack_sender, keyframe_request_sender)); |
| process_thread_->RegisterModule(nack_module_.get(), RTC_FROM_HERE); |
| } |
| |
| packet_buffer_ = video_coding::PacketBuffer::Create( |
| clock_, kPacketBufferStartSize, kPacketBufferMaxSixe, this); |
| reference_finder_.reset(new video_coding::RtpFrameReferenceFinder(this)); |
| } |
| |
| RtpVideoStreamReceiver::~RtpVideoStreamReceiver() { |
| RTC_DCHECK(secondary_sinks_.empty()); |
| |
| if (nack_module_) { |
| process_thread_->DeRegisterModule(nack_module_.get()); |
| } |
| |
| process_thread_->DeRegisterModule(rtp_rtcp_.get()); |
| |
| packet_router_->RemoveReceiveRtpModule(rtp_rtcp_.get()); |
| UpdateHistograms(); |
| } |
| |
| bool RtpVideoStreamReceiver::AddReceiveCodec( |
| const VideoCodec& video_codec, |
| const std::map<std::string, std::string>& codec_params) { |
| pt_codec_params_.insert(make_pair(video_codec.plType, codec_params)); |
| return AddReceiveCodec(video_codec); |
| } |
| |
| bool RtpVideoStreamReceiver::AddReceiveCodec(const VideoCodec& video_codec) { |
| int8_t old_pltype = -1; |
| if (rtp_payload_registry_.ReceivePayloadType(video_codec, &old_pltype) != |
| -1) { |
| rtp_payload_registry_.DeRegisterReceivePayload(old_pltype); |
| } |
| return rtp_payload_registry_.RegisterReceivePayload(video_codec) == 0; |
| } |
| |
| uint32_t RtpVideoStreamReceiver::GetRemoteSsrc() const { |
| return config_.rtp.remote_ssrc; |
| } |
| |
| int RtpVideoStreamReceiver::GetCsrcs(uint32_t* csrcs) const { |
| return rtp_receiver_->CSRCs(csrcs); |
| } |
| |
| RtpReceiver* RtpVideoStreamReceiver::GetRtpReceiver() const { |
| return rtp_receiver_.get(); |
| } |
| |
| int32_t RtpVideoStreamReceiver::OnReceivedPayloadData( |
| const uint8_t* payload_data, |
| size_t payload_size, |
| const WebRtcRTPHeader* rtp_header) { |
| WebRtcRTPHeader rtp_header_with_ntp = *rtp_header; |
| rtp_header_with_ntp.ntp_time_ms = |
| ntp_estimator_.Estimate(rtp_header->header.timestamp); |
| VCMPacket packet(payload_data, payload_size, rtp_header_with_ntp); |
| packet.timesNacked = |
| nack_module_ ? nack_module_->OnReceivedPacket(packet) : -1; |
| packet.receive_time_ms = clock_->TimeInMilliseconds(); |
| |
| // In the case of a video stream without picture ids and no rtx the |
| // RtpFrameReferenceFinder will need to know about padding to |
| // correctly calculate frame references. |
| if (packet.sizeBytes == 0) { |
| reference_finder_->PaddingReceived(packet.seqNum); |
| packet_buffer_->PaddingReceived(packet.seqNum); |
| return 0; |
| } |
| |
| if (packet.codec == kVideoCodecH264) { |
| // Only when we start to receive packets will we know what payload type |
| // that will be used. When we know the payload type insert the correct |
| // sps/pps into the tracker. |
| if (packet.payloadType != last_payload_type_) { |
| last_payload_type_ = packet.payloadType; |
| InsertSpsPpsIntoTracker(packet.payloadType); |
| } |
| |
| switch (tracker_.CopyAndFixBitstream(&packet)) { |
| case video_coding::H264SpsPpsTracker::kRequestKeyframe: |
| keyframe_request_sender_->RequestKeyFrame(); |
| RTC_FALLTHROUGH(); |
| case video_coding::H264SpsPpsTracker::kDrop: |
| return 0; |
| case video_coding::H264SpsPpsTracker::kInsert: |
| break; |
| } |
| |
| } else { |
| uint8_t* data = new uint8_t[packet.sizeBytes]; |
| memcpy(data, packet.dataPtr, packet.sizeBytes); |
| packet.dataPtr = data; |
| } |
| |
| packet_buffer_->InsertPacket(&packet); |
| return 0; |
| } |
| |
| void RtpVideoStreamReceiver::OnRecoveredPacket(const uint8_t* rtp_packet, |
| size_t rtp_packet_length) { |
| RtpPacketReceived packet; |
| if (!packet.Parse(rtp_packet, rtp_packet_length)) |
| return; |
| packet.IdentifyExtensions(rtp_header_extensions_); |
| packet.set_payload_type_frequency(kVideoPayloadTypeFrequency); |
| |
| RTPHeader header; |
| packet.GetHeader(&header); |
| ReceivePacket(rtp_packet, rtp_packet_length, header); |
| } |
| |
| // TODO(pbos): Remove as soon as audio can handle a changing payload type |
| // without this callback. |
| int32_t RtpVideoStreamReceiver::OnInitializeDecoder( |
| const int payload_type, |
| const SdpAudioFormat& audio_format, |
| const uint32_t rate) { |
| RTC_NOTREACHED(); |
| return 0; |
| } |
| |
| // This method handles both regular RTP packets and packets recovered |
| // via FlexFEC. |
| void RtpVideoStreamReceiver::OnRtpPacket(const RtpPacketReceived& packet) { |
| RTC_DCHECK_CALLED_SEQUENTIALLY(&worker_task_checker_); |
| |
| if (!receiving_) { |
| return; |
| } |
| |
| if (!packet.recovered()) { |
| int64_t now_ms = clock_->TimeInMilliseconds(); |
| |
| // Periodically log the RTP header of incoming packets. |
| if (now_ms - last_packet_log_ms_ > kPacketLogIntervalMs) { |
| std::stringstream ss; |
| ss << "Packet received on SSRC: " << packet.Ssrc() |
| << " with payload type: " << static_cast<int>(packet.PayloadType()) |
| << ", timestamp: " << packet.Timestamp() |
| << ", sequence number: " << packet.SequenceNumber() |
| << ", arrival time: " << packet.arrival_time_ms(); |
| int32_t time_offset; |
| if (packet.GetExtension<TransmissionOffset>(&time_offset)) { |
| ss << ", toffset: " << time_offset; |
| } |
| uint32_t send_time; |
| if (packet.GetExtension<AbsoluteSendTime>(&send_time)) { |
| ss << ", abs send time: " << send_time; |
| } |
| RTC_LOG(LS_INFO) << ss.str(); |
| last_packet_log_ms_ = now_ms; |
| } |
| } |
| |
| // TODO(nisse): Delete use of GetHeader, but needs refactoring of |
| // ReceivePacket and IncomingPacket methods below. |
| RTPHeader header; |
| packet.GetHeader(&header); |
| |
| header.payload_type_frequency = kVideoPayloadTypeFrequency; |
| |
| bool in_order = IsPacketInOrder(header); |
| |
| ReceivePacket(packet.data(), packet.size(), header); |
| // Update receive statistics after ReceivePacket. |
| // Receive statistics will be reset if the payload type changes (make sure |
| // that the first packet is included in the stats). |
| if (!packet.recovered()) { |
| // TODO(nisse): We should pass a recovered flag to stats, to aid |
| // fixing bug bugs.webrtc.org/6339. |
| rtp_receive_statistics_->IncomingPacket( |
| header, packet.size(), IsPacketRetransmitted(header, in_order)); |
| } |
| |
| for (RtpPacketSinkInterface* secondary_sink : secondary_sinks_) { |
| secondary_sink->OnRtpPacket(packet); |
| } |
| } |
| |
| int32_t RtpVideoStreamReceiver::RequestKeyFrame() { |
| return rtp_rtcp_->RequestKeyFrame(); |
| } |
| |
| bool RtpVideoStreamReceiver::IsUlpfecEnabled() const { |
| return config_.rtp.ulpfec_payload_type != -1; |
| } |
| |
| bool RtpVideoStreamReceiver::IsRedEnabled() const { |
| return config_.rtp.red_payload_type != -1; |
| } |
| |
| bool RtpVideoStreamReceiver::IsRetransmissionsEnabled() const { |
| return config_.rtp.nack.rtp_history_ms > 0; |
| } |
| |
| void RtpVideoStreamReceiver::RequestPacketRetransmit( |
| const std::vector<uint16_t>& sequence_numbers) { |
| rtp_rtcp_->SendNack(sequence_numbers); |
| } |
| |
| int32_t RtpVideoStreamReceiver::ResendPackets(const uint16_t* sequence_numbers, |
| uint16_t length) { |
| return rtp_rtcp_->SendNACK(sequence_numbers, length); |
| } |
| |
| void RtpVideoStreamReceiver::OnReceivedFrame( |
| std::unique_ptr<video_coding::RtpFrameObject> frame) { |
| if (!has_received_frame_) { |
| has_received_frame_ = true; |
| if (frame->FrameType() != kVideoFrameKey) |
| keyframe_request_sender_->RequestKeyFrame(); |
| } |
| |
| reference_finder_->ManageFrame(std::move(frame)); |
| } |
| |
| void RtpVideoStreamReceiver::OnCompleteFrame( |
| std::unique_ptr<video_coding::EncodedFrame> frame) { |
| { |
| rtc::CritScope lock(&last_seq_num_cs_); |
| video_coding::RtpFrameObject* rtp_frame = |
| static_cast<video_coding::RtpFrameObject*>(frame.get()); |
| last_seq_num_for_pic_id_[rtp_frame->picture_id] = rtp_frame->last_seq_num(); |
| } |
| complete_frame_callback_->OnCompleteFrame(std::move(frame)); |
| } |
| |
| void RtpVideoStreamReceiver::OnRttUpdate(int64_t avg_rtt_ms, |
| int64_t max_rtt_ms) { |
| if (nack_module_) |
| nack_module_->UpdateRtt(max_rtt_ms); |
| } |
| |
| rtc::Optional<int64_t> RtpVideoStreamReceiver::LastReceivedPacketMs() const { |
| return packet_buffer_->LastReceivedPacketMs(); |
| } |
| |
| rtc::Optional<int64_t> RtpVideoStreamReceiver::LastReceivedKeyframePacketMs() |
| const { |
| return packet_buffer_->LastReceivedKeyframePacketMs(); |
| } |
| |
| void RtpVideoStreamReceiver::AddSecondarySink(RtpPacketSinkInterface* sink) { |
| RTC_DCHECK_CALLED_SEQUENTIALLY(&worker_task_checker_); |
| RTC_DCHECK(std::find(secondary_sinks_.cbegin(), secondary_sinks_.cend(), |
| sink) == secondary_sinks_.cend()); |
| secondary_sinks_.push_back(sink); |
| } |
| |
| void RtpVideoStreamReceiver::RemoveSecondarySink( |
| const RtpPacketSinkInterface* sink) { |
| RTC_DCHECK_CALLED_SEQUENTIALLY(&worker_task_checker_); |
| auto it = std::find(secondary_sinks_.begin(), secondary_sinks_.end(), sink); |
| if (it == secondary_sinks_.end()) { |
| // We might be rolling-back a call whose setup failed mid-way. In such a |
| // case, it's simpler to remove "everything" rather than remember what |
| // has already been added. |
| RTC_LOG(LS_WARNING) << "Removal of unknown sink."; |
| return; |
| } |
| secondary_sinks_.erase(it); |
| } |
| |
| void RtpVideoStreamReceiver::ReceivePacket(const uint8_t* packet, |
| size_t packet_length, |
| const RTPHeader& header) { |
| if (rtp_payload_registry_.IsRed(header)) { |
| ParseAndHandleEncapsulatingHeader(packet, packet_length, header); |
| return; |
| } |
| const uint8_t* payload = packet + header.headerLength; |
| assert(packet_length >= header.headerLength); |
| size_t payload_length = packet_length - header.headerLength; |
| const auto pl = |
| rtp_payload_registry_.PayloadTypeToPayload(header.payloadType); |
| if (pl) { |
| rtp_receiver_->IncomingRtpPacket(header, payload, payload_length, |
| pl->typeSpecific); |
| } |
| } |
| |
| void RtpVideoStreamReceiver::ParseAndHandleEncapsulatingHeader( |
| const uint8_t* packet, size_t packet_length, const RTPHeader& header) { |
| RTC_DCHECK_CALLED_SEQUENTIALLY(&worker_task_checker_); |
| if (rtp_payload_registry_.IsRed(header)) { |
| int8_t ulpfec_pt = rtp_payload_registry_.ulpfec_payload_type(); |
| if (packet[header.headerLength] == ulpfec_pt) { |
| rtp_receive_statistics_->FecPacketReceived(header, packet_length); |
| // Notify video_receiver about received FEC packets to avoid NACKing these |
| // packets. |
| NotifyReceiverOfFecPacket(header); |
| } |
| if (ulpfec_receiver_->AddReceivedRedPacket(header, packet, packet_length, |
| ulpfec_pt) != 0) { |
| return; |
| } |
| ulpfec_receiver_->ProcessReceivedFec(); |
| } |
| } |
| |
| void RtpVideoStreamReceiver::NotifyReceiverOfFecPacket( |
| const RTPHeader& header) { |
| int8_t last_media_payload_type = |
| rtp_payload_registry_.last_received_media_payload_type(); |
| if (last_media_payload_type < 0) { |
| RTC_LOG(LS_WARNING) << "Failed to get last media payload type."; |
| return; |
| } |
| // Fake an empty media packet. |
| WebRtcRTPHeader rtp_header = {}; |
| rtp_header.header = header; |
| rtp_header.header.payloadType = last_media_payload_type; |
| rtp_header.header.paddingLength = 0; |
| const auto pl = |
| rtp_payload_registry_.PayloadTypeToPayload(last_media_payload_type); |
| if (!pl) { |
| RTC_LOG(LS_WARNING) << "Failed to get payload specifics."; |
| return; |
| } |
| rtp_header.type.Video.codec = pl->typeSpecific.video_payload().videoCodecType; |
| rtp_header.type.Video.rotation = kVideoRotation_0; |
| if (header.extension.hasVideoRotation) { |
| rtp_header.type.Video.rotation = header.extension.videoRotation; |
| } |
| rtp_header.type.Video.content_type = VideoContentType::UNSPECIFIED; |
| if (header.extension.hasVideoContentType) { |
| rtp_header.type.Video.content_type = header.extension.videoContentType; |
| } |
| rtp_header.type.Video.video_timing = {0u, 0u, 0u, 0u, 0u, 0u, false}; |
| if (header.extension.has_video_timing) { |
| rtp_header.type.Video.video_timing = header.extension.video_timing; |
| } |
| rtp_header.type.Video.playout_delay = header.extension.playout_delay; |
| |
| OnReceivedPayloadData(nullptr, 0, &rtp_header); |
| } |
| |
| bool RtpVideoStreamReceiver::DeliverRtcp(const uint8_t* rtcp_packet, |
| size_t rtcp_packet_length) { |
| RTC_DCHECK_CALLED_SEQUENTIALLY(&worker_task_checker_); |
| |
| if (!receiving_) { |
| return false; |
| } |
| |
| rtp_rtcp_->IncomingRtcpPacket(rtcp_packet, rtcp_packet_length); |
| |
| int64_t rtt = 0; |
| rtp_rtcp_->RTT(rtp_receiver_->SSRC(), &rtt, nullptr, nullptr, nullptr); |
| if (rtt == 0) { |
| // Waiting for valid rtt. |
| return true; |
| } |
| uint32_t ntp_secs = 0; |
| uint32_t ntp_frac = 0; |
| uint32_t rtp_timestamp = 0; |
| uint32_t recieved_ntp_secs = 0; |
| uint32_t recieved_ntp_frac = 0; |
| if (rtp_rtcp_->RemoteNTP(&ntp_secs, &ntp_frac, &recieved_ntp_secs, |
| &recieved_ntp_frac, &rtp_timestamp) != 0) { |
| // Waiting for RTCP. |
| return true; |
| } |
| NtpTime recieved_ntp(recieved_ntp_secs, recieved_ntp_frac); |
| int64_t time_since_recieved = |
| clock_->CurrentNtpInMilliseconds() - recieved_ntp.ToMs(); |
| // Don't use old SRs to estimate time. |
| if (time_since_recieved <= 1) { |
| ntp_estimator_.UpdateRtcpTimestamp(rtt, ntp_secs, ntp_frac, rtp_timestamp); |
| } |
| |
| return true; |
| } |
| |
| void RtpVideoStreamReceiver::FrameContinuous(int64_t picture_id) { |
| if (!nack_module_) |
| return; |
| |
| int seq_num = -1; |
| { |
| rtc::CritScope lock(&last_seq_num_cs_); |
| auto seq_num_it = last_seq_num_for_pic_id_.find(picture_id); |
| if (seq_num_it != last_seq_num_for_pic_id_.end()) |
| seq_num = seq_num_it->second; |
| } |
| if (seq_num != -1) |
| nack_module_->ClearUpTo(seq_num); |
| } |
| |
| void RtpVideoStreamReceiver::FrameDecoded(int64_t picture_id) { |
| int seq_num = -1; |
| { |
| rtc::CritScope lock(&last_seq_num_cs_); |
| auto seq_num_it = last_seq_num_for_pic_id_.find(picture_id); |
| if (seq_num_it != last_seq_num_for_pic_id_.end()) { |
| seq_num = seq_num_it->second; |
| last_seq_num_for_pic_id_.erase(last_seq_num_for_pic_id_.begin(), |
| ++seq_num_it); |
| } |
| } |
| if (seq_num != -1) { |
| packet_buffer_->ClearTo(seq_num); |
| reference_finder_->ClearTo(seq_num); |
| } |
| } |
| |
| void RtpVideoStreamReceiver::SignalNetworkState(NetworkState state) { |
| rtp_rtcp_->SetRTCPStatus(state == kNetworkUp ? config_.rtp.rtcp_mode |
| : RtcpMode::kOff); |
| } |
| |
| int RtpVideoStreamReceiver::GetUniqueFramesSeen() const { |
| return packet_buffer_->GetUniqueFramesSeen(); |
| } |
| |
| void RtpVideoStreamReceiver::StartReceive() { |
| RTC_DCHECK_CALLED_SEQUENTIALLY(&worker_task_checker_); |
| receiving_ = true; |
| } |
| |
| void RtpVideoStreamReceiver::StopReceive() { |
| RTC_DCHECK_CALLED_SEQUENTIALLY(&worker_task_checker_); |
| receiving_ = false; |
| } |
| |
| bool RtpVideoStreamReceiver::IsPacketInOrder(const RTPHeader& header) const { |
| StreamStatistician* statistician = |
| rtp_receive_statistics_->GetStatistician(header.ssrc); |
| if (!statistician) |
| return false; |
| return statistician->IsPacketInOrder(header.sequenceNumber); |
| } |
| |
| bool RtpVideoStreamReceiver::IsPacketRetransmitted(const RTPHeader& header, |
| bool in_order) const { |
| // Retransmissions are handled separately if RTX is enabled. |
| if (rtp_payload_registry_.RtxEnabled()) |
| return false; |
| StreamStatistician* statistician = |
| rtp_receive_statistics_->GetStatistician(header.ssrc); |
| if (!statistician) |
| return false; |
| // Check if this is a retransmission. |
| int64_t min_rtt = 0; |
| rtp_rtcp_->RTT(config_.rtp.remote_ssrc, nullptr, nullptr, &min_rtt, nullptr); |
| return !in_order && |
| statistician->IsRetransmitOfOldPacket(header, min_rtt); |
| } |
| |
| void RtpVideoStreamReceiver::UpdateHistograms() { |
| FecPacketCounter counter = ulpfec_receiver_->GetPacketCounter(); |
| if (counter.first_packet_time_ms == -1) |
| return; |
| |
| int64_t elapsed_sec = |
| (clock_->TimeInMilliseconds() - counter.first_packet_time_ms) / 1000; |
| if (elapsed_sec < metrics::kMinRunTimeInSeconds) |
| return; |
| |
| if (counter.num_packets > 0) { |
| RTC_HISTOGRAM_PERCENTAGE( |
| "WebRTC.Video.ReceivedFecPacketsInPercent", |
| static_cast<int>(counter.num_fec_packets * 100 / counter.num_packets)); |
| } |
| if (counter.num_fec_packets > 0) { |
| RTC_HISTOGRAM_PERCENTAGE("WebRTC.Video.RecoveredMediaPacketsInPercentOfFec", |
| static_cast<int>(counter.num_recovered_packets * |
| 100 / counter.num_fec_packets)); |
| } |
| } |
| |
| void RtpVideoStreamReceiver::InsertSpsPpsIntoTracker(uint8_t payload_type) { |
| auto codec_params_it = pt_codec_params_.find(payload_type); |
| if (codec_params_it == pt_codec_params_.end()) |
| return; |
| |
| RTC_LOG(LS_INFO) << "Found out of band supplied codec parameters for" |
| << " payload type: " << static_cast<int>(payload_type); |
| |
| H264SpropParameterSets sprop_decoder; |
| auto sprop_base64_it = |
| codec_params_it->second.find(cricket::kH264FmtpSpropParameterSets); |
| |
| if (sprop_base64_it == codec_params_it->second.end()) |
| return; |
| |
| if (!sprop_decoder.DecodeSprop(sprop_base64_it->second.c_str())) |
| return; |
| |
| tracker_.InsertSpsPpsNalus(sprop_decoder.sps_nalu(), |
| sprop_decoder.pps_nalu()); |
| } |
| |
| } // namespace webrtc |