| /* |
| * Copyright (c) 2015 The WebRTC project authors. All Rights Reserved. |
| * |
| * Use of this source code is governed by a BSD-style license |
| * that can be found in the LICENSE file in the root of the source |
| * tree. An additional intellectual property rights grant can be found |
| * in the file PATENTS. All contributing project authors may |
| * be found in the AUTHORS file in the root of the source tree. |
| */ |
| |
| #include "webrtc/video/payload_router.h" |
| |
| #include "webrtc/modules/rtp_rtcp/include/rtp_rtcp.h" |
| #include "webrtc/modules/rtp_rtcp/include/rtp_rtcp_defines.h" |
| #include "webrtc/modules/video_coding/include/video_codec_interface.h" |
| #include "webrtc/rtc_base/checks.h" |
| |
| namespace webrtc { |
| |
| namespace { |
| // Map information from info into rtp. |
| void CopyCodecSpecific(const CodecSpecificInfo* info, RTPVideoHeader* rtp) { |
| RTC_DCHECK(info); |
| switch (info->codecType) { |
| case kVideoCodecVP8: { |
| rtp->codec = kRtpVideoVp8; |
| rtp->codecHeader.VP8.InitRTPVideoHeaderVP8(); |
| rtp->codecHeader.VP8.pictureId = info->codecSpecific.VP8.pictureId; |
| rtp->codecHeader.VP8.nonReference = info->codecSpecific.VP8.nonReference; |
| rtp->codecHeader.VP8.temporalIdx = info->codecSpecific.VP8.temporalIdx; |
| rtp->codecHeader.VP8.layerSync = info->codecSpecific.VP8.layerSync; |
| rtp->codecHeader.VP8.tl0PicIdx = info->codecSpecific.VP8.tl0PicIdx; |
| rtp->codecHeader.VP8.keyIdx = info->codecSpecific.VP8.keyIdx; |
| rtp->simulcastIdx = info->codecSpecific.VP8.simulcastIdx; |
| return; |
| } |
| case kVideoCodecVP9: { |
| rtp->codec = kRtpVideoVp9; |
| rtp->codecHeader.VP9.InitRTPVideoHeaderVP9(); |
| rtp->codecHeader.VP9.inter_pic_predicted = |
| info->codecSpecific.VP9.inter_pic_predicted; |
| rtp->codecHeader.VP9.flexible_mode = |
| info->codecSpecific.VP9.flexible_mode; |
| rtp->codecHeader.VP9.ss_data_available = |
| info->codecSpecific.VP9.ss_data_available; |
| rtp->codecHeader.VP9.picture_id = info->codecSpecific.VP9.picture_id; |
| rtp->codecHeader.VP9.tl0_pic_idx = info->codecSpecific.VP9.tl0_pic_idx; |
| rtp->codecHeader.VP9.temporal_idx = info->codecSpecific.VP9.temporal_idx; |
| rtp->codecHeader.VP9.spatial_idx = info->codecSpecific.VP9.spatial_idx; |
| rtp->codecHeader.VP9.temporal_up_switch = |
| info->codecSpecific.VP9.temporal_up_switch; |
| rtp->codecHeader.VP9.inter_layer_predicted = |
| info->codecSpecific.VP9.inter_layer_predicted; |
| rtp->codecHeader.VP9.gof_idx = info->codecSpecific.VP9.gof_idx; |
| rtp->codecHeader.VP9.num_spatial_layers = |
| info->codecSpecific.VP9.num_spatial_layers; |
| |
| if (info->codecSpecific.VP9.ss_data_available) { |
| rtp->codecHeader.VP9.spatial_layer_resolution_present = |
| info->codecSpecific.VP9.spatial_layer_resolution_present; |
| if (info->codecSpecific.VP9.spatial_layer_resolution_present) { |
| for (size_t i = 0; i < info->codecSpecific.VP9.num_spatial_layers; |
| ++i) { |
| rtp->codecHeader.VP9.width[i] = info->codecSpecific.VP9.width[i]; |
| rtp->codecHeader.VP9.height[i] = info->codecSpecific.VP9.height[i]; |
| } |
| } |
| rtp->codecHeader.VP9.gof.CopyGofInfoVP9(info->codecSpecific.VP9.gof); |
| } |
| |
| rtp->codecHeader.VP9.num_ref_pics = info->codecSpecific.VP9.num_ref_pics; |
| for (int i = 0; i < info->codecSpecific.VP9.num_ref_pics; ++i) |
| rtp->codecHeader.VP9.pid_diff[i] = info->codecSpecific.VP9.p_diff[i]; |
| return; |
| } |
| case kVideoCodecH264: |
| rtp->codec = kRtpVideoH264; |
| rtp->codecHeader.H264.packetization_mode = |
| info->codecSpecific.H264.packetization_mode; |
| return; |
| case kVideoCodecGeneric: |
| rtp->codec = kRtpVideoGeneric; |
| rtp->simulcastIdx = info->codecSpecific.generic.simulcast_idx; |
| return; |
| default: |
| return; |
| } |
| } |
| |
| } // namespace |
| |
| PayloadRouter::PayloadRouter(const std::vector<RtpRtcp*>& rtp_modules, |
| int payload_type) |
| : active_(false), |
| rtp_modules_(rtp_modules), |
| payload_type_(payload_type) { |
| } |
| |
| PayloadRouter::~PayloadRouter() {} |
| |
| void PayloadRouter::SetActive(bool active) { |
| rtc::CritScope lock(&crit_); |
| if (active_ == active) |
| return; |
| active_ = active; |
| |
| for (auto& module : rtp_modules_) { |
| module->SetSendingStatus(active_); |
| module->SetSendingMediaStatus(active_); |
| } |
| } |
| |
| bool PayloadRouter::IsActive() { |
| rtc::CritScope lock(&crit_); |
| return active_ && !rtp_modules_.empty(); |
| } |
| |
| EncodedImageCallback::Result PayloadRouter::OnEncodedImage( |
| const EncodedImage& encoded_image, |
| const CodecSpecificInfo* codec_specific_info, |
| const RTPFragmentationHeader* fragmentation) { |
| rtc::CritScope lock(&crit_); |
| RTC_DCHECK(!rtp_modules_.empty()); |
| if (!active_) |
| return Result(Result::ERROR_SEND_FAILED); |
| |
| RTPVideoHeader rtp_video_header; |
| memset(&rtp_video_header, 0, sizeof(RTPVideoHeader)); |
| if (codec_specific_info) |
| CopyCodecSpecific(codec_specific_info, &rtp_video_header); |
| rtp_video_header.rotation = encoded_image.rotation_; |
| rtp_video_header.content_type = encoded_image.content_type_; |
| if (encoded_image.timing_.flags != TimingFrameFlags::kInvalid) { |
| rtp_video_header.video_timing.encode_start_delta_ms = |
| VideoSendTiming::GetDeltaCappedMs( |
| encoded_image.capture_time_ms_, |
| encoded_image.timing_.encode_start_ms); |
| rtp_video_header.video_timing.encode_finish_delta_ms = |
| VideoSendTiming::GetDeltaCappedMs( |
| encoded_image.capture_time_ms_, |
| encoded_image.timing_.encode_finish_ms); |
| rtp_video_header.video_timing.packetization_finish_delta_ms = 0; |
| rtp_video_header.video_timing.pacer_exit_delta_ms = 0; |
| rtp_video_header.video_timing.network_timstamp_delta_ms = 0; |
| rtp_video_header.video_timing.network2_timstamp_delta_ms = 0; |
| } |
| rtp_video_header.video_timing.flags = encoded_image.timing_.flags; |
| rtp_video_header.playout_delay = encoded_image.playout_delay_; |
| |
| int stream_index = rtp_video_header.simulcastIdx; |
| RTC_DCHECK_LT(stream_index, rtp_modules_.size()); |
| uint32_t frame_id; |
| bool send_result = rtp_modules_[stream_index]->SendOutgoingData( |
| encoded_image._frameType, payload_type_, encoded_image._timeStamp, |
| encoded_image.capture_time_ms_, encoded_image._buffer, |
| encoded_image._length, fragmentation, &rtp_video_header, &frame_id); |
| if (!send_result) |
| return Result(Result::ERROR_SEND_FAILED); |
| |
| return Result(Result::OK, frame_id); |
| } |
| |
| void PayloadRouter::OnBitrateAllocationUpdated( |
| const BitrateAllocation& bitrate) { |
| rtc::CritScope lock(&crit_); |
| if (IsActive()) { |
| if (rtp_modules_.size() == 1) { |
| // If spatial scalability is enabled, it is covered by a single stream. |
| rtp_modules_[0]->SetVideoBitrateAllocation(bitrate); |
| } else { |
| // Simulcast is in use, split the BitrateAllocation into one struct per |
| // rtp stream, moving over the temporal layer allocation. |
| for (size_t si = 0; si < rtp_modules_.size(); ++si) { |
| // Don't send empty TargetBitrate messages on streams not being relayed. |
| if (bitrate.GetSpatialLayerSum(si) == 0) |
| break; |
| |
| BitrateAllocation layer_bitrate; |
| for (int tl = 0; tl < kMaxTemporalStreams; ++tl) |
| layer_bitrate.SetBitrate(0, tl, bitrate.GetBitrate(si, tl)); |
| rtp_modules_[si]->SetVideoBitrateAllocation(layer_bitrate); |
| } |
| } |
| } |
| } |
| |
| } // namespace webrtc |