| /* |
| * Copyright (c) 2014 The WebRTC project authors. All Rights Reserved. |
| * |
| * Use of this source code is governed by a BSD-style license |
| * that can be found in the LICENSE file in the root of the source |
| * tree. An additional intellectual property rights grant can be found |
| * in the file PATENTS. All contributing project authors may |
| * be found in the AUTHORS file in the root of the source tree. |
| */ |
| #include "webrtc/video/video_capture_input.h" |
| |
| #include <vector> |
| |
| #include "testing/gmock/include/gmock/gmock.h" |
| #include "testing/gtest/include/gtest/gtest.h" |
| #include "webrtc/base/scoped_ptr.h" |
| #include "webrtc/common.h" |
| #include "webrtc/modules/utility/include/mock/mock_process_thread.h" |
| #include "webrtc/system_wrappers/include/critical_section_wrapper.h" |
| #include "webrtc/system_wrappers/include/event_wrapper.h" |
| #include "webrtc/system_wrappers/include/ref_count.h" |
| #include "webrtc/system_wrappers/include/scoped_vector.h" |
| #include "webrtc/test/fake_texture_frame.h" |
| #include "webrtc/video/send_statistics_proxy.h" |
| |
| using ::testing::_; |
| using ::testing::Invoke; |
| using ::testing::NiceMock; |
| using ::testing::Return; |
| using ::testing::WithArg; |
| |
| // If an output frame does not arrive in 500ms, the test will fail. |
| #define FRAME_TIMEOUT_MS 500 |
| |
| namespace webrtc { |
| |
| class MockVideoCaptureCallback : public VideoCaptureCallback { |
| public: |
| MOCK_METHOD1(DeliverFrame, void(VideoFrame video_frame)); |
| }; |
| |
| bool EqualFrames(const VideoFrame& frame1, const VideoFrame& frame2); |
| bool EqualTextureFrames(const VideoFrame& frame1, const VideoFrame& frame2); |
| bool EqualBufferFrames(const VideoFrame& frame1, const VideoFrame& frame2); |
| bool EqualFramesVector(const ScopedVector<VideoFrame>& frames1, |
| const ScopedVector<VideoFrame>& frames2); |
| VideoFrame* CreateVideoFrame(uint8_t length); |
| |
| class VideoCaptureInputTest : public ::testing::Test { |
| protected: |
| VideoCaptureInputTest() |
| : mock_process_thread_(new NiceMock<MockProcessThread>), |
| mock_frame_callback_(new NiceMock<MockVideoCaptureCallback>), |
| output_frame_event_(EventWrapper::Create()), |
| stats_proxy_(Clock::GetRealTimeClock(), |
| webrtc::VideoSendStream::Config(nullptr)) {} |
| |
| virtual void SetUp() { |
| EXPECT_CALL(*mock_frame_callback_, DeliverFrame(_)) |
| .WillRepeatedly( |
| WithArg<0>(Invoke(this, &VideoCaptureInputTest::AddOutputFrame))); |
| |
| Config config; |
| input_.reset(new internal::VideoCaptureInput( |
| mock_process_thread_.get(), mock_frame_callback_.get(), nullptr, |
| &stats_proxy_, nullptr, nullptr)); |
| } |
| |
| virtual void TearDown() { |
| // VideoCaptureInput accesses |mock_process_thread_| in destructor and |
| // should |
| // be deleted first. |
| input_.reset(); |
| } |
| |
| void AddInputFrame(VideoFrame* frame) { |
| input_->IncomingCapturedFrame(*frame); |
| } |
| |
| void AddOutputFrame(const VideoFrame& frame) { |
| if (frame.native_handle() == NULL) |
| output_frame_ybuffers_.push_back(frame.buffer(kYPlane)); |
| output_frames_.push_back(new VideoFrame(frame)); |
| output_frame_event_->Set(); |
| } |
| |
| void WaitOutputFrame() { |
| EXPECT_EQ(kEventSignaled, output_frame_event_->Wait(FRAME_TIMEOUT_MS)); |
| } |
| |
| rtc::scoped_ptr<MockProcessThread> mock_process_thread_; |
| rtc::scoped_ptr<MockVideoCaptureCallback> mock_frame_callback_; |
| |
| // Used to send input capture frames to VideoCaptureInput. |
| rtc::scoped_ptr<internal::VideoCaptureInput> input_; |
| |
| // Input capture frames of VideoCaptureInput. |
| ScopedVector<VideoFrame> input_frames_; |
| |
| // Indicate an output frame has arrived. |
| rtc::scoped_ptr<EventWrapper> output_frame_event_; |
| |
| // Output delivered frames of VideoCaptureInput. |
| ScopedVector<VideoFrame> output_frames_; |
| |
| // The pointers of Y plane buffers of output frames. This is used to verify |
| // the frame are swapped and not copied. |
| std::vector<const uint8_t*> output_frame_ybuffers_; |
| SendStatisticsProxy stats_proxy_; |
| }; |
| |
| TEST_F(VideoCaptureInputTest, DoesNotRetainHandleNorCopyBuffer) { |
| // Indicate an output frame has arrived. |
| rtc::scoped_ptr<EventWrapper> frame_destroyed_event(EventWrapper::Create()); |
| class TestBuffer : public webrtc::I420Buffer { |
| public: |
| explicit TestBuffer(EventWrapper* event) |
| : I420Buffer(5, 5), event_(event) {} |
| |
| private: |
| friend class rtc::RefCountedObject<TestBuffer>; |
| ~TestBuffer() override { event_->Set(); } |
| EventWrapper* event_; |
| }; |
| |
| VideoFrame frame( |
| new rtc::RefCountedObject<TestBuffer>(frame_destroyed_event.get()), 1, 1, |
| kVideoRotation_0); |
| |
| AddInputFrame(&frame); |
| WaitOutputFrame(); |
| |
| EXPECT_EQ(output_frames_[0]->video_frame_buffer().get(), |
| frame.video_frame_buffer().get()); |
| output_frames_.clear(); |
| frame.Reset(); |
| EXPECT_EQ(kEventSignaled, frame_destroyed_event->Wait(FRAME_TIMEOUT_MS)); |
| } |
| |
| TEST_F(VideoCaptureInputTest, TestNtpTimeStampSetIfRenderTimeSet) { |
| input_frames_.push_back(CreateVideoFrame(0)); |
| input_frames_[0]->set_render_time_ms(5); |
| input_frames_[0]->set_ntp_time_ms(0); |
| |
| AddInputFrame(input_frames_[0]); |
| WaitOutputFrame(); |
| EXPECT_GT(output_frames_[0]->ntp_time_ms(), |
| input_frames_[0]->render_time_ms()); |
| } |
| |
| TEST_F(VideoCaptureInputTest, TestRtpTimeStampSet) { |
| input_frames_.push_back(CreateVideoFrame(0)); |
| input_frames_[0]->set_render_time_ms(0); |
| input_frames_[0]->set_ntp_time_ms(1); |
| input_frames_[0]->set_timestamp(0); |
| |
| AddInputFrame(input_frames_[0]); |
| WaitOutputFrame(); |
| EXPECT_EQ(output_frames_[0]->timestamp(), |
| input_frames_[0]->ntp_time_ms() * 90); |
| } |
| |
| TEST_F(VideoCaptureInputTest, DropsFramesWithSameOrOldNtpTimestamp) { |
| input_frames_.push_back(CreateVideoFrame(0)); |
| |
| input_frames_[0]->set_ntp_time_ms(17); |
| AddInputFrame(input_frames_[0]); |
| WaitOutputFrame(); |
| EXPECT_EQ(output_frames_[0]->timestamp(), |
| input_frames_[0]->ntp_time_ms() * 90); |
| |
| // Repeat frame with the same NTP timestamp should drop. |
| AddInputFrame(input_frames_[0]); |
| EXPECT_EQ(kEventTimeout, output_frame_event_->Wait(FRAME_TIMEOUT_MS)); |
| |
| // As should frames with a decreased NTP timestamp. |
| input_frames_[0]->set_ntp_time_ms(input_frames_[0]->ntp_time_ms() - 1); |
| AddInputFrame(input_frames_[0]); |
| EXPECT_EQ(kEventTimeout, output_frame_event_->Wait(FRAME_TIMEOUT_MS)); |
| |
| // But delivering with an increased NTP timestamp should succeed. |
| input_frames_[0]->set_ntp_time_ms(4711); |
| AddInputFrame(input_frames_[0]); |
| WaitOutputFrame(); |
| EXPECT_EQ(output_frames_[1]->timestamp(), |
| input_frames_[0]->ntp_time_ms() * 90); |
| } |
| |
| TEST_F(VideoCaptureInputTest, TestTextureFrames) { |
| const int kNumFrame = 3; |
| for (int i = 0 ; i < kNumFrame; ++i) { |
| test::FakeNativeHandle* dummy_handle = new test::FakeNativeHandle(); |
| // Add one to |i| so that width/height > 0. |
| input_frames_.push_back(new VideoFrame(test::CreateFakeNativeHandleFrame( |
| dummy_handle, i + 1, i + 1, i + 1, i + 1, webrtc::kVideoRotation_0))); |
| AddInputFrame(input_frames_[i]); |
| WaitOutputFrame(); |
| EXPECT_EQ(dummy_handle, output_frames_[i]->native_handle()); |
| } |
| |
| EXPECT_TRUE(EqualFramesVector(input_frames_, output_frames_)); |
| } |
| |
| TEST_F(VideoCaptureInputTest, TestI420Frames) { |
| const int kNumFrame = 4; |
| std::vector<const uint8_t*> ybuffer_pointers; |
| for (int i = 0; i < kNumFrame; ++i) { |
| input_frames_.push_back(CreateVideoFrame(static_cast<uint8_t>(i + 1))); |
| const VideoFrame* const_input_frame = input_frames_[i]; |
| ybuffer_pointers.push_back(const_input_frame->buffer(kYPlane)); |
| AddInputFrame(input_frames_[i]); |
| WaitOutputFrame(); |
| } |
| |
| EXPECT_TRUE(EqualFramesVector(input_frames_, output_frames_)); |
| // Make sure the buffer is not copied. |
| for (int i = 0; i < kNumFrame; ++i) |
| EXPECT_EQ(ybuffer_pointers[i], output_frame_ybuffers_[i]); |
| } |
| |
| TEST_F(VideoCaptureInputTest, TestI420FrameAfterTextureFrame) { |
| test::FakeNativeHandle* dummy_handle = new test::FakeNativeHandle(); |
| input_frames_.push_back(new VideoFrame(test::CreateFakeNativeHandleFrame( |
| dummy_handle, 1, 1, 1, 1, webrtc::kVideoRotation_0))); |
| AddInputFrame(input_frames_[0]); |
| WaitOutputFrame(); |
| EXPECT_EQ(dummy_handle, output_frames_[0]->native_handle()); |
| |
| input_frames_.push_back(CreateVideoFrame(2)); |
| AddInputFrame(input_frames_[1]); |
| WaitOutputFrame(); |
| |
| EXPECT_TRUE(EqualFramesVector(input_frames_, output_frames_)); |
| } |
| |
| TEST_F(VideoCaptureInputTest, TestTextureFrameAfterI420Frame) { |
| input_frames_.push_back(CreateVideoFrame(1)); |
| AddInputFrame(input_frames_[0]); |
| WaitOutputFrame(); |
| |
| test::FakeNativeHandle* dummy_handle = new test::FakeNativeHandle(); |
| input_frames_.push_back(new VideoFrame(test::CreateFakeNativeHandleFrame( |
| dummy_handle, 1, 1, 2, 2, webrtc::kVideoRotation_0))); |
| AddInputFrame(input_frames_[1]); |
| WaitOutputFrame(); |
| |
| EXPECT_TRUE(EqualFramesVector(input_frames_, output_frames_)); |
| } |
| |
| bool EqualFrames(const VideoFrame& frame1, const VideoFrame& frame2) { |
| if (frame1.native_handle() != NULL || frame2.native_handle() != NULL) |
| return EqualTextureFrames(frame1, frame2); |
| return EqualBufferFrames(frame1, frame2); |
| } |
| |
| bool EqualTextureFrames(const VideoFrame& frame1, const VideoFrame& frame2) { |
| return ((frame1.native_handle() == frame2.native_handle()) && |
| (frame1.width() == frame2.width()) && |
| (frame1.height() == frame2.height()) && |
| (frame1.render_time_ms() == frame2.render_time_ms())); |
| } |
| |
| bool EqualBufferFrames(const VideoFrame& frame1, const VideoFrame& frame2) { |
| return ((frame1.width() == frame2.width()) && |
| (frame1.height() == frame2.height()) && |
| (frame1.stride(kYPlane) == frame2.stride(kYPlane)) && |
| (frame1.stride(kUPlane) == frame2.stride(kUPlane)) && |
| (frame1.stride(kVPlane) == frame2.stride(kVPlane)) && |
| (frame1.render_time_ms() == frame2.render_time_ms()) && |
| (frame1.allocated_size(kYPlane) == frame2.allocated_size(kYPlane)) && |
| (frame1.allocated_size(kUPlane) == frame2.allocated_size(kUPlane)) && |
| (frame1.allocated_size(kVPlane) == frame2.allocated_size(kVPlane)) && |
| (memcmp(frame1.buffer(kYPlane), frame2.buffer(kYPlane), |
| frame1.allocated_size(kYPlane)) == 0) && |
| (memcmp(frame1.buffer(kUPlane), frame2.buffer(kUPlane), |
| frame1.allocated_size(kUPlane)) == 0) && |
| (memcmp(frame1.buffer(kVPlane), frame2.buffer(kVPlane), |
| frame1.allocated_size(kVPlane)) == 0)); |
| } |
| |
| bool EqualFramesVector(const ScopedVector<VideoFrame>& frames1, |
| const ScopedVector<VideoFrame>& frames2) { |
| if (frames1.size() != frames2.size()) |
| return false; |
| for (size_t i = 0; i < frames1.size(); ++i) { |
| if (!EqualFrames(*frames1[i], *frames2[i])) |
| return false; |
| } |
| return true; |
| } |
| |
| VideoFrame* CreateVideoFrame(uint8_t data) { |
| VideoFrame* frame = new VideoFrame(); |
| const int width = 36; |
| const int height = 24; |
| const int kSizeY = width * height * 2; |
| uint8_t buffer[kSizeY]; |
| memset(buffer, data, kSizeY); |
| frame->CreateFrame(buffer, buffer, buffer, width, height, width, width / 2, |
| width / 2); |
| frame->set_render_time_ms(data); |
| return frame; |
| } |
| |
| } // namespace webrtc |