blob: 6ebc1ae1415204b0f6dac022aa976e04aced925a [file] [log] [blame]
/*
* Copyright (c) 2011 The WebRTC project authors. All Rights Reserved.
*
* Use of this source code is governed by a BSD-style license
* that can be found in the LICENSE file in the root of the source
* tree. An additional intellectual property rights grant can be found
* in the file PATENTS. All contributing project authors may
* be found in the AUTHORS file in the root of the source tree.
*/
#ifndef WEBRTC_MEDIA_BASE_FAKEVIDEORENDERER_H_
#define WEBRTC_MEDIA_BASE_FAKEVIDEORENDERER_H_
#include "webrtc/base/logging.h"
#include "webrtc/base/sigslot.h"
#include "webrtc/media/base/videoframe.h"
#include "webrtc/media/base/videosinkinterface.h"
namespace cricket {
// Faked video renderer that has a callback for actions on rendering.
class FakeVideoRenderer : public rtc::VideoSinkInterface<cricket::VideoFrame> {
public:
FakeVideoRenderer()
: errors_(0),
width_(0),
height_(0),
rotation_(webrtc::kVideoRotation_0),
timestamp_us_(0),
num_rendered_frames_(0),
black_frame_(false) {}
virtual void OnFrame(const VideoFrame& frame) {
rtc::CritScope cs(&crit_);
// TODO(zhurunz) Check with VP8 team to see if we can remove this
// tolerance on Y values. Some unit tests produce Y values close
// to 16 rather than close to zero, for supposedly black frames.
// Largest value observed is 34, e.g., running
// P2PTestConductor.LocalP2PTest16To9 (peerconnection_unittests).
black_frame_ = CheckFrameColorYuv(0, 48, 128, 128, 128, 128, &frame);
// Treat unexpected frame size as error.
++num_rendered_frames_;
width_ = frame.width();
height_ = frame.height();
rotation_ = frame.rotation();
timestamp_us_ = frame.timestamp_us();
SignalRenderFrame(&frame);
}
int errors() const { return errors_; }
int width() const {
rtc::CritScope cs(&crit_);
return width_;
}
int height() const {
rtc::CritScope cs(&crit_);
return height_;
}
webrtc::VideoRotation rotation() const {
rtc::CritScope cs(&crit_);
return rotation_;
}
int64_t timestamp_us() const {
rtc::CritScope cs(&crit_);
return timestamp_us_;
}
int num_rendered_frames() const {
rtc::CritScope cs(&crit_);
return num_rendered_frames_;
}
bool black_frame() const {
rtc::CritScope cs(&crit_);
return black_frame_;
}
sigslot::signal3<int, int, int> SignalSetSize;
sigslot::signal1<const VideoFrame*> SignalRenderFrame;
private:
static bool CheckFrameColorYuv(uint8_t y_min,
uint8_t y_max,
uint8_t u_min,
uint8_t u_max,
uint8_t v_min,
uint8_t v_max,
const cricket::VideoFrame* frame) {
if (!frame || !frame->video_frame_buffer()) {
return false;
}
// Y
int y_width = frame->width();
int y_height = frame->height();
const uint8_t* y_plane = frame->video_frame_buffer()->DataY();
const uint8_t* y_pos = y_plane;
int32_t y_pitch = frame->video_frame_buffer()->StrideY();
for (int i = 0; i < y_height; ++i) {
for (int j = 0; j < y_width; ++j) {
uint8_t y_value = *(y_pos + j);
if (y_value < y_min || y_value > y_max) {
return false;
}
}
y_pos += y_pitch;
}
// U and V
int chroma_width = (frame->width() + 1)/2;
int chroma_height = (frame->height() + 1)/2;
const uint8_t* u_plane = frame->video_frame_buffer()->DataU();
const uint8_t* v_plane = frame->video_frame_buffer()->DataV();
const uint8_t* u_pos = u_plane;
const uint8_t* v_pos = v_plane;
int32_t u_pitch = frame->video_frame_buffer()->StrideU();
int32_t v_pitch = frame->video_frame_buffer()->StrideV();
for (int i = 0; i < chroma_height; ++i) {
for (int j = 0; j < chroma_width; ++j) {
uint8_t u_value = *(u_pos + j);
if (u_value < u_min || u_value > u_max) {
return false;
}
uint8_t v_value = *(v_pos + j);
if (v_value < v_min || v_value > v_max) {
return false;
}
}
u_pos += u_pitch;
v_pos += v_pitch;
}
return true;
}
int errors_;
int width_;
int height_;
webrtc::VideoRotation rotation_;
int64_t timestamp_us_;
int num_rendered_frames_;
bool black_frame_;
rtc::CriticalSection crit_;
};
} // namespace cricket
#endif // WEBRTC_MEDIA_BASE_FAKEVIDEORENDERER_H_