blob: 2e22d1a2db07a695557aa46f89c50460f6d26cf7 [file] [log] [blame]
/*
* Copyright 2013 The WebRTC project authors. All Rights Reserved.
*
* Use of this source code is governed by a BSD-style license
* that can be found in the LICENSE file in the root of the source
* tree. An additional intellectual property rights grant can be found
* in the file PATENTS. All contributing project authors may
* be found in the AUTHORS file in the root of the source tree.
*/
package org.webrtc;
import androidx.annotation.Nullable;
/**
* Java wrapper of native AndroidVideoTrackSource.
*/
public class VideoSource extends MediaSource {
/** Simple aspect ratio clas for use in constraining output format. */
public static class AspectRatio {
public static final AspectRatio UNDEFINED = new AspectRatio(/* width= */ 0, /* height= */ 0);
public final int width;
public final int height;
public AspectRatio(int width, int height) {
this.width = width;
this.height = height;
}
}
private final NativeAndroidVideoTrackSource nativeAndroidVideoTrackSource;
private final Object videoProcessorLock = new Object();
@Nullable private VideoProcessor videoProcessor;
private boolean isCapturerRunning;
private final CapturerObserver capturerObserver = new CapturerObserver() {
@Override
public void onCapturerStarted(boolean success) {
nativeAndroidVideoTrackSource.setState(success);
synchronized (videoProcessorLock) {
isCapturerRunning = success;
if (videoProcessor != null) {
videoProcessor.onCapturerStarted(success);
}
}
}
@Override
public void onCapturerStopped() {
nativeAndroidVideoTrackSource.setState(/* isLive= */ false);
synchronized (videoProcessorLock) {
isCapturerRunning = false;
if (videoProcessor != null) {
videoProcessor.onCapturerStopped();
}
}
}
@Override
public void onFrameCaptured(VideoFrame frame) {
final VideoProcessor.FrameAdaptationParameters parameters =
nativeAndroidVideoTrackSource.adaptFrame(frame);
synchronized (videoProcessorLock) {
if (videoProcessor != null) {
videoProcessor.onFrameCaptured(frame, parameters);
return;
}
}
VideoFrame adaptedFrame = VideoProcessor.applyFrameAdaptationParameters(frame, parameters);
if (adaptedFrame != null) {
nativeAndroidVideoTrackSource.onFrameCaptured(adaptedFrame);
adaptedFrame.release();
}
}
};
public VideoSource(long nativeSource) {
super(nativeSource);
this.nativeAndroidVideoTrackSource = new NativeAndroidVideoTrackSource(nativeSource);
}
/**
* Calling this function will cause frames to be scaled down to the requested resolution. Also,
* frames will be cropped to match the requested aspect ratio, and frames will be dropped to match
* the requested fps. The requested aspect ratio is orientation agnostic and will be adjusted to
* maintain the input orientation, so it doesn't matter if e.g. 1280x720 or 720x1280 is requested.
*/
public void adaptOutputFormat(int width, int height, int fps) {
final int maxSide = Math.max(width, height);
final int minSide = Math.min(width, height);
adaptOutputFormat(maxSide, minSide, minSide, maxSide, fps);
}
/**
* Same as above, but allows setting two different target resolutions depending on incoming
* frame orientation. This gives more fine-grained control and can e.g. be used to force landscape
* video to be cropped to portrait video.
*/
public void adaptOutputFormat(
int landscapeWidth, int landscapeHeight, int portraitWidth, int portraitHeight, int fps) {
adaptOutputFormat(new AspectRatio(landscapeWidth, landscapeHeight),
/* maxLandscapePixelCount= */ landscapeWidth * landscapeHeight,
new AspectRatio(portraitWidth, portraitHeight),
/* maxPortraitPixelCount= */ portraitWidth * portraitHeight, fps);
}
/** Same as above, with even more control as each constraint is optional. */
public void adaptOutputFormat(AspectRatio targetLandscapeAspectRatio,
@Nullable Integer maxLandscapePixelCount, AspectRatio targetPortraitAspectRatio,
@Nullable Integer maxPortraitPixelCount, @Nullable Integer maxFps) {
nativeAndroidVideoTrackSource.adaptOutputFormat(targetLandscapeAspectRatio,
maxLandscapePixelCount, targetPortraitAspectRatio, maxPortraitPixelCount, maxFps);
}
public void setIsScreencast(boolean isScreencast) {
nativeAndroidVideoTrackSource.setIsScreencast(isScreencast);
}
/**
* Hook for injecting a custom video processor before frames are passed onto WebRTC. The frames
* will be cropped and scaled depending on CPU and network conditions before they are passed to
* the video processor. Frames will be delivered to the video processor on the same thread they
* are passed to this object. The video processor is allowed to deliver the processed frames
* back on any thread.
*/
public void setVideoProcessor(@Nullable VideoProcessor newVideoProcessor) {
synchronized (videoProcessorLock) {
if (videoProcessor != null) {
videoProcessor.setSink(/* sink= */ null);
if (isCapturerRunning) {
videoProcessor.onCapturerStopped();
}
}
videoProcessor = newVideoProcessor;
if (newVideoProcessor != null) {
newVideoProcessor.setSink(
(frame)
-> runWithReference(() -> nativeAndroidVideoTrackSource.onFrameCaptured(frame)));
if (isCapturerRunning) {
newVideoProcessor.onCapturerStarted(/* success= */ true);
}
}
}
}
public CapturerObserver getCapturerObserver() {
return capturerObserver;
}
/** Returns a pointer to webrtc::VideoTrackSourceInterface. */
long getNativeVideoTrackSource() {
return getNativeMediaSource();
}
@Override
public void dispose() {
setVideoProcessor(/* newVideoProcessor= */ null);
super.dispose();
}
}