| /* |
| * Copyright 2014 The WebRTC Project Authors. All rights reserved. |
| * |
| * Use of this source code is governed by a BSD-style license |
| * that can be found in the LICENSE file in the root of the source |
| * tree. An additional intellectual property rights grant can be found |
| * in the file PATENTS. All contributing project authors may |
| * be found in the AUTHORS file in the root of the source tree. |
| */ |
| |
| package org.appspot.apprtc; |
| |
| import android.content.Context; |
| import android.os.Environment; |
| import android.os.ParcelFileDescriptor; |
| import android.preference.PreferenceManager; |
| import android.util.Log; |
| import java.io.File; |
| import java.io.IOException; |
| import java.nio.ByteBuffer; |
| import java.nio.charset.Charset; |
| import java.text.DateFormat; |
| import java.text.SimpleDateFormat; |
| import java.util.ArrayList; |
| import java.util.Arrays; |
| import java.util.Collections; |
| import java.util.Date; |
| import java.util.Iterator; |
| import java.util.List; |
| import java.util.Locale; |
| import java.util.Timer; |
| import java.util.TimerTask; |
| import java.util.concurrent.ExecutorService; |
| import java.util.concurrent.Executors; |
| import java.util.regex.Matcher; |
| import java.util.regex.Pattern; |
| import org.appspot.apprtc.AppRTCClient.SignalingParameters; |
| import org.appspot.apprtc.RecordedAudioToFileController; |
| import org.webrtc.AudioSource; |
| import org.webrtc.AudioTrack; |
| import org.webrtc.CameraVideoCapturer; |
| import org.webrtc.DataChannel; |
| import org.webrtc.DefaultVideoDecoderFactory; |
| import org.webrtc.DefaultVideoEncoderFactory; |
| import org.webrtc.EglBase; |
| import org.webrtc.IceCandidate; |
| import org.webrtc.Logging; |
| import org.webrtc.MediaConstraints; |
| import org.webrtc.MediaStream; |
| import org.webrtc.MediaStreamTrack; |
| import org.webrtc.MediaStreamTrack.MediaType; |
| import org.webrtc.PeerConnection; |
| import org.webrtc.PeerConnection.IceConnectionState; |
| import org.webrtc.PeerConnectionFactory; |
| import org.webrtc.RtpParameters; |
| import org.webrtc.RtpReceiver; |
| import org.webrtc.RtpSender; |
| import org.webrtc.RtpTransceiver; |
| import org.webrtc.SdpObserver; |
| import org.webrtc.SessionDescription; |
| import org.webrtc.SoftwareVideoDecoderFactory; |
| import org.webrtc.SoftwareVideoEncoderFactory; |
| import org.webrtc.StatsObserver; |
| import org.webrtc.StatsReport; |
| import org.webrtc.VideoCapturer; |
| import org.webrtc.VideoDecoderFactory; |
| import org.webrtc.VideoEncoderFactory; |
| import org.webrtc.VideoRenderer; |
| import org.webrtc.VideoSink; |
| import org.webrtc.VideoSource; |
| import org.webrtc.VideoTrack; |
| import org.webrtc.voiceengine.WebRtcAudioManager; |
| import org.webrtc.voiceengine.WebRtcAudioRecord; |
| import org.webrtc.voiceengine.WebRtcAudioRecord.AudioRecordStartErrorCode; |
| import org.webrtc.voiceengine.WebRtcAudioRecord.WebRtcAudioRecordErrorCallback; |
| import org.webrtc.voiceengine.WebRtcAudioTrack; |
| import org.webrtc.voiceengine.WebRtcAudioTrack.AudioTrackStartErrorCode; |
| import org.webrtc.voiceengine.WebRtcAudioUtils; |
| |
| /** |
| * Peer connection client implementation. |
| * |
| * <p>All public methods are routed to local looper thread. |
| * All PeerConnectionEvents callbacks are invoked from the same looper thread. |
| * This class is a singleton. |
| */ |
| public class PeerConnectionClient { |
| public static final String VIDEO_TRACK_ID = "ARDAMSv0"; |
| public static final String AUDIO_TRACK_ID = "ARDAMSa0"; |
| public static final String VIDEO_TRACK_TYPE = "video"; |
| private static final String TAG = "PCRTCClient"; |
| private static final String VIDEO_CODEC_VP8 = "VP8"; |
| private static final String VIDEO_CODEC_VP9 = "VP9"; |
| private static final String VIDEO_CODEC_H264 = "H264"; |
| private static final String VIDEO_CODEC_H264_BASELINE = "H264 Baseline"; |
| private static final String VIDEO_CODEC_H264_HIGH = "H264 High"; |
| private static final String AUDIO_CODEC_OPUS = "opus"; |
| private static final String AUDIO_CODEC_ISAC = "ISAC"; |
| private static final String VIDEO_CODEC_PARAM_START_BITRATE = "x-google-start-bitrate"; |
| private static final String VIDEO_FLEXFEC_FIELDTRIAL = |
| "WebRTC-FlexFEC-03-Advertised/Enabled/WebRTC-FlexFEC-03/Enabled/"; |
| private static final String VIDEO_VP8_INTEL_HW_ENCODER_FIELDTRIAL = "WebRTC-IntelVP8/Enabled/"; |
| private static final String VIDEO_H264_HIGH_PROFILE_FIELDTRIAL = |
| "WebRTC-H264HighProfile/Enabled/"; |
| private static final String DISABLE_WEBRTC_AGC_FIELDTRIAL = |
| "WebRTC-Audio-MinimizeResamplingOnMobile/Enabled/"; |
| private static final String AUDIO_CODEC_PARAM_BITRATE = "maxaveragebitrate"; |
| private static final String AUDIO_ECHO_CANCELLATION_CONSTRAINT = "googEchoCancellation"; |
| private static final String AUDIO_AUTO_GAIN_CONTROL_CONSTRAINT = "googAutoGainControl"; |
| private static final String AUDIO_HIGH_PASS_FILTER_CONSTRAINT = "googHighpassFilter"; |
| private static final String AUDIO_NOISE_SUPPRESSION_CONSTRAINT = "googNoiseSuppression"; |
| private static final String DTLS_SRTP_KEY_AGREEMENT_CONSTRAINT = "DtlsSrtpKeyAgreement"; |
| private static final int HD_VIDEO_WIDTH = 1280; |
| private static final int HD_VIDEO_HEIGHT = 720; |
| private static final int BPS_IN_KBPS = 1000; |
| private static final String RTCEVENTLOG_OUTPUT_DIR_NAME = "rtc_event_log"; |
| |
| // Executor thread is started once in private ctor and is used for all |
| // peer connection API calls to ensure new peer connection factory is |
| // created on the same thread as previously destroyed factory. |
| private static final ExecutorService executor = Executors.newSingleThreadExecutor(); |
| |
| private final PCObserver pcObserver = new PCObserver(); |
| private final SDPObserver sdpObserver = new SDPObserver(); |
| |
| private final EglBase rootEglBase; |
| private final Context appContext; |
| private PeerConnectionFactory factory; |
| private PeerConnection peerConnection; |
| PeerConnectionFactory.Options options = null; |
| private AudioSource audioSource; |
| private VideoSource videoSource; |
| private boolean videoCallEnabled; |
| private boolean preferIsac; |
| private String preferredVideoCodec; |
| private boolean videoCapturerStopped; |
| private boolean isError; |
| private Timer statsTimer; |
| private VideoSink localRender; |
| private List<VideoRenderer.Callbacks> remoteRenders; |
| private SignalingParameters signalingParameters; |
| private int videoWidth; |
| private int videoHeight; |
| private int videoFps; |
| private MediaConstraints audioConstraints; |
| private MediaConstraints sdpMediaConstraints; |
| private PeerConnectionParameters peerConnectionParameters; |
| // Queued remote ICE candidates are consumed only after both local and |
| // remote descriptions are set. Similarly local ICE candidates are sent to |
| // remote peer after both local and remote description are set. |
| private List<IceCandidate> queuedRemoteCandidates; |
| private PeerConnectionEvents events; |
| private boolean isInitiator; |
| private SessionDescription localSdp; // either offer or answer SDP |
| private MediaStream mediaStream; |
| private VideoCapturer videoCapturer; |
| // enableVideo is set to true if video should be rendered and sent. |
| private boolean renderVideo; |
| private VideoTrack localVideoTrack; |
| private VideoTrack remoteVideoTrack; |
| private RtpSender localVideoSender; |
| // enableAudio is set to true if audio should be sent. |
| private boolean enableAudio; |
| private AudioTrack localAudioTrack; |
| private DataChannel dataChannel; |
| private boolean dataChannelEnabled; |
| // Enable RtcEventLog. |
| private RtcEventLog rtcEventLog; |
| // Implements the WebRtcAudioRecordSamplesReadyCallback interface and writes |
| // recorded audio samples to an output file. |
| private RecordedAudioToFileController saveRecordedAudioToFile = null; |
| |
| /** |
| * Peer connection parameters. |
| */ |
| public static class DataChannelParameters { |
| public final boolean ordered; |
| public final int maxRetransmitTimeMs; |
| public final int maxRetransmits; |
| public final String protocol; |
| public final boolean negotiated; |
| public final int id; |
| |
| public DataChannelParameters(boolean ordered, int maxRetransmitTimeMs, int maxRetransmits, |
| String protocol, boolean negotiated, int id) { |
| this.ordered = ordered; |
| this.maxRetransmitTimeMs = maxRetransmitTimeMs; |
| this.maxRetransmits = maxRetransmits; |
| this.protocol = protocol; |
| this.negotiated = negotiated; |
| this.id = id; |
| } |
| } |
| |
| /** |
| * Peer connection parameters. |
| */ |
| public static class PeerConnectionParameters { |
| public final boolean videoCallEnabled; |
| public final boolean loopback; |
| public final boolean tracing; |
| public final int videoWidth; |
| public final int videoHeight; |
| public final int videoFps; |
| public final int videoMaxBitrate; |
| public final String videoCodec; |
| public final boolean videoCodecHwAcceleration; |
| public final boolean videoFlexfecEnabled; |
| public final int audioStartBitrate; |
| public final String audioCodec; |
| public final boolean noAudioProcessing; |
| public final boolean aecDump; |
| public final boolean saveInputAudioToFile; |
| public final boolean useOpenSLES; |
| public final boolean disableBuiltInAEC; |
| public final boolean disableBuiltInAGC; |
| public final boolean disableBuiltInNS; |
| public final boolean disableWebRtcAGCAndHPF; |
| public final boolean enableRtcEventLog; |
| private final DataChannelParameters dataChannelParameters; |
| |
| public PeerConnectionParameters(boolean videoCallEnabled, boolean loopback, boolean tracing, |
| int videoWidth, int videoHeight, int videoFps, int videoMaxBitrate, String videoCodec, |
| boolean videoCodecHwAcceleration, boolean videoFlexfecEnabled, int audioStartBitrate, |
| String audioCodec, boolean noAudioProcessing, boolean aecDump, boolean saveInputAudioToFile, |
| boolean useOpenSLES, boolean disableBuiltInAEC, boolean disableBuiltInAGC, |
| boolean disableBuiltInNS, boolean disableWebRtcAGCAndHPF, boolean enableRtcEventLog, |
| DataChannelParameters dataChannelParameters) { |
| this.videoCallEnabled = videoCallEnabled; |
| this.loopback = loopback; |
| this.tracing = tracing; |
| this.videoWidth = videoWidth; |
| this.videoHeight = videoHeight; |
| this.videoFps = videoFps; |
| this.videoMaxBitrate = videoMaxBitrate; |
| this.videoCodec = videoCodec; |
| this.videoFlexfecEnabled = videoFlexfecEnabled; |
| this.videoCodecHwAcceleration = videoCodecHwAcceleration; |
| this.audioStartBitrate = audioStartBitrate; |
| this.audioCodec = audioCodec; |
| this.noAudioProcessing = noAudioProcessing; |
| this.aecDump = aecDump; |
| this.saveInputAudioToFile = saveInputAudioToFile; |
| this.useOpenSLES = useOpenSLES; |
| this.disableBuiltInAEC = disableBuiltInAEC; |
| this.disableBuiltInAGC = disableBuiltInAGC; |
| this.disableBuiltInNS = disableBuiltInNS; |
| this.disableWebRtcAGCAndHPF = disableWebRtcAGCAndHPF; |
| this.enableRtcEventLog = enableRtcEventLog; |
| this.dataChannelParameters = dataChannelParameters; |
| } |
| } |
| |
| /** |
| * Peer connection events. |
| */ |
| public interface PeerConnectionEvents { |
| /** |
| * Callback fired once local SDP is created and set. |
| */ |
| void onLocalDescription(final SessionDescription sdp); |
| |
| /** |
| * Callback fired once local Ice candidate is generated. |
| */ |
| void onIceCandidate(final IceCandidate candidate); |
| |
| /** |
| * Callback fired once local ICE candidates are removed. |
| */ |
| void onIceCandidatesRemoved(final IceCandidate[] candidates); |
| |
| /** |
| * Callback fired once connection is established (IceConnectionState is |
| * CONNECTED). |
| */ |
| void onIceConnected(); |
| |
| /** |
| * Callback fired once connection is closed (IceConnectionState is |
| * DISCONNECTED). |
| */ |
| void onIceDisconnected(); |
| |
| /** |
| * Callback fired once peer connection is closed. |
| */ |
| void onPeerConnectionClosed(); |
| |
| /** |
| * Callback fired once peer connection statistics is ready. |
| */ |
| void onPeerConnectionStatsReady(final StatsReport[] reports); |
| |
| /** |
| * Callback fired once peer connection error happened. |
| */ |
| void onPeerConnectionError(final String description); |
| } |
| |
| public PeerConnectionClient(Context appContext) { |
| if (appContext == null) { |
| throw new NullPointerException("The application context is null"); |
| } |
| rootEglBase = EglBase.create(); |
| this.appContext = appContext; |
| } |
| |
| public void setPeerConnectionFactoryOptions(PeerConnectionFactory.Options options) { |
| this.options = options; |
| } |
| |
| public void createPeerConnectionFactory( |
| final PeerConnectionParameters peerConnectionParameters, final PeerConnectionEvents events) { |
| this.peerConnectionParameters = peerConnectionParameters; |
| this.events = events; |
| videoCallEnabled = peerConnectionParameters.videoCallEnabled; |
| dataChannelEnabled = peerConnectionParameters.dataChannelParameters != null; |
| // Reset variables to initial states. |
| factory = null; |
| peerConnection = null; |
| preferIsac = false; |
| videoCapturerStopped = false; |
| isError = false; |
| queuedRemoteCandidates = null; |
| localSdp = null; // either offer or answer SDP |
| mediaStream = null; |
| videoCapturer = null; |
| renderVideo = true; |
| localVideoTrack = null; |
| remoteVideoTrack = null; |
| localVideoSender = null; |
| enableAudio = true; |
| localAudioTrack = null; |
| statsTimer = new Timer(); |
| |
| executor.execute(new Runnable() { |
| @Override |
| public void run() { |
| createPeerConnectionFactoryInternal(); |
| } |
| }); |
| } |
| |
| public void createPeerConnection(final VideoSink localRender, |
| final VideoRenderer.Callbacks remoteRender, final VideoCapturer videoCapturer, |
| final SignalingParameters signalingParameters) { |
| createPeerConnection( |
| localRender, Collections.singletonList(remoteRender), videoCapturer, signalingParameters); |
| } |
| |
| public void createPeerConnection(final VideoSink localRender, |
| final List<VideoRenderer.Callbacks> remoteRenders, final VideoCapturer videoCapturer, |
| final SignalingParameters signalingParameters) { |
| if (peerConnectionParameters == null) { |
| Log.e(TAG, "Creating peer connection without initializing factory."); |
| return; |
| } |
| this.localRender = localRender; |
| this.remoteRenders = remoteRenders; |
| this.videoCapturer = videoCapturer; |
| this.signalingParameters = signalingParameters; |
| executor.execute(new Runnable() { |
| @Override |
| public void run() { |
| try { |
| createMediaConstraintsInternal(); |
| createPeerConnectionInternal(); |
| maybeCreateAndStartRtcEventLog(); |
| } catch (Exception e) { |
| reportError("Failed to create peer connection: " + e.getMessage()); |
| throw e; |
| } |
| } |
| }); |
| } |
| |
| public void close() { |
| executor.execute(new Runnable() { |
| @Override |
| public void run() { |
| closeInternal(); |
| } |
| }); |
| } |
| |
| public boolean isVideoCallEnabled() { |
| return videoCallEnabled; |
| } |
| |
| private void createPeerConnectionFactoryInternal() { |
| isError = false; |
| |
| // Initialize field trials. |
| String fieldTrials = ""; |
| if (peerConnectionParameters.videoFlexfecEnabled) { |
| fieldTrials += VIDEO_FLEXFEC_FIELDTRIAL; |
| Log.d(TAG, "Enable FlexFEC field trial."); |
| } |
| fieldTrials += VIDEO_VP8_INTEL_HW_ENCODER_FIELDTRIAL; |
| if (peerConnectionParameters.disableWebRtcAGCAndHPF) { |
| fieldTrials += DISABLE_WEBRTC_AGC_FIELDTRIAL; |
| Log.d(TAG, "Disable WebRTC AGC field trial."); |
| } |
| |
| // Check preferred video codec. |
| preferredVideoCodec = VIDEO_CODEC_VP8; |
| if (videoCallEnabled && peerConnectionParameters.videoCodec != null) { |
| switch (peerConnectionParameters.videoCodec) { |
| case VIDEO_CODEC_VP8: |
| preferredVideoCodec = VIDEO_CODEC_VP8; |
| break; |
| case VIDEO_CODEC_VP9: |
| preferredVideoCodec = VIDEO_CODEC_VP9; |
| break; |
| case VIDEO_CODEC_H264_BASELINE: |
| preferredVideoCodec = VIDEO_CODEC_H264; |
| break; |
| case VIDEO_CODEC_H264_HIGH: |
| // TODO(magjed): Strip High from SDP when selecting Baseline instead of using field trial. |
| fieldTrials += VIDEO_H264_HIGH_PROFILE_FIELDTRIAL; |
| preferredVideoCodec = VIDEO_CODEC_H264; |
| break; |
| default: |
| preferredVideoCodec = VIDEO_CODEC_VP8; |
| } |
| } |
| Log.d(TAG, "Preferred video codec: " + preferredVideoCodec); |
| |
| // Initialize WebRTC |
| Log.d(TAG, |
| "Initialize WebRTC. Field trials: " + fieldTrials + " Enable video HW acceleration: " |
| + peerConnectionParameters.videoCodecHwAcceleration); |
| PeerConnectionFactory.initialize( |
| PeerConnectionFactory.InitializationOptions.builder(appContext) |
| .setFieldTrials(fieldTrials) |
| .setEnableVideoHwAcceleration(peerConnectionParameters.videoCodecHwAcceleration) |
| .setEnableInternalTracer(true) |
| .createInitializationOptions()); |
| if (peerConnectionParameters.tracing) { |
| PeerConnectionFactory.startInternalTracingCapture( |
| Environment.getExternalStorageDirectory().getAbsolutePath() + File.separator |
| + "webrtc-trace.txt"); |
| } |
| |
| // Check if ISAC is used by default. |
| preferIsac = peerConnectionParameters.audioCodec != null |
| && peerConnectionParameters.audioCodec.equals(AUDIO_CODEC_ISAC); |
| |
| // Enable/disable OpenSL ES playback. |
| if (!peerConnectionParameters.useOpenSLES) { |
| Log.d(TAG, "Disable OpenSL ES audio even if device supports it"); |
| WebRtcAudioManager.setBlacklistDeviceForOpenSLESUsage(true /* enable */); |
| } else { |
| Log.d(TAG, "Allow OpenSL ES audio if device supports it"); |
| WebRtcAudioManager.setBlacklistDeviceForOpenSLESUsage(false); |
| } |
| |
| if (peerConnectionParameters.disableBuiltInAEC) { |
| Log.d(TAG, "Disable built-in AEC even if device supports it"); |
| WebRtcAudioUtils.setWebRtcBasedAcousticEchoCanceler(true); |
| } else { |
| Log.d(TAG, "Enable built-in AEC if device supports it"); |
| WebRtcAudioUtils.setWebRtcBasedAcousticEchoCanceler(false); |
| } |
| |
| if (peerConnectionParameters.disableBuiltInAGC) { |
| Log.d(TAG, "Disable built-in AGC even if device supports it"); |
| WebRtcAudioUtils.setWebRtcBasedAutomaticGainControl(true); |
| } else { |
| Log.d(TAG, "Enable built-in AGC if device supports it"); |
| WebRtcAudioUtils.setWebRtcBasedAutomaticGainControl(false); |
| } |
| |
| if (peerConnectionParameters.disableBuiltInNS) { |
| Log.d(TAG, "Disable built-in NS even if device supports it"); |
| WebRtcAudioUtils.setWebRtcBasedNoiseSuppressor(true); |
| } else { |
| Log.d(TAG, "Enable built-in NS if device supports it"); |
| WebRtcAudioUtils.setWebRtcBasedNoiseSuppressor(false); |
| } |
| |
| // Set audio record error callbacks. |
| WebRtcAudioRecord.setErrorCallback(new WebRtcAudioRecordErrorCallback() { |
| @Override |
| public void onWebRtcAudioRecordInitError(String errorMessage) { |
| Log.e(TAG, "onWebRtcAudioRecordInitError: " + errorMessage); |
| reportError(errorMessage); |
| } |
| |
| @Override |
| public void onWebRtcAudioRecordStartError( |
| AudioRecordStartErrorCode errorCode, String errorMessage) { |
| Log.e(TAG, "onWebRtcAudioRecordStartError: " + errorCode + ". " + errorMessage); |
| reportError(errorMessage); |
| } |
| |
| @Override |
| public void onWebRtcAudioRecordError(String errorMessage) { |
| Log.e(TAG, "onWebRtcAudioRecordError: " + errorMessage); |
| reportError(errorMessage); |
| } |
| }); |
| |
| // It is possible to save a copy in raw PCM format on a file by checking |
| // the "Save input audio to file" checkbox in the Settings UI. A callback |
| // interface is set when this flag is enabled. As a result, a copy of recorded |
| // audio samples are provided to this client directly from the native audio |
| // layer in Java. |
| if (peerConnectionParameters.saveInputAudioToFile) { |
| if (!peerConnectionParameters.useOpenSLES) { |
| Log.d(TAG, "Enable recording of microphone input audio to file"); |
| saveRecordedAudioToFile = new RecordedAudioToFileController(executor); |
| } else { |
| // TODO(henrika): ensure that the UI reflects that if OpenSL ES is selected, |
| // then the "Save inut audio to file" option shall be grayed out. |
| Log.e(TAG, "Recording of input audio is not supported for OpenSL ES"); |
| } |
| } |
| |
| WebRtcAudioTrack.setErrorCallback(new WebRtcAudioTrack.ErrorCallback() { |
| @Override |
| public void onWebRtcAudioTrackInitError(String errorMessage) { |
| Log.e(TAG, "onWebRtcAudioTrackInitError: " + errorMessage); |
| reportError(errorMessage); |
| } |
| |
| @Override |
| public void onWebRtcAudioTrackStartError( |
| AudioTrackStartErrorCode errorCode, String errorMessage) { |
| Log.e(TAG, "onWebRtcAudioTrackStartError: " + errorCode + ". " + errorMessage); |
| reportError(errorMessage); |
| } |
| |
| @Override |
| public void onWebRtcAudioTrackError(String errorMessage) { |
| Log.e(TAG, "onWebRtcAudioTrackError: " + errorMessage); |
| reportError(errorMessage); |
| } |
| }); |
| |
| // Create peer connection factory. |
| if (options != null) { |
| Log.d(TAG, "Factory networkIgnoreMask option: " + options.networkIgnoreMask); |
| } |
| final boolean enableH264HighProfile = |
| VIDEO_CODEC_H264_HIGH.equals(peerConnectionParameters.videoCodec); |
| final VideoEncoderFactory encoderFactory; |
| final VideoDecoderFactory decoderFactory; |
| |
| if (peerConnectionParameters.videoCodecHwAcceleration) { |
| encoderFactory = new DefaultVideoEncoderFactory( |
| rootEglBase.getEglBaseContext(), true /* enableIntelVp8Encoder */, enableH264HighProfile); |
| decoderFactory = new DefaultVideoDecoderFactory(rootEglBase.getEglBaseContext()); |
| } else { |
| encoderFactory = new SoftwareVideoEncoderFactory(); |
| decoderFactory = new SoftwareVideoDecoderFactory(); |
| } |
| |
| factory = new PeerConnectionFactory(options, encoderFactory, decoderFactory); |
| Log.d(TAG, "Peer connection factory created."); |
| } |
| |
| private void createMediaConstraintsInternal() { |
| // Check if there is a camera on device and disable video call if not. |
| if (videoCapturer == null) { |
| Log.w(TAG, "No camera on device. Switch to audio only call."); |
| videoCallEnabled = false; |
| } |
| // Create video constraints if video call is enabled. |
| if (videoCallEnabled) { |
| videoWidth = peerConnectionParameters.videoWidth; |
| videoHeight = peerConnectionParameters.videoHeight; |
| videoFps = peerConnectionParameters.videoFps; |
| |
| // If video resolution is not specified, default to HD. |
| if (videoWidth == 0 || videoHeight == 0) { |
| videoWidth = HD_VIDEO_WIDTH; |
| videoHeight = HD_VIDEO_HEIGHT; |
| } |
| |
| // If fps is not specified, default to 30. |
| if (videoFps == 0) { |
| videoFps = 30; |
| } |
| Logging.d(TAG, "Capturing format: " + videoWidth + "x" + videoHeight + "@" + videoFps); |
| } |
| |
| // Create audio constraints. |
| audioConstraints = new MediaConstraints(); |
| // added for audio performance measurements |
| if (peerConnectionParameters.noAudioProcessing) { |
| Log.d(TAG, "Disabling audio processing"); |
| audioConstraints.mandatory.add( |
| new MediaConstraints.KeyValuePair(AUDIO_ECHO_CANCELLATION_CONSTRAINT, "false")); |
| audioConstraints.mandatory.add( |
| new MediaConstraints.KeyValuePair(AUDIO_AUTO_GAIN_CONTROL_CONSTRAINT, "false")); |
| audioConstraints.mandatory.add( |
| new MediaConstraints.KeyValuePair(AUDIO_HIGH_PASS_FILTER_CONSTRAINT, "false")); |
| audioConstraints.mandatory.add( |
| new MediaConstraints.KeyValuePair(AUDIO_NOISE_SUPPRESSION_CONSTRAINT, "false")); |
| } |
| // Create SDP constraints. |
| sdpMediaConstraints = new MediaConstraints(); |
| sdpMediaConstraints.mandatory.add( |
| new MediaConstraints.KeyValuePair("OfferToReceiveAudio", "true")); |
| if (videoCallEnabled || peerConnectionParameters.loopback) { |
| sdpMediaConstraints.mandatory.add( |
| new MediaConstraints.KeyValuePair("OfferToReceiveVideo", "true")); |
| } else { |
| sdpMediaConstraints.mandatory.add( |
| new MediaConstraints.KeyValuePair("OfferToReceiveVideo", "false")); |
| } |
| } |
| |
| private void createPeerConnectionInternal() { |
| if (factory == null || isError) { |
| Log.e(TAG, "Peerconnection factory is not created"); |
| return; |
| } |
| Log.d(TAG, "Create peer connection."); |
| |
| queuedRemoteCandidates = new ArrayList<>(); |
| |
| if (videoCallEnabled) { |
| factory.setVideoHwAccelerationOptions( |
| rootEglBase.getEglBaseContext(), rootEglBase.getEglBaseContext()); |
| } |
| |
| PeerConnection.RTCConfiguration rtcConfig = |
| new PeerConnection.RTCConfiguration(signalingParameters.iceServers); |
| // TCP candidates are only useful when connecting to a server that supports |
| // ICE-TCP. |
| rtcConfig.tcpCandidatePolicy = PeerConnection.TcpCandidatePolicy.DISABLED; |
| rtcConfig.bundlePolicy = PeerConnection.BundlePolicy.MAXBUNDLE; |
| rtcConfig.rtcpMuxPolicy = PeerConnection.RtcpMuxPolicy.REQUIRE; |
| rtcConfig.continualGatheringPolicy = PeerConnection.ContinualGatheringPolicy.GATHER_CONTINUALLY; |
| // Use ECDSA encryption. |
| rtcConfig.keyType = PeerConnection.KeyType.ECDSA; |
| // Enable DTLS for normal calls and disable for loopback calls. |
| rtcConfig.enableDtlsSrtp = !peerConnectionParameters.loopback; |
| rtcConfig.sdpSemantics = PeerConnection.SdpSemantics.UNIFIED_PLAN; |
| |
| peerConnection = factory.createPeerConnection(rtcConfig, pcObserver); |
| |
| if (dataChannelEnabled) { |
| DataChannel.Init init = new DataChannel.Init(); |
| init.ordered = peerConnectionParameters.dataChannelParameters.ordered; |
| init.negotiated = peerConnectionParameters.dataChannelParameters.negotiated; |
| init.maxRetransmits = peerConnectionParameters.dataChannelParameters.maxRetransmits; |
| init.maxRetransmitTimeMs = peerConnectionParameters.dataChannelParameters.maxRetransmitTimeMs; |
| init.id = peerConnectionParameters.dataChannelParameters.id; |
| init.protocol = peerConnectionParameters.dataChannelParameters.protocol; |
| dataChannel = peerConnection.createDataChannel("ApprtcDemo data", init); |
| } |
| isInitiator = false; |
| |
| // Set INFO libjingle logging. |
| // NOTE: this _must_ happen while |factory| is alive! |
| Logging.enableLogToDebugOutput(Logging.Severity.LS_INFO); |
| |
| List<String> mediaStreamLabels = Collections.singletonList("ARDAMS"); |
| if (videoCallEnabled) { |
| peerConnection.addTrack(createVideoTrack(videoCapturer), mediaStreamLabels); |
| // We can add the renderers right away because we don't need to wait for an |
| // answer to get the remote track. |
| remoteVideoTrack = getRemoteVideoTrack(); |
| remoteVideoTrack.setEnabled(renderVideo); |
| for (VideoRenderer.Callbacks remoteRender : remoteRenders) { |
| remoteVideoTrack.addRenderer(new VideoRenderer(remoteRender)); |
| } |
| } |
| peerConnection.addTrack(createAudioTrack(), mediaStreamLabels); |
| if (videoCallEnabled) { |
| findVideoSender(); |
| } |
| |
| if (peerConnectionParameters.aecDump) { |
| try { |
| ParcelFileDescriptor aecDumpFileDescriptor = |
| ParcelFileDescriptor.open(new File(Environment.getExternalStorageDirectory().getPath() |
| + File.separator + "Download/audio.aecdump"), |
| ParcelFileDescriptor.MODE_READ_WRITE | ParcelFileDescriptor.MODE_CREATE |
| | ParcelFileDescriptor.MODE_TRUNCATE); |
| factory.startAecDump(aecDumpFileDescriptor.detachFd(), -1); |
| } catch (IOException e) { |
| Log.e(TAG, "Can not open aecdump file", e); |
| } |
| } |
| |
| if (saveRecordedAudioToFile != null) { |
| if (saveRecordedAudioToFile.start()) { |
| Log.d(TAG, "Recording input audio to file is activated"); |
| } |
| } |
| Log.d(TAG, "Peer connection created."); |
| } |
| |
| private File createRtcEventLogOutputFile() { |
| DateFormat dateFormat = new SimpleDateFormat("yyyyMMdd_hhmm_ss", Locale.getDefault()); |
| Date date = new Date(); |
| final String outputFileName = "event_log_" + dateFormat.format(date) + ".log"; |
| return new File( |
| appContext.getDir(RTCEVENTLOG_OUTPUT_DIR_NAME, Context.MODE_PRIVATE), outputFileName); |
| } |
| |
| private void maybeCreateAndStartRtcEventLog() { |
| if (appContext == null || peerConnection == null) { |
| return; |
| } |
| if (!peerConnectionParameters.enableRtcEventLog) { |
| Log.d(TAG, "RtcEventLog is disabled."); |
| return; |
| } |
| rtcEventLog = new RtcEventLog(peerConnection); |
| rtcEventLog.start(createRtcEventLogOutputFile()); |
| } |
| |
| private void closeInternal() { |
| if (factory != null && peerConnectionParameters.aecDump) { |
| factory.stopAecDump(); |
| } |
| Log.d(TAG, "Closing peer connection."); |
| statsTimer.cancel(); |
| if (dataChannel != null) { |
| dataChannel.dispose(); |
| dataChannel = null; |
| } |
| if (rtcEventLog != null) { |
| // RtcEventLog should stop before the peer connection is disposed. |
| rtcEventLog.stop(); |
| rtcEventLog = null; |
| } |
| if (peerConnection != null) { |
| peerConnection.dispose(); |
| peerConnection = null; |
| } |
| Log.d(TAG, "Closing audio source."); |
| if (audioSource != null) { |
| audioSource.dispose(); |
| audioSource = null; |
| } |
| Log.d(TAG, "Stopping capture."); |
| if (videoCapturer != null) { |
| try { |
| videoCapturer.stopCapture(); |
| } catch (InterruptedException e) { |
| throw new RuntimeException(e); |
| } |
| videoCapturerStopped = true; |
| videoCapturer.dispose(); |
| videoCapturer = null; |
| } |
| Log.d(TAG, "Closing video source."); |
| if (videoSource != null) { |
| videoSource.dispose(); |
| videoSource = null; |
| } |
| if (saveRecordedAudioToFile != null) { |
| Log.d(TAG, "Closing audio file for recorded input audio."); |
| saveRecordedAudioToFile.stop(); |
| saveRecordedAudioToFile = null; |
| } |
| localRender = null; |
| remoteRenders = null; |
| Log.d(TAG, "Closing peer connection factory."); |
| if (factory != null) { |
| factory.dispose(); |
| factory = null; |
| } |
| options = null; |
| rootEglBase.release(); |
| Log.d(TAG, "Closing peer connection done."); |
| events.onPeerConnectionClosed(); |
| PeerConnectionFactory.stopInternalTracingCapture(); |
| PeerConnectionFactory.shutdownInternalTracer(); |
| events = null; |
| } |
| |
| public boolean isHDVideo() { |
| return videoCallEnabled && videoWidth * videoHeight >= 1280 * 720; |
| } |
| |
| public EglBase.Context getRenderContext() { |
| return rootEglBase.getEglBaseContext(); |
| } |
| |
| @SuppressWarnings("deprecation") // TODO(sakal): getStats is deprecated. |
| private void getStats() { |
| if (peerConnection == null || isError) { |
| return; |
| } |
| boolean success = peerConnection.getStats(new StatsObserver() { |
| @Override |
| public void onComplete(final StatsReport[] reports) { |
| events.onPeerConnectionStatsReady(reports); |
| } |
| }, null); |
| if (!success) { |
| Log.e(TAG, "getStats() returns false!"); |
| } |
| } |
| |
| public void enableStatsEvents(boolean enable, int periodMs) { |
| if (enable) { |
| try { |
| statsTimer.schedule(new TimerTask() { |
| @Override |
| public void run() { |
| executor.execute(new Runnable() { |
| @Override |
| public void run() { |
| getStats(); |
| } |
| }); |
| } |
| }, 0, periodMs); |
| } catch (Exception e) { |
| Log.e(TAG, "Can not schedule statistics timer", e); |
| } |
| } else { |
| statsTimer.cancel(); |
| } |
| } |
| |
| public void setAudioEnabled(final boolean enable) { |
| executor.execute(new Runnable() { |
| @Override |
| public void run() { |
| enableAudio = enable; |
| if (localAudioTrack != null) { |
| localAudioTrack.setEnabled(enableAudio); |
| } |
| } |
| }); |
| } |
| |
| public void setVideoEnabled(final boolean enable) { |
| executor.execute(new Runnable() { |
| @Override |
| public void run() { |
| renderVideo = enable; |
| if (localVideoTrack != null) { |
| localVideoTrack.setEnabled(renderVideo); |
| } |
| if (remoteVideoTrack != null) { |
| remoteVideoTrack.setEnabled(renderVideo); |
| } |
| } |
| }); |
| } |
| |
| public void createOffer() { |
| executor.execute(new Runnable() { |
| @Override |
| public void run() { |
| if (peerConnection != null && !isError) { |
| Log.d(TAG, "PC Create OFFER"); |
| isInitiator = true; |
| peerConnection.createOffer(sdpObserver, sdpMediaConstraints); |
| } |
| } |
| }); |
| } |
| |
| public void createAnswer() { |
| executor.execute(new Runnable() { |
| @Override |
| public void run() { |
| if (peerConnection != null && !isError) { |
| Log.d(TAG, "PC create ANSWER"); |
| isInitiator = false; |
| peerConnection.createAnswer(sdpObserver, sdpMediaConstraints); |
| } |
| } |
| }); |
| } |
| |
| public void addRemoteIceCandidate(final IceCandidate candidate) { |
| executor.execute(new Runnable() { |
| @Override |
| public void run() { |
| if (peerConnection != null && !isError) { |
| if (queuedRemoteCandidates != null) { |
| queuedRemoteCandidates.add(candidate); |
| } else { |
| peerConnection.addIceCandidate(candidate); |
| } |
| } |
| } |
| }); |
| } |
| |
| public void removeRemoteIceCandidates(final IceCandidate[] candidates) { |
| executor.execute(new Runnable() { |
| @Override |
| public void run() { |
| if (peerConnection == null || isError) { |
| return; |
| } |
| // Drain the queued remote candidates if there is any so that |
| // they are processed in the proper order. |
| drainCandidates(); |
| peerConnection.removeIceCandidates(candidates); |
| } |
| }); |
| } |
| |
| public void setRemoteDescription(final SessionDescription sdp) { |
| executor.execute(new Runnable() { |
| @Override |
| public void run() { |
| if (peerConnection == null || isError) { |
| return; |
| } |
| String sdpDescription = sdp.description; |
| if (preferIsac) { |
| sdpDescription = preferCodec(sdpDescription, AUDIO_CODEC_ISAC, true); |
| } |
| if (videoCallEnabled) { |
| sdpDescription = preferCodec(sdpDescription, preferredVideoCodec, false); |
| } |
| if (peerConnectionParameters.audioStartBitrate > 0) { |
| sdpDescription = setStartBitrate( |
| AUDIO_CODEC_OPUS, false, sdpDescription, peerConnectionParameters.audioStartBitrate); |
| } |
| Log.d(TAG, "Set remote SDP."); |
| SessionDescription sdpRemote = new SessionDescription(sdp.type, sdpDescription); |
| peerConnection.setRemoteDescription(sdpObserver, sdpRemote); |
| } |
| }); |
| } |
| |
| public void stopVideoSource() { |
| executor.execute(new Runnable() { |
| @Override |
| public void run() { |
| if (videoCapturer != null && !videoCapturerStopped) { |
| Log.d(TAG, "Stop video source."); |
| try { |
| videoCapturer.stopCapture(); |
| } catch (InterruptedException e) { |
| } |
| videoCapturerStopped = true; |
| } |
| } |
| }); |
| } |
| |
| public void startVideoSource() { |
| executor.execute(new Runnable() { |
| @Override |
| public void run() { |
| if (videoCapturer != null && videoCapturerStopped) { |
| Log.d(TAG, "Restart video source."); |
| videoCapturer.startCapture(videoWidth, videoHeight, videoFps); |
| videoCapturerStopped = false; |
| } |
| } |
| }); |
| } |
| |
| public void setVideoMaxBitrate(final Integer maxBitrateKbps) { |
| executor.execute(new Runnable() { |
| @Override |
| public void run() { |
| if (peerConnection == null || localVideoSender == null || isError) { |
| return; |
| } |
| Log.d(TAG, "Requested max video bitrate: " + maxBitrateKbps); |
| if (localVideoSender == null) { |
| Log.w(TAG, "Sender is not ready."); |
| return; |
| } |
| |
| RtpParameters parameters = localVideoSender.getParameters(); |
| if (parameters.encodings.size() == 0) { |
| Log.w(TAG, "RtpParameters are not ready."); |
| return; |
| } |
| |
| for (RtpParameters.Encoding encoding : parameters.encodings) { |
| // Null value means no limit. |
| encoding.maxBitrateBps = maxBitrateKbps == null ? null : maxBitrateKbps * BPS_IN_KBPS; |
| } |
| if (!localVideoSender.setParameters(parameters)) { |
| Log.e(TAG, "RtpSender.setParameters failed."); |
| } |
| Log.d(TAG, "Configured max video bitrate to: " + maxBitrateKbps); |
| } |
| }); |
| } |
| |
| private void reportError(final String errorMessage) { |
| Log.e(TAG, "Peerconnection error: " + errorMessage); |
| executor.execute(new Runnable() { |
| @Override |
| public void run() { |
| if (!isError) { |
| events.onPeerConnectionError(errorMessage); |
| isError = true; |
| } |
| } |
| }); |
| } |
| |
| private AudioTrack createAudioTrack() { |
| audioSource = factory.createAudioSource(audioConstraints); |
| localAudioTrack = factory.createAudioTrack(AUDIO_TRACK_ID, audioSource); |
| localAudioTrack.setEnabled(enableAudio); |
| return localAudioTrack; |
| } |
| |
| private VideoTrack createVideoTrack(VideoCapturer capturer) { |
| videoSource = factory.createVideoSource(capturer); |
| capturer.startCapture(videoWidth, videoHeight, videoFps); |
| |
| localVideoTrack = factory.createVideoTrack(VIDEO_TRACK_ID, videoSource); |
| localVideoTrack.setEnabled(renderVideo); |
| localVideoTrack.addSink(localRender); |
| return localVideoTrack; |
| } |
| |
| private void findVideoSender() { |
| for (RtpSender sender : peerConnection.getSenders()) { |
| if (sender.track() != null) { |
| String trackType = sender.track().kind(); |
| if (trackType.equals(VIDEO_TRACK_TYPE)) { |
| Log.d(TAG, "Found video sender."); |
| localVideoSender = sender; |
| } |
| } |
| } |
| } |
| |
| // Returns the remote VideoTrack, assuming there is only one. |
| private VideoTrack getRemoteVideoTrack() { |
| for (RtpTransceiver transceiver : peerConnection.getTransceivers()) { |
| MediaStreamTrack track = transceiver.getReceiver().track(); |
| if (track instanceof VideoTrack) { |
| return (VideoTrack) track; |
| } |
| } |
| return null; |
| } |
| |
| @SuppressWarnings("StringSplitter") |
| private static String setStartBitrate( |
| String codec, boolean isVideoCodec, String sdpDescription, int bitrateKbps) { |
| String[] lines = sdpDescription.split("\r\n"); |
| int rtpmapLineIndex = -1; |
| boolean sdpFormatUpdated = false; |
| String codecRtpMap = null; |
| // Search for codec rtpmap in format |
| // a=rtpmap:<payload type> <encoding name>/<clock rate> [/<encoding parameters>] |
| String regex = "^a=rtpmap:(\\d+) " + codec + "(/\\d+)+[\r]?$"; |
| Pattern codecPattern = Pattern.compile(regex); |
| for (int i = 0; i < lines.length; i++) { |
| Matcher codecMatcher = codecPattern.matcher(lines[i]); |
| if (codecMatcher.matches()) { |
| codecRtpMap = codecMatcher.group(1); |
| rtpmapLineIndex = i; |
| break; |
| } |
| } |
| if (codecRtpMap == null) { |
| Log.w(TAG, "No rtpmap for " + codec + " codec"); |
| return sdpDescription; |
| } |
| Log.d(TAG, "Found " + codec + " rtpmap " + codecRtpMap + " at " + lines[rtpmapLineIndex]); |
| |
| // Check if a=fmtp string already exist in remote SDP for this codec and |
| // update it with new bitrate parameter. |
| regex = "^a=fmtp:" + codecRtpMap + " \\w+=\\d+.*[\r]?$"; |
| codecPattern = Pattern.compile(regex); |
| for (int i = 0; i < lines.length; i++) { |
| Matcher codecMatcher = codecPattern.matcher(lines[i]); |
| if (codecMatcher.matches()) { |
| Log.d(TAG, "Found " + codec + " " + lines[i]); |
| if (isVideoCodec) { |
| lines[i] += "; " + VIDEO_CODEC_PARAM_START_BITRATE + "=" + bitrateKbps; |
| } else { |
| lines[i] += "; " + AUDIO_CODEC_PARAM_BITRATE + "=" + (bitrateKbps * 1000); |
| } |
| Log.d(TAG, "Update remote SDP line: " + lines[i]); |
| sdpFormatUpdated = true; |
| break; |
| } |
| } |
| |
| StringBuilder newSdpDescription = new StringBuilder(); |
| for (int i = 0; i < lines.length; i++) { |
| newSdpDescription.append(lines[i]).append("\r\n"); |
| // Append new a=fmtp line if no such line exist for a codec. |
| if (!sdpFormatUpdated && i == rtpmapLineIndex) { |
| String bitrateSet; |
| if (isVideoCodec) { |
| bitrateSet = |
| "a=fmtp:" + codecRtpMap + " " + VIDEO_CODEC_PARAM_START_BITRATE + "=" + bitrateKbps; |
| } else { |
| bitrateSet = "a=fmtp:" + codecRtpMap + " " + AUDIO_CODEC_PARAM_BITRATE + "=" |
| + (bitrateKbps * 1000); |
| } |
| Log.d(TAG, "Add remote SDP line: " + bitrateSet); |
| newSdpDescription.append(bitrateSet).append("\r\n"); |
| } |
| } |
| return newSdpDescription.toString(); |
| } |
| |
| /** Returns the line number containing "m=audio|video", or -1 if no such line exists. */ |
| private static int findMediaDescriptionLine(boolean isAudio, String[] sdpLines) { |
| final String mediaDescription = isAudio ? "m=audio " : "m=video "; |
| for (int i = 0; i < sdpLines.length; ++i) { |
| if (sdpLines[i].startsWith(mediaDescription)) { |
| return i; |
| } |
| } |
| return -1; |
| } |
| |
| private static String joinString( |
| Iterable<? extends CharSequence> s, String delimiter, boolean delimiterAtEnd) { |
| Iterator<? extends CharSequence> iter = s.iterator(); |
| if (!iter.hasNext()) { |
| return ""; |
| } |
| StringBuilder buffer = new StringBuilder(iter.next()); |
| while (iter.hasNext()) { |
| buffer.append(delimiter).append(iter.next()); |
| } |
| if (delimiterAtEnd) { |
| buffer.append(delimiter); |
| } |
| return buffer.toString(); |
| } |
| |
| private static String movePayloadTypesToFront(List<String> preferredPayloadTypes, String mLine) { |
| // The format of the media description line should be: m=<media> <port> <proto> <fmt> ... |
| final List<String> origLineParts = Arrays.asList(mLine.split(" ")); |
| if (origLineParts.size() <= 3) { |
| Log.e(TAG, "Wrong SDP media description format: " + mLine); |
| return null; |
| } |
| final List<String> header = origLineParts.subList(0, 3); |
| final List<String> unpreferredPayloadTypes = |
| new ArrayList<>(origLineParts.subList(3, origLineParts.size())); |
| unpreferredPayloadTypes.removeAll(preferredPayloadTypes); |
| // Reconstruct the line with |preferredPayloadTypes| moved to the beginning of the payload |
| // types. |
| final List<String> newLineParts = new ArrayList<>(); |
| newLineParts.addAll(header); |
| newLineParts.addAll(preferredPayloadTypes); |
| newLineParts.addAll(unpreferredPayloadTypes); |
| return joinString(newLineParts, " ", false /* delimiterAtEnd */); |
| } |
| |
| private static String preferCodec(String sdpDescription, String codec, boolean isAudio) { |
| final String[] lines = sdpDescription.split("\r\n"); |
| final int mLineIndex = findMediaDescriptionLine(isAudio, lines); |
| if (mLineIndex == -1) { |
| Log.w(TAG, "No mediaDescription line, so can't prefer " + codec); |
| return sdpDescription; |
| } |
| // A list with all the payload types with name |codec|. The payload types are integers in the |
| // range 96-127, but they are stored as strings here. |
| final List<String> codecPayloadTypes = new ArrayList<>(); |
| // a=rtpmap:<payload type> <encoding name>/<clock rate> [/<encoding parameters>] |
| final Pattern codecPattern = Pattern.compile("^a=rtpmap:(\\d+) " + codec + "(/\\d+)+[\r]?$"); |
| for (String line : lines) { |
| Matcher codecMatcher = codecPattern.matcher(line); |
| if (codecMatcher.matches()) { |
| codecPayloadTypes.add(codecMatcher.group(1)); |
| } |
| } |
| if (codecPayloadTypes.isEmpty()) { |
| Log.w(TAG, "No payload types with name " + codec); |
| return sdpDescription; |
| } |
| |
| final String newMLine = movePayloadTypesToFront(codecPayloadTypes, lines[mLineIndex]); |
| if (newMLine == null) { |
| return sdpDescription; |
| } |
| Log.d(TAG, "Change media description from: " + lines[mLineIndex] + " to " + newMLine); |
| lines[mLineIndex] = newMLine; |
| return joinString(Arrays.asList(lines), "\r\n", true /* delimiterAtEnd */); |
| } |
| |
| private void drainCandidates() { |
| if (queuedRemoteCandidates != null) { |
| Log.d(TAG, "Add " + queuedRemoteCandidates.size() + " remote candidates"); |
| for (IceCandidate candidate : queuedRemoteCandidates) { |
| peerConnection.addIceCandidate(candidate); |
| } |
| queuedRemoteCandidates = null; |
| } |
| } |
| |
| private void switchCameraInternal() { |
| if (videoCapturer instanceof CameraVideoCapturer) { |
| if (!videoCallEnabled || isError) { |
| Log.e(TAG, "Failed to switch camera. Video: " + videoCallEnabled + ". Error : " + isError); |
| return; // No video is sent or only one camera is available or error happened. |
| } |
| Log.d(TAG, "Switch camera"); |
| CameraVideoCapturer cameraVideoCapturer = (CameraVideoCapturer) videoCapturer; |
| cameraVideoCapturer.switchCamera(null); |
| } else { |
| Log.d(TAG, "Will not switch camera, video caputurer is not a camera"); |
| } |
| } |
| |
| public void switchCamera() { |
| executor.execute(new Runnable() { |
| @Override |
| public void run() { |
| switchCameraInternal(); |
| } |
| }); |
| } |
| |
| public void changeCaptureFormat(final int width, final int height, final int framerate) { |
| executor.execute(new Runnable() { |
| @Override |
| public void run() { |
| changeCaptureFormatInternal(width, height, framerate); |
| } |
| }); |
| } |
| |
| private void changeCaptureFormatInternal(int width, int height, int framerate) { |
| if (!videoCallEnabled || isError || videoCapturer == null) { |
| Log.e(TAG, |
| "Failed to change capture format. Video: " + videoCallEnabled + ". Error : " + isError); |
| return; |
| } |
| Log.d(TAG, "changeCaptureFormat: " + width + "x" + height + "@" + framerate); |
| videoSource.adaptOutputFormat(width, height, framerate); |
| } |
| |
| // Implementation detail: observe ICE & stream changes and react accordingly. |
| private class PCObserver implements PeerConnection.Observer { |
| @Override |
| public void onIceCandidate(final IceCandidate candidate) { |
| executor.execute(new Runnable() { |
| @Override |
| public void run() { |
| events.onIceCandidate(candidate); |
| } |
| }); |
| } |
| |
| @Override |
| public void onIceCandidatesRemoved(final IceCandidate[] candidates) { |
| executor.execute(new Runnable() { |
| @Override |
| public void run() { |
| events.onIceCandidatesRemoved(candidates); |
| } |
| }); |
| } |
| |
| @Override |
| public void onSignalingChange(PeerConnection.SignalingState newState) { |
| Log.d(TAG, "SignalingState: " + newState); |
| } |
| |
| @Override |
| public void onIceConnectionChange(final PeerConnection.IceConnectionState newState) { |
| executor.execute(new Runnable() { |
| @Override |
| public void run() { |
| Log.d(TAG, "IceConnectionState: " + newState); |
| if (newState == IceConnectionState.CONNECTED) { |
| events.onIceConnected(); |
| } else if (newState == IceConnectionState.DISCONNECTED) { |
| events.onIceDisconnected(); |
| } else if (newState == IceConnectionState.FAILED) { |
| reportError("ICE connection failed."); |
| } |
| } |
| }); |
| } |
| |
| @Override |
| public void onIceGatheringChange(PeerConnection.IceGatheringState newState) { |
| Log.d(TAG, "IceGatheringState: " + newState); |
| } |
| |
| @Override |
| public void onIceConnectionReceivingChange(boolean receiving) { |
| Log.d(TAG, "IceConnectionReceiving changed to " + receiving); |
| } |
| |
| @Override |
| public void onAddStream(final MediaStream stream) {} |
| |
| @Override |
| public void onRemoveStream(final MediaStream stream) {} |
| |
| @Override |
| public void onDataChannel(final DataChannel dc) { |
| Log.d(TAG, "New Data channel " + dc.label()); |
| |
| if (!dataChannelEnabled) |
| return; |
| |
| dc.registerObserver(new DataChannel.Observer() { |
| @Override |
| public void onBufferedAmountChange(long previousAmount) { |
| Log.d(TAG, "Data channel buffered amount changed: " + dc.label() + ": " + dc.state()); |
| } |
| |
| @Override |
| public void onStateChange() { |
| Log.d(TAG, "Data channel state changed: " + dc.label() + ": " + dc.state()); |
| } |
| |
| @Override |
| public void onMessage(final DataChannel.Buffer buffer) { |
| if (buffer.binary) { |
| Log.d(TAG, "Received binary msg over " + dc); |
| return; |
| } |
| ByteBuffer data = buffer.data; |
| final byte[] bytes = new byte[data.capacity()]; |
| data.get(bytes); |
| String strData = new String(bytes, Charset.forName("UTF-8")); |
| Log.d(TAG, "Got msg: " + strData + " over " + dc); |
| } |
| }); |
| } |
| |
| @Override |
| public void onRenegotiationNeeded() { |
| // No need to do anything; AppRTC follows a pre-agreed-upon |
| // signaling/negotiation protocol. |
| } |
| |
| @Override |
| public void onAddTrack(final RtpReceiver receiver, final MediaStream[] mediaStreams) {} |
| } |
| |
| // Implementation detail: handle offer creation/signaling and answer setting, |
| // as well as adding remote ICE candidates once the answer SDP is set. |
| private class SDPObserver implements SdpObserver { |
| @Override |
| public void onCreateSuccess(final SessionDescription origSdp) { |
| if (localSdp != null) { |
| reportError("Multiple SDP create."); |
| return; |
| } |
| String sdpDescription = origSdp.description; |
| if (preferIsac) { |
| sdpDescription = preferCodec(sdpDescription, AUDIO_CODEC_ISAC, true); |
| } |
| if (videoCallEnabled) { |
| sdpDescription = preferCodec(sdpDescription, preferredVideoCodec, false); |
| } |
| final SessionDescription sdp = new SessionDescription(origSdp.type, sdpDescription); |
| localSdp = sdp; |
| executor.execute(new Runnable() { |
| @Override |
| public void run() { |
| if (peerConnection != null && !isError) { |
| Log.d(TAG, "Set local SDP from " + sdp.type); |
| peerConnection.setLocalDescription(sdpObserver, sdp); |
| } |
| } |
| }); |
| } |
| |
| @Override |
| public void onSetSuccess() { |
| executor.execute(new Runnable() { |
| @Override |
| public void run() { |
| if (peerConnection == null || isError) { |
| return; |
| } |
| if (isInitiator) { |
| // For offering peer connection we first create offer and set |
| // local SDP, then after receiving answer set remote SDP. |
| if (peerConnection.getRemoteDescription() == null) { |
| // We've just set our local SDP so time to send it. |
| Log.d(TAG, "Local SDP set succesfully"); |
| events.onLocalDescription(localSdp); |
| } else { |
| // We've just set remote description, so drain remote |
| // and send local ICE candidates. |
| Log.d(TAG, "Remote SDP set succesfully"); |
| drainCandidates(); |
| } |
| } else { |
| // For answering peer connection we set remote SDP and then |
| // create answer and set local SDP. |
| if (peerConnection.getLocalDescription() != null) { |
| // We've just set our local SDP so time to send it, drain |
| // remote and send local ICE candidates. |
| Log.d(TAG, "Local SDP set succesfully"); |
| events.onLocalDescription(localSdp); |
| drainCandidates(); |
| } else { |
| // We've just set remote SDP - do nothing for now - |
| // answer will be created soon. |
| Log.d(TAG, "Remote SDP set succesfully"); |
| } |
| } |
| } |
| }); |
| } |
| |
| @Override |
| public void onCreateFailure(final String error) { |
| reportError("createSDP error: " + error); |
| } |
| |
| @Override |
| public void onSetFailure(final String error) { |
| reportError("setSDP error: " + error); |
| } |
| } |
| } |