Fixes logging levels in WebRtcAudioXXX.java classes
BUG=NONE
R=magjed@webrtc.org
Review URL: https://codereview.webrtc.org/1363673005 .
Cr-Commit-Position: refs/heads/master@{#10082}
diff --git a/webrtc/modules/audio_device/android/java/src/org/webrtc/voiceengine/WebRtcAudioRecord.java b/webrtc/modules/audio_device/android/java/src/org/webrtc/voiceengine/WebRtcAudioRecord.java
index 3df9e16..84e3fb8 100644
--- a/webrtc/modules/audio_device/android/java/src/org/webrtc/voiceengine/WebRtcAudioRecord.java
+++ b/webrtc/modules/audio_device/android/java/src/org/webrtc/voiceengine/WebRtcAudioRecord.java
@@ -70,7 +70,7 @@
@Override
public void run() {
Process.setThreadPriority(Process.THREAD_PRIORITY_URGENT_AUDIO);
- Logging.w(TAG, "AudioRecordThread" + WebRtcAudioUtils.getThreadInfo());
+ Logging.d(TAG, "AudioRecordThread" + WebRtcAudioUtils.getThreadInfo());
assertTrue(audioRecord.getRecordingState()
== AudioRecord.RECORDSTATE_RECORDING);
@@ -90,7 +90,7 @@
long durationInMs =
TimeUnit.NANOSECONDS.toMillis((nowTime - lastTime));
lastTime = nowTime;
- Logging.w(TAG, "bytesRead[" + durationInMs + "] " + bytesRead);
+ Logging.d(TAG, "bytesRead[" + durationInMs + "] " + bytesRead);
}
}
@@ -114,7 +114,7 @@
}
WebRtcAudioRecord(Context context, long nativeAudioRecord) {
- Logging.w(TAG, "ctor" + WebRtcAudioUtils.getThreadInfo());
+ Logging.d(TAG, "ctor" + WebRtcAudioUtils.getThreadInfo());
this.context = context;
this.nativeAudioRecord = nativeAudioRecord;
if (DEBUG) {
@@ -124,7 +124,7 @@
}
private boolean enableBuiltInAEC(boolean enable) {
- Logging.w(TAG, "enableBuiltInAEC(" + enable + ')');
+ Logging.d(TAG, "enableBuiltInAEC(" + enable + ')');
if (effects == null) {
Logging.e(TAG,"Built-in AEC is not supported on this platform");
return false;
@@ -133,7 +133,7 @@
}
private boolean enableBuiltInAGC(boolean enable) {
- Logging.w(TAG, "enableBuiltInAGC(" + enable + ')');
+ Logging.d(TAG, "enableBuiltInAGC(" + enable + ')');
if (effects == null) {
Logging.e(TAG,"Built-in AGC is not supported on this platform");
return false;
@@ -142,7 +142,7 @@
}
private boolean enableBuiltInNS(boolean enable) {
- Logging.w(TAG, "enableBuiltInNS(" + enable + ')');
+ Logging.d(TAG, "enableBuiltInNS(" + enable + ')');
if (effects == null) {
Logging.e(TAG,"Built-in NS is not supported on this platform");
return false;
@@ -151,7 +151,7 @@
}
private int initRecording(int sampleRate, int channels) {
- Logging.w(TAG, "initRecording(sampleRate=" + sampleRate + ", channels=" +
+ Logging.d(TAG, "initRecording(sampleRate=" + sampleRate + ", channels=" +
channels + ")");
if (!WebRtcAudioUtils.hasPermission(
context, android.Manifest.permission.RECORD_AUDIO)) {
@@ -165,7 +165,7 @@
final int bytesPerFrame = channels * (BITS_PER_SAMPLE / 8);
final int framesPerBuffer = sampleRate / BUFFERS_PER_SECOND;
byteBuffer = ByteBuffer.allocateDirect(bytesPerFrame * framesPerBuffer);
- Logging.w(TAG, "byteBuffer.capacity: " + byteBuffer.capacity());
+ Logging.d(TAG, "byteBuffer.capacity: " + byteBuffer.capacity());
// Rather than passing the ByteBuffer with every callback (requiring
// the potentially expensive GetDirectBufferAddress) we simply have the
// the native class cache the address to the memory once.
@@ -183,14 +183,14 @@
Logging.e(TAG, "AudioRecord.getMinBufferSize failed: " + minBufferSize);
return -1;
}
- Logging.w(TAG, "AudioRecord.getMinBufferSize: " + minBufferSize);
+ Logging.d(TAG, "AudioRecord.getMinBufferSize: " + minBufferSize);
// Use a larger buffer size than the minimum required when creating the
// AudioRecord instance to ensure smooth recording under load. It has been
// verified that it does not increase the actual recording latency.
int bufferSizeInBytes =
Math.max(BUFFER_SIZE_FACTOR * minBufferSize, byteBuffer.capacity());
- Logging.w(TAG, "bufferSizeInBytes: " + bufferSizeInBytes);
+ Logging.d(TAG, "bufferSizeInBytes: " + bufferSizeInBytes);
try {
audioRecord = new AudioRecord(AudioSource.VOICE_COMMUNICATION,
sampleRate,
@@ -206,7 +206,7 @@
Logging.e(TAG,"Failed to create a new AudioRecord instance");
return -1;
}
- Logging.w(TAG, "AudioRecord "
+ Logging.d(TAG, "AudioRecord "
+ "session ID: " + audioRecord.getAudioSessionId() + ", "
+ "audio format: " + audioRecord.getAudioFormat() + ", "
+ "channels: " + audioRecord.getChannelCount() + ", "
@@ -227,7 +227,7 @@
}
private boolean startRecording() {
- Logging.w(TAG, "startRecording");
+ Logging.d(TAG, "startRecording");
assertTrue(audioRecord != null);
assertTrue(audioThread == null);
try {
@@ -246,7 +246,7 @@
}
private boolean stopRecording() {
- Logging.w(TAG, "stopRecording");
+ Logging.d(TAG, "stopRecording");
assertTrue(audioThread != null);
audioThread.joinThread();
audioThread = null;
diff --git a/webrtc/modules/audio_device/android/java/src/org/webrtc/voiceengine/WebRtcAudioTrack.java b/webrtc/modules/audio_device/android/java/src/org/webrtc/voiceengine/WebRtcAudioTrack.java
index e99b9d7..43c1a19 100644
--- a/webrtc/modules/audio_device/android/java/src/org/webrtc/voiceengine/WebRtcAudioTrack.java
+++ b/webrtc/modules/audio_device/android/java/src/org/webrtc/voiceengine/WebRtcAudioTrack.java
@@ -61,7 +61,7 @@
@Override
public void run() {
Process.setThreadPriority(Process.THREAD_PRIORITY_URGENT_AUDIO);
- Logd("AudioTrackThread" + WebRtcAudioUtils.getThreadInfo());
+ Logging.d(TAG, "AudioTrackThread" + WebRtcAudioUtils.getThreadInfo());
try {
// In MODE_STREAM mode we can optionally prime the output buffer by
@@ -71,7 +71,7 @@
audioTrack.play();
assertTrue(audioTrack.getPlayState() == AudioTrack.PLAYSTATE_PLAYING);
} catch (IllegalStateException e) {
- Loge("AudioTrack.play failed: " + e.getMessage());
+ Logging.e(TAG, "AudioTrack.play failed: " + e.getMessage());
return;
}
@@ -99,7 +99,7 @@
sizeInBytes);
}
if (bytesWritten != sizeInBytes) {
- Loge("AudioTrack.write failed: " + bytesWritten);
+ Logging.e(TAG, "AudioTrack.write failed: " + bytesWritten);
if (bytesWritten == AudioTrack.ERROR_INVALID_OPERATION) {
keepAlive = false;
}
@@ -117,7 +117,7 @@
try {
audioTrack.stop();
} catch (IllegalStateException e) {
- Loge("AudioTrack.stop failed: " + e.getMessage());
+ Logging.e(TAG, "AudioTrack.stop failed: " + e.getMessage());
}
assertTrue(audioTrack.getPlayState() == AudioTrack.PLAYSTATE_STOPPED);
audioTrack.flush();
@@ -136,7 +136,7 @@
}
WebRtcAudioTrack(Context context, long nativeAudioTrack) {
- Logd("ctor" + WebRtcAudioUtils.getThreadInfo());
+ Logging.d(TAG, "ctor" + WebRtcAudioUtils.getThreadInfo());
this.context = context;
this.nativeAudioTrack = nativeAudioTrack;
audioManager = (AudioManager) context.getSystemService(
@@ -147,12 +147,12 @@
}
private void initPlayout(int sampleRate, int channels) {
- Logd("initPlayout(sampleRate=" + sampleRate + ", channels=" +
- channels + ")");
+ Logging.d(TAG, "initPlayout(sampleRate=" + sampleRate + ", channels="
+ + channels + ")");
final int bytesPerFrame = channels * (BITS_PER_SAMPLE / 8);
byteBuffer = byteBuffer.allocateDirect(
bytesPerFrame * (sampleRate / BUFFERS_PER_SECOND));
- Logd("byteBuffer.capacity: " + byteBuffer.capacity());
+ Logging.d(TAG, "byteBuffer.capacity: " + byteBuffer.capacity());
// Rather than passing the ByteBuffer with every callback (requiring
// the potentially expensive GetDirectBufferAddress) we simply have the
// the native class cache the address to the memory once.
@@ -166,7 +166,7 @@
sampleRate,
AudioFormat.CHANNEL_OUT_MONO,
AudioFormat.ENCODING_PCM_16BIT);
- Logd("AudioTrack.getMinBufferSize: " + minBufferSizeInBytes);
+ Logging.d(TAG, "AudioTrack.getMinBufferSize: " + minBufferSizeInBytes);
assertTrue(audioTrack == null);
// For the streaming mode, data must be written to the audio sink in
@@ -184,7 +184,7 @@
minBufferSizeInBytes,
AudioTrack.MODE_STREAM);
} catch (IllegalArgumentException e) {
- Logd(e.getMessage());
+ Logging.d(TAG, e.getMessage());
return;
}
assertTrue(audioTrack.getState() == AudioTrack.STATE_INITIALIZED);
@@ -193,7 +193,7 @@
}
private boolean startPlayout() {
- Logd("startPlayout");
+ Logging.d(TAG, "startPlayout");
assertTrue(audioTrack != null);
assertTrue(audioThread == null);
audioThread = new AudioTrackThread("AudioTrackJavaThread");
@@ -202,7 +202,7 @@
}
private boolean stopPlayout() {
- Logd("stopPlayout");
+ Logging.d(TAG, "stopPlayout");
assertTrue(audioThread != null);
audioThread.joinThread();
audioThread = null;
@@ -215,18 +215,18 @@
/** Get max possible volume index for a phone call audio stream. */
private int getStreamMaxVolume() {
- Logd("getStreamMaxVolume");
+ Logging.d(TAG, "getStreamMaxVolume");
assertTrue(audioManager != null);
return audioManager.getStreamMaxVolume(AudioManager.STREAM_VOICE_CALL);
}
/** Set current volume level for a phone call audio stream. */
private boolean setStreamVolume(int volume) {
- Logd("setStreamVolume(" + volume + ")");
+ Logging.d(TAG, "setStreamVolume(" + volume + ")");
assertTrue(audioManager != null);
if (WebRtcAudioUtils.runningOnLollipopOrHigher()) {
if (audioManager.isVolumeFixed()) {
- Loge("The device implements a fixed volume policy.");
+ Logging.e(TAG, "The device implements a fixed volume policy.");
return false;
}
}
@@ -236,7 +236,7 @@
/** Get current volume level for a phone call audio stream. */
private int getStreamVolume() {
- Logd("getStreamVolume");
+ Logging.d(TAG, "getStreamVolume");
assertTrue(audioManager != null);
return audioManager.getStreamVolume(AudioManager.STREAM_VOICE_CALL);
}
@@ -248,14 +248,6 @@
}
}
- private static void Logd(String msg) {
- Logging.d(TAG, msg);
- }
-
- private static void Loge(String msg) {
- Logging.e(TAG, msg);
- }
-
private native void nativeCacheDirectBufferAddress(
ByteBuffer byteBuffer, long nativeAudioRecord);