| /* |
| * Copyright (c) 2012 The WebRTC project authors. All Rights Reserved. |
| * |
| * Use of this source code is governed by a BSD-style license |
| * that can be found in the LICENSE file in the root of the source |
| * tree. An additional intellectual property rights grant can be found |
| * in the file PATENTS. All contributing project authors may |
| * be found in the AUTHORS file in the root of the source tree. |
| */ |
| |
| #include "voe_base_impl.h" |
| |
| #include "audio_coding_module.h" |
| #include "audio_processing.h" |
| #include "channel.h" |
| #include "critical_section_wrapper.h" |
| #include "file_wrapper.h" |
| #include "modules/audio_device/audio_device_impl.h" |
| #include "output_mixer.h" |
| #include "signal_processing_library.h" |
| #include "trace.h" |
| #include "transmit_mixer.h" |
| #include "utility.h" |
| #include "voe_errors.h" |
| #include "voice_engine_impl.h" |
| |
| #if (defined(_WIN32) && defined(_DLL) && (_MSC_VER == 1400)) |
| // Fix for VS 2005 MD/MDd link problem |
| #include <stdio.h> |
| extern "C" |
| { FILE _iob[3] = { __iob_func()[0], __iob_func()[1], __iob_func()[2]}; } |
| #endif |
| |
| namespace webrtc |
| { |
| |
| VoEBase* VoEBase::GetInterface(VoiceEngine* voiceEngine) |
| { |
| if (NULL == voiceEngine) |
| { |
| return NULL; |
| } |
| VoiceEngineImpl* s = static_cast<VoiceEngineImpl*>(voiceEngine); |
| s->AddRef(); |
| return s; |
| } |
| |
| VoEBaseImpl::VoEBaseImpl(voe::SharedData* shared) : |
| _voiceEngineObserverPtr(NULL), |
| _callbackCritSect(*CriticalSectionWrapper::CreateCriticalSection()), |
| _voiceEngineObserver(false), _oldVoEMicLevel(0), _oldMicLevel(0), |
| _shared(shared) |
| { |
| WEBRTC_TRACE(kTraceMemory, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "VoEBaseImpl() - ctor"); |
| } |
| |
| VoEBaseImpl::~VoEBaseImpl() |
| { |
| WEBRTC_TRACE(kTraceMemory, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "~VoEBaseImpl() - dtor"); |
| |
| TerminateInternal(); |
| |
| delete &_callbackCritSect; |
| } |
| |
| void VoEBaseImpl::OnErrorIsReported(const ErrorCode error) |
| { |
| CriticalSectionScoped cs(&_callbackCritSect); |
| if (_voiceEngineObserver) |
| { |
| if (_voiceEngineObserverPtr) |
| { |
| int errCode(0); |
| if (error == AudioDeviceObserver::kRecordingError) |
| { |
| errCode = VE_RUNTIME_REC_ERROR; |
| WEBRTC_TRACE(kTraceInfo, kTraceVoice, |
| VoEId(_shared->instance_id(), -1), |
| "VoEBaseImpl::OnErrorIsReported() => VE_RUNTIME_REC_ERROR"); |
| } |
| else if (error == AudioDeviceObserver::kPlayoutError) |
| { |
| errCode = VE_RUNTIME_PLAY_ERROR; |
| WEBRTC_TRACE(kTraceInfo, kTraceVoice, |
| VoEId(_shared->instance_id(), -1), |
| "VoEBaseImpl::OnErrorIsReported() => " |
| "VE_RUNTIME_PLAY_ERROR"); |
| } |
| // Deliver callback (-1 <=> no channel dependency) |
| _voiceEngineObserverPtr->CallbackOnError(-1, errCode); |
| } |
| } |
| } |
| |
| void VoEBaseImpl::OnWarningIsReported(const WarningCode warning) |
| { |
| CriticalSectionScoped cs(&_callbackCritSect); |
| if (_voiceEngineObserver) |
| { |
| if (_voiceEngineObserverPtr) |
| { |
| int warningCode(0); |
| if (warning == AudioDeviceObserver::kRecordingWarning) |
| { |
| warningCode = VE_RUNTIME_REC_WARNING; |
| WEBRTC_TRACE(kTraceInfo, kTraceVoice, |
| VoEId(_shared->instance_id(), -1), |
| "VoEBaseImpl::OnErrorIsReported() => " |
| "VE_RUNTIME_REC_WARNING"); |
| } |
| else if (warning == AudioDeviceObserver::kPlayoutWarning) |
| { |
| warningCode = VE_RUNTIME_PLAY_WARNING; |
| WEBRTC_TRACE(kTraceInfo, kTraceVoice, |
| VoEId(_shared->instance_id(), -1), |
| "VoEBaseImpl::OnErrorIsReported() => " |
| "VE_RUNTIME_PLAY_WARNING"); |
| } |
| // Deliver callback (-1 <=> no channel dependency) |
| _voiceEngineObserverPtr->CallbackOnError(-1, warningCode); |
| } |
| } |
| } |
| |
| int32_t VoEBaseImpl::RecordedDataIsAvailable( |
| const void* audioSamples, |
| const uint32_t nSamples, |
| const uint8_t nBytesPerSample, |
| const uint8_t nChannels, |
| const uint32_t samplesPerSec, |
| const uint32_t totalDelayMS, |
| const int32_t clockDrift, |
| const uint32_t currentMicLevel, |
| uint32_t& newMicLevel) |
| { |
| WEBRTC_TRACE(kTraceStream, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "VoEBaseImpl::RecordedDataIsAvailable(nSamples=%u, " |
| "nBytesPerSample=%u, nChannels=%u, samplesPerSec=%u, " |
| "totalDelayMS=%u, clockDrift=%d, currentMicLevel=%u)", |
| nSamples, nBytesPerSample, nChannels, samplesPerSec, |
| totalDelayMS, clockDrift, currentMicLevel); |
| |
| assert(_shared->transmit_mixer() != NULL); |
| assert(_shared->audio_device() != NULL); |
| |
| bool isAnalogAGC(false); |
| uint32_t maxVolume(0); |
| uint16_t currentVoEMicLevel(0); |
| uint32_t newVoEMicLevel(0); |
| |
| if (_shared->audio_processing() && |
| (_shared->audio_processing()->gain_control()->mode() |
| == GainControl::kAdaptiveAnalog)) |
| { |
| isAnalogAGC = true; |
| } |
| |
| // Will only deal with the volume in adaptive analog mode |
| if (isAnalogAGC) |
| { |
| // Scale from ADM to VoE level range |
| if (_shared->audio_device()->MaxMicrophoneVolume(&maxVolume) == 0) |
| { |
| if (0 != maxVolume) |
| { |
| currentVoEMicLevel = (uint16_t) ((currentMicLevel |
| * kMaxVolumeLevel + (int) (maxVolume / 2)) |
| / (maxVolume)); |
| } |
| } |
| // We learned that on certain systems (e.g Linux) the currentVoEMicLevel |
| // can be greater than the maxVolumeLevel therefore |
| // we are going to cap the currentVoEMicLevel to the maxVolumeLevel |
| // and change the maxVolume to currentMicLevel if it turns out that |
| // the currentVoEMicLevel is indeed greater than the maxVolumeLevel. |
| if (currentVoEMicLevel > kMaxVolumeLevel) |
| { |
| currentVoEMicLevel = kMaxVolumeLevel; |
| maxVolume = currentMicLevel; |
| } |
| } |
| |
| // Keep track if the MicLevel has been changed by the AGC, if not, |
| // use the old value AGC returns to let AGC continue its trend, |
| // so eventually the AGC is able to change the mic level. This handles |
| // issues with truncation introduced by the scaling. |
| if (_oldMicLevel == currentMicLevel) |
| { |
| currentVoEMicLevel = (uint16_t) _oldVoEMicLevel; |
| } |
| |
| // Perform channel-independent operations |
| // (APM, mix with file, record to file, mute, etc.) |
| _shared->transmit_mixer()->PrepareDemux(audioSamples, nSamples, nChannels, |
| samplesPerSec, static_cast<uint16_t>(totalDelayMS), clockDrift, |
| currentVoEMicLevel); |
| |
| // Copy the audio frame to each sending channel and perform |
| // channel-dependent operations (file mixing, mute, etc.) to prepare |
| // for encoding. |
| _shared->transmit_mixer()->DemuxAndMix(); |
| // Do the encoding and packetize+transmit the RTP packet when encoding |
| // is done. |
| _shared->transmit_mixer()->EncodeAndSend(); |
| |
| // Will only deal with the volume in adaptive analog mode |
| if (isAnalogAGC) |
| { |
| // Scale from VoE to ADM level range |
| newVoEMicLevel = _shared->transmit_mixer()->CaptureLevel(); |
| if (newVoEMicLevel != currentVoEMicLevel) |
| { |
| // Add (kMaxVolumeLevel/2) to round the value |
| newMicLevel = (uint32_t) ((newVoEMicLevel * maxVolume |
| + (int) (kMaxVolumeLevel / 2)) / (kMaxVolumeLevel)); |
| } |
| else |
| { |
| // Pass zero if the level is unchanged |
| newMicLevel = 0; |
| } |
| |
| // Keep track of the value AGC returns |
| _oldVoEMicLevel = newVoEMicLevel; |
| _oldMicLevel = currentMicLevel; |
| } |
| |
| return 0; |
| } |
| |
| int32_t VoEBaseImpl::NeedMorePlayData( |
| const uint32_t nSamples, |
| const uint8_t nBytesPerSample, |
| const uint8_t nChannels, |
| const uint32_t samplesPerSec, |
| void* audioSamples, |
| uint32_t& nSamplesOut) |
| { |
| WEBRTC_TRACE(kTraceStream, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "VoEBaseImpl::NeedMorePlayData(nSamples=%u, " |
| "nBytesPerSample=%d, nChannels=%d, samplesPerSec=%u)", |
| nSamples, nBytesPerSample, nChannels, samplesPerSec); |
| |
| assert(_shared->output_mixer() != NULL); |
| |
| // TODO(andrew): if the device is running in mono, we should tell the mixer |
| // here so that it will only request mono from AudioCodingModule. |
| // Perform mixing of all active participants (channel-based mixing) |
| _shared->output_mixer()->MixActiveChannels(); |
| |
| // Additional operations on the combined signal |
| _shared->output_mixer()->DoOperationsOnCombinedSignal(); |
| |
| // Retrieve the final output mix (resampled to match the ADM) |
| _shared->output_mixer()->GetMixedAudio(samplesPerSec, nChannels, |
| &_audioFrame); |
| |
| assert(static_cast<int>(nSamples) == _audioFrame.samples_per_channel_); |
| assert(samplesPerSec == |
| static_cast<uint32_t>(_audioFrame.sample_rate_hz_)); |
| |
| // Deliver audio (PCM) samples to the ADM |
| memcpy( |
| (int16_t*) audioSamples, |
| (const int16_t*) _audioFrame.data_, |
| sizeof(int16_t) * (_audioFrame.samples_per_channel_ |
| * _audioFrame.num_channels_)); |
| |
| nSamplesOut = _audioFrame.samples_per_channel_; |
| |
| return 0; |
| } |
| |
| int VoEBaseImpl::RegisterVoiceEngineObserver(VoiceEngineObserver& observer) |
| { |
| WEBRTC_TRACE(kTraceApiCall, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "RegisterVoiceEngineObserver(observer=0x%d)", &observer); |
| CriticalSectionScoped cs(&_callbackCritSect); |
| if (_voiceEngineObserverPtr) |
| { |
| _shared->SetLastError(VE_INVALID_OPERATION, kTraceError, |
| "RegisterVoiceEngineObserver() observer already enabled"); |
| return -1; |
| } |
| |
| // Register the observer in all active channels |
| voe::ScopedChannel sc(_shared->channel_manager()); |
| void* iterator(NULL); |
| voe::Channel* channelPtr = sc.GetFirstChannel(iterator); |
| while (channelPtr != NULL) |
| { |
| channelPtr->RegisterVoiceEngineObserver(observer); |
| channelPtr = sc.GetNextChannel(iterator); |
| } |
| _shared->transmit_mixer()->RegisterVoiceEngineObserver(observer); |
| |
| _voiceEngineObserverPtr = &observer; |
| _voiceEngineObserver = true; |
| |
| return 0; |
| } |
| |
| int VoEBaseImpl::DeRegisterVoiceEngineObserver() |
| { |
| WEBRTC_TRACE(kTraceApiCall, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "DeRegisterVoiceEngineObserver()"); |
| CriticalSectionScoped cs(&_callbackCritSect); |
| if (!_voiceEngineObserverPtr) |
| { |
| _shared->SetLastError(VE_INVALID_OPERATION, kTraceError, |
| "DeRegisterVoiceEngineObserver() observer already disabled"); |
| return 0; |
| } |
| |
| _voiceEngineObserver = false; |
| _voiceEngineObserverPtr = NULL; |
| |
| // Deregister the observer in all active channels |
| voe::ScopedChannel sc(_shared->channel_manager()); |
| void* iterator(NULL); |
| voe::Channel* channelPtr = sc.GetFirstChannel(iterator); |
| while (channelPtr != NULL) |
| { |
| channelPtr->DeRegisterVoiceEngineObserver(); |
| channelPtr = sc.GetNextChannel(iterator); |
| } |
| |
| return 0; |
| } |
| |
| int VoEBaseImpl::Init(AudioDeviceModule* external_adm, |
| AudioProcessing* audioproc) |
| { |
| WEBRTC_TRACE(kTraceApiCall, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "Init(external_adm=0x%p)", external_adm); |
| CriticalSectionScoped cs(_shared->crit_sec()); |
| |
| WebRtcSpl_Init(); |
| |
| if (_shared->statistics().Initialized()) |
| { |
| return 0; |
| } |
| |
| if (_shared->process_thread()) |
| { |
| if (_shared->process_thread()->Start() != 0) |
| { |
| _shared->SetLastError(VE_THREAD_ERROR, kTraceError, |
| "Init() failed to start module process thread"); |
| return -1; |
| } |
| } |
| |
| // Create an internal ADM if the user has not added an external |
| // ADM implementation as input to Init(). |
| if (external_adm == NULL) |
| { |
| // Create the internal ADM implementation. |
| _shared->set_audio_device(AudioDeviceModuleImpl::Create( |
| VoEId(_shared->instance_id(), -1), _shared->audio_device_layer())); |
| |
| if (_shared->audio_device() == NULL) |
| { |
| _shared->SetLastError(VE_NO_MEMORY, kTraceCritical, |
| "Init() failed to create the ADM"); |
| return -1; |
| } |
| } |
| else |
| { |
| // Use the already existing external ADM implementation. |
| _shared->set_audio_device(external_adm); |
| WEBRTC_TRACE(kTraceInfo, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "An external ADM implementation will be used in VoiceEngine"); |
| } |
| |
| // Register the ADM to the process thread, which will drive the error |
| // callback mechanism |
| if (_shared->process_thread() && |
| _shared->process_thread()->RegisterModule(_shared->audio_device()) != 0) |
| { |
| _shared->SetLastError(VE_AUDIO_DEVICE_MODULE_ERROR, kTraceError, |
| "Init() failed to register the ADM"); |
| return -1; |
| } |
| |
| bool available(false); |
| |
| // -------------------- |
| // Reinitialize the ADM |
| |
| // Register the AudioObserver implementation |
| if (_shared->audio_device()->RegisterEventObserver(this) != 0) { |
| _shared->SetLastError(VE_AUDIO_DEVICE_MODULE_ERROR, kTraceWarning, |
| "Init() failed to register event observer for the ADM"); |
| } |
| |
| // Register the AudioTransport implementation |
| if (_shared->audio_device()->RegisterAudioCallback(this) != 0) { |
| _shared->SetLastError(VE_AUDIO_DEVICE_MODULE_ERROR, kTraceWarning, |
| "Init() failed to register audio callback for the ADM"); |
| } |
| |
| // ADM initialization |
| if (_shared->audio_device()->Init() != 0) |
| { |
| _shared->SetLastError(VE_AUDIO_DEVICE_MODULE_ERROR, kTraceError, |
| "Init() failed to initialize the ADM"); |
| return -1; |
| } |
| |
| // Initialize the default speaker |
| if (_shared->audio_device()->SetPlayoutDevice( |
| WEBRTC_VOICE_ENGINE_DEFAULT_DEVICE) != 0) |
| { |
| _shared->SetLastError(VE_AUDIO_DEVICE_MODULE_ERROR, kTraceInfo, |
| "Init() failed to set the default output device"); |
| } |
| if (_shared->audio_device()->SpeakerIsAvailable(&available) != 0) |
| { |
| _shared->SetLastError(VE_CANNOT_ACCESS_SPEAKER_VOL, kTraceInfo, |
| "Init() failed to check speaker availability, trying to " |
| "initialize speaker anyway"); |
| } |
| else if (!available) |
| { |
| _shared->SetLastError(VE_CANNOT_ACCESS_SPEAKER_VOL, kTraceInfo, |
| "Init() speaker not available, trying to initialize speaker " |
| "anyway"); |
| } |
| if (_shared->audio_device()->InitSpeaker() != 0) |
| { |
| _shared->SetLastError(VE_CANNOT_ACCESS_SPEAKER_VOL, kTraceInfo, |
| "Init() failed to initialize the speaker"); |
| } |
| |
| // Initialize the default microphone |
| if (_shared->audio_device()->SetRecordingDevice( |
| WEBRTC_VOICE_ENGINE_DEFAULT_DEVICE) != 0) |
| { |
| _shared->SetLastError(VE_SOUNDCARD_ERROR, kTraceInfo, |
| "Init() failed to set the default input device"); |
| } |
| if (_shared->audio_device()->MicrophoneIsAvailable(&available) != 0) |
| { |
| _shared->SetLastError(VE_CANNOT_ACCESS_MIC_VOL, kTraceInfo, |
| "Init() failed to check microphone availability, trying to " |
| "initialize microphone anyway"); |
| } |
| else if (!available) |
| { |
| _shared->SetLastError(VE_CANNOT_ACCESS_MIC_VOL, kTraceInfo, |
| "Init() microphone not available, trying to initialize " |
| "microphone anyway"); |
| } |
| if (_shared->audio_device()->InitMicrophone() != 0) |
| { |
| _shared->SetLastError(VE_CANNOT_ACCESS_MIC_VOL, kTraceInfo, |
| "Init() failed to initialize the microphone"); |
| } |
| |
| // Set number of channels |
| if (_shared->audio_device()->StereoPlayoutIsAvailable(&available) != 0) { |
| _shared->SetLastError(VE_SOUNDCARD_ERROR, kTraceWarning, |
| "Init() failed to query stereo playout mode"); |
| } |
| if (_shared->audio_device()->SetStereoPlayout(available) != 0) |
| { |
| _shared->SetLastError(VE_SOUNDCARD_ERROR, kTraceWarning, |
| "Init() failed to set mono/stereo playout mode"); |
| } |
| |
| // TODO(andrew): These functions don't tell us whether stereo recording |
| // is truly available. We simply set the AudioProcessing input to stereo |
| // here, because we have to wait until receiving the first frame to |
| // determine the actual number of channels anyway. |
| // |
| // These functions may be changed; tracked here: |
| // http://code.google.com/p/webrtc/issues/detail?id=204 |
| _shared->audio_device()->StereoRecordingIsAvailable(&available); |
| if (_shared->audio_device()->SetStereoRecording(available) != 0) |
| { |
| _shared->SetLastError(VE_SOUNDCARD_ERROR, kTraceWarning, |
| "Init() failed to set mono/stereo recording mode"); |
| } |
| |
| if (!audioproc) { |
| audioproc = AudioProcessing::Create(VoEId(_shared->instance_id(), -1)); |
| if (!audioproc) { |
| LOG(LS_ERROR) << "Failed to create AudioProcessing."; |
| _shared->SetLastError(VE_NO_MEMORY); |
| return -1; |
| } |
| } |
| _shared->set_audio_processing(audioproc); |
| |
| // Set the error state for any failures in this block. |
| _shared->SetLastError(VE_APM_ERROR); |
| if (audioproc->echo_cancellation()->set_device_sample_rate_hz(48000)) { |
| LOG_FERR1(LS_ERROR, set_device_sample_rate_hz, 48000); |
| return -1; |
| } |
| // Assume 16 kHz mono until the audio frames are received from the capture |
| // device, at which point this can be updated. |
| if (audioproc->set_sample_rate_hz(16000)) { |
| LOG_FERR1(LS_ERROR, set_sample_rate_hz, 16000); |
| return -1; |
| } |
| if (audioproc->set_num_channels(1, 1) != 0) { |
| LOG_FERR2(LS_ERROR, set_num_channels, 1, 1); |
| return -1; |
| } |
| if (audioproc->set_num_reverse_channels(1) != 0) { |
| LOG_FERR1(LS_ERROR, set_num_reverse_channels, 1); |
| return -1; |
| } |
| |
| // Configure AudioProcessing components. All are disabled by default. |
| if (audioproc->high_pass_filter()->Enable(true) != 0) { |
| LOG_FERR1(LS_ERROR, high_pass_filter()->Enable, true); |
| return -1; |
| } |
| if (audioproc->echo_cancellation()->enable_drift_compensation(false) != 0) { |
| LOG_FERR1(LS_ERROR, enable_drift_compensation, false); |
| return -1; |
| } |
| if (audioproc->noise_suppression()->set_level(kDefaultNsMode) != 0) { |
| LOG_FERR1(LS_ERROR, noise_suppression()->set_level, kDefaultNsMode); |
| return -1; |
| } |
| GainControl* agc = audioproc->gain_control(); |
| if (agc->set_analog_level_limits(kMinVolumeLevel, kMaxVolumeLevel) != 0) { |
| LOG_FERR2(LS_ERROR, agc->set_analog_level_limits, kMinVolumeLevel, |
| kMaxVolumeLevel); |
| return -1; |
| } |
| if (agc->set_mode(kDefaultAgcMode) != 0) { |
| LOG_FERR1(LS_ERROR, agc->set_mode, kDefaultAgcMode); |
| return -1; |
| } |
| if (agc->Enable(kDefaultAgcState) != 0) { |
| LOG_FERR1(LS_ERROR, agc->Enable, kDefaultAgcState); |
| return -1; |
| } |
| _shared->SetLastError(0); // Clear error state. |
| |
| #ifdef WEBRTC_VOICE_ENGINE_AGC |
| bool agc_enabled = agc->mode() == GainControl::kAdaptiveAnalog && |
| agc->is_enabled(); |
| if (_shared->audio_device()->SetAGC(agc_enabled) != 0) { |
| LOG_FERR1(LS_ERROR, audio_device()->SetAGC, agc_enabled); |
| _shared->SetLastError(VE_AUDIO_DEVICE_MODULE_ERROR); |
| // TODO(ajm): No error return here due to |
| // https://code.google.com/p/webrtc/issues/detail?id=1464 |
| } |
| #endif |
| |
| return _shared->statistics().SetInitialized(); |
| } |
| |
| int VoEBaseImpl::Terminate() |
| { |
| WEBRTC_TRACE(kTraceApiCall, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "Terminate()"); |
| CriticalSectionScoped cs(_shared->crit_sec()); |
| return TerminateInternal(); |
| } |
| |
| int VoEBaseImpl::MaxNumOfChannels() |
| { |
| WEBRTC_TRACE(kTraceApiCall, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "MaxNumOfChannels()"); |
| int32_t maxNumOfChannels = |
| _shared->channel_manager().MaxNumOfChannels(); |
| WEBRTC_TRACE(kTraceStateInfo, kTraceVoice, |
| VoEId(_shared->instance_id(), -1), |
| "MaxNumOfChannels() => %d", maxNumOfChannels); |
| return (maxNumOfChannels); |
| } |
| |
| int VoEBaseImpl::CreateChannel() |
| { |
| WEBRTC_TRACE(kTraceApiCall, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "CreateChannel()"); |
| CriticalSectionScoped cs(_shared->crit_sec()); |
| |
| if (!_shared->statistics().Initialized()) |
| { |
| _shared->SetLastError(VE_NOT_INITED, kTraceError); |
| return -1; |
| } |
| |
| int32_t channelId = -1; |
| |
| if (!_shared->channel_manager().CreateChannel(channelId)) |
| { |
| _shared->SetLastError(VE_CHANNEL_NOT_CREATED, kTraceError, |
| "CreateChannel() failed to allocate memory for channel"); |
| return -1; |
| } |
| |
| bool destroyChannel(false); |
| { |
| voe::ScopedChannel sc(_shared->channel_manager(), channelId); |
| voe::Channel* channelPtr = sc.ChannelPtr(); |
| if (channelPtr == NULL) |
| { |
| _shared->SetLastError(VE_CHANNEL_NOT_CREATED, kTraceError, |
| "CreateChannel() failed to allocate memory for channel"); |
| return -1; |
| } |
| else if (channelPtr->SetEngineInformation(_shared->statistics(), |
| *_shared->output_mixer(), |
| *_shared->transmit_mixer(), |
| *_shared->process_thread(), |
| *_shared->audio_device(), |
| _voiceEngineObserverPtr, |
| &_callbackCritSect) != 0) |
| { |
| destroyChannel = true; |
| _shared->SetLastError(VE_CHANNEL_NOT_CREATED, kTraceError, |
| "CreateChannel() failed to associate engine and channel." |
| " Destroying channel."); |
| } |
| else if (channelPtr->Init() != 0) |
| { |
| destroyChannel = true; |
| _shared->SetLastError(VE_CHANNEL_NOT_CREATED, kTraceError, |
| "CreateChannel() failed to initialize channel. Destroying" |
| " channel."); |
| } |
| } |
| if (destroyChannel) |
| { |
| _shared->channel_manager().DestroyChannel(channelId); |
| return -1; |
| } |
| WEBRTC_TRACE(kTraceStateInfo, kTraceVoice, |
| VoEId(_shared->instance_id(), -1), |
| "CreateChannel() => %d", channelId); |
| return channelId; |
| } |
| |
| int VoEBaseImpl::DeleteChannel(int channel) |
| { |
| WEBRTC_TRACE(kTraceApiCall, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "DeleteChannel(channel=%d)", channel); |
| CriticalSectionScoped cs(_shared->crit_sec()); |
| |
| if (!_shared->statistics().Initialized()) |
| { |
| _shared->SetLastError(VE_NOT_INITED, kTraceError); |
| return -1; |
| } |
| |
| { |
| voe::ScopedChannel sc(_shared->channel_manager(), channel); |
| voe::Channel* channelPtr = sc.ChannelPtr(); |
| if (channelPtr == NULL) |
| { |
| _shared->SetLastError(VE_CHANNEL_NOT_VALID, kTraceError, |
| "DeleteChannel() failed to locate channel"); |
| return -1; |
| } |
| } |
| |
| if (_shared->channel_manager().DestroyChannel(channel) != 0) |
| { |
| _shared->SetLastError(VE_CHANNEL_NOT_VALID, kTraceError, |
| "DeleteChannel() failed to destroy channel"); |
| return -1; |
| } |
| |
| if (StopSend() != 0) |
| { |
| return -1; |
| } |
| |
| if (StopPlayout() != 0) |
| { |
| return -1; |
| } |
| |
| return 0; |
| } |
| |
| int VoEBaseImpl::StartReceive(int channel) |
| { |
| WEBRTC_TRACE(kTraceApiCall, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "StartReceive(channel=%d)", channel); |
| CriticalSectionScoped cs(_shared->crit_sec()); |
| if (!_shared->statistics().Initialized()) |
| { |
| _shared->SetLastError(VE_NOT_INITED, kTraceError); |
| return -1; |
| } |
| voe::ScopedChannel sc(_shared->channel_manager(), channel); |
| voe::Channel* channelPtr = sc.ChannelPtr(); |
| if (channelPtr == NULL) |
| { |
| _shared->SetLastError(VE_CHANNEL_NOT_VALID, kTraceError, |
| "StartReceive() failed to locate channel"); |
| return -1; |
| } |
| return channelPtr->StartReceiving(); |
| } |
| |
| int VoEBaseImpl::StopReceive(int channel) |
| { |
| WEBRTC_TRACE(kTraceApiCall, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "StopListen(channel=%d)", channel); |
| CriticalSectionScoped cs(_shared->crit_sec()); |
| if (!_shared->statistics().Initialized()) |
| { |
| _shared->SetLastError(VE_NOT_INITED, kTraceError); |
| return -1; |
| } |
| voe::ScopedChannel sc(_shared->channel_manager(), channel); |
| voe::Channel* channelPtr = sc.ChannelPtr(); |
| if (channelPtr == NULL) |
| { |
| _shared->SetLastError(VE_CHANNEL_NOT_VALID, kTraceError, |
| "SetLocalReceiver() failed to locate channel"); |
| return -1; |
| } |
| return channelPtr->StopReceiving(); |
| } |
| |
| int VoEBaseImpl::StartPlayout(int channel) |
| { |
| WEBRTC_TRACE(kTraceApiCall, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "StartPlayout(channel=%d)", channel); |
| CriticalSectionScoped cs(_shared->crit_sec()); |
| if (!_shared->statistics().Initialized()) |
| { |
| _shared->SetLastError(VE_NOT_INITED, kTraceError); |
| return -1; |
| } |
| voe::ScopedChannel sc(_shared->channel_manager(), channel); |
| voe::Channel* channelPtr = sc.ChannelPtr(); |
| if (channelPtr == NULL) |
| { |
| _shared->SetLastError(VE_CHANNEL_NOT_VALID, kTraceError, |
| "StartPlayout() failed to locate channel"); |
| return -1; |
| } |
| if (channelPtr->Playing()) |
| { |
| return 0; |
| } |
| if (StartPlayout() != 0) |
| { |
| _shared->SetLastError(VE_AUDIO_DEVICE_MODULE_ERROR, kTraceError, |
| "StartPlayout() failed to start playout"); |
| return -1; |
| } |
| return channelPtr->StartPlayout(); |
| } |
| |
| int VoEBaseImpl::StopPlayout(int channel) |
| { |
| WEBRTC_TRACE(kTraceApiCall, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "StopPlayout(channel=%d)", channel); |
| CriticalSectionScoped cs(_shared->crit_sec()); |
| if (!_shared->statistics().Initialized()) |
| { |
| _shared->SetLastError(VE_NOT_INITED, kTraceError); |
| return -1; |
| } |
| voe::ScopedChannel sc(_shared->channel_manager(), channel); |
| voe::Channel* channelPtr = sc.ChannelPtr(); |
| if (channelPtr == NULL) |
| { |
| _shared->SetLastError(VE_CHANNEL_NOT_VALID, kTraceError, |
| "StopPlayout() failed to locate channel"); |
| return -1; |
| } |
| if (channelPtr->StopPlayout() != 0) |
| { |
| WEBRTC_TRACE(kTraceWarning, kTraceVoice, |
| VoEId(_shared->instance_id(), -1), |
| "StopPlayout() failed to stop playout for channel %d", channel); |
| } |
| return StopPlayout(); |
| } |
| |
| int VoEBaseImpl::StartSend(int channel) |
| { |
| WEBRTC_TRACE(kTraceApiCall, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "StartSend(channel=%d)", channel); |
| CriticalSectionScoped cs(_shared->crit_sec()); |
| if (!_shared->statistics().Initialized()) |
| { |
| _shared->SetLastError(VE_NOT_INITED, kTraceError); |
| return -1; |
| } |
| voe::ScopedChannel sc(_shared->channel_manager(), channel); |
| voe::Channel* channelPtr = sc.ChannelPtr(); |
| if (channelPtr == NULL) |
| { |
| _shared->SetLastError(VE_CHANNEL_NOT_VALID, kTraceError, |
| "StartSend() failed to locate channel"); |
| return -1; |
| } |
| if (channelPtr->Sending()) |
| { |
| return 0; |
| } |
| if (StartSend() != 0) |
| { |
| _shared->SetLastError(VE_AUDIO_DEVICE_MODULE_ERROR, kTraceError, |
| "StartSend() failed to start recording"); |
| return -1; |
| } |
| return channelPtr->StartSend(); |
| } |
| |
| int VoEBaseImpl::StopSend(int channel) |
| { |
| WEBRTC_TRACE(kTraceApiCall, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "StopSend(channel=%d)", channel); |
| CriticalSectionScoped cs(_shared->crit_sec()); |
| if (!_shared->statistics().Initialized()) |
| { |
| _shared->SetLastError(VE_NOT_INITED, kTraceError); |
| return -1; |
| } |
| voe::ScopedChannel sc(_shared->channel_manager(), channel); |
| voe::Channel* channelPtr = sc.ChannelPtr(); |
| if (channelPtr == NULL) |
| { |
| _shared->SetLastError(VE_CHANNEL_NOT_VALID, kTraceError, |
| "StopSend() failed to locate channel"); |
| return -1; |
| } |
| if (channelPtr->StopSend() != 0) |
| { |
| WEBRTC_TRACE(kTraceWarning, kTraceVoice, |
| VoEId(_shared->instance_id(), -1), |
| "StopSend() failed to stop sending for channel %d", channel); |
| } |
| return StopSend(); |
| } |
| |
| int VoEBaseImpl::GetVersion(char version[1024]) |
| { |
| WEBRTC_TRACE(kTraceApiCall, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "GetVersion(version=?)"); |
| assert(kVoiceEngineVersionMaxMessageSize == 1024); |
| |
| if (version == NULL) |
| { |
| _shared->SetLastError(VE_INVALID_ARGUMENT, kTraceError); |
| return (-1); |
| } |
| |
| char versionBuf[kVoiceEngineVersionMaxMessageSize]; |
| char* versionPtr = versionBuf; |
| |
| int32_t len = 0; |
| int32_t accLen = 0; |
| |
| len = AddVoEVersion(versionPtr); |
| if (len == -1) |
| { |
| return -1; |
| } |
| versionPtr += len; |
| accLen += len; |
| assert(accLen < kVoiceEngineVersionMaxMessageSize); |
| |
| len = AddBuildInfo(versionPtr); |
| if (len == -1) |
| { |
| return -1; |
| } |
| versionPtr += len; |
| accLen += len; |
| assert(accLen < kVoiceEngineVersionMaxMessageSize); |
| |
| #ifdef WEBRTC_EXTERNAL_TRANSPORT |
| len = AddExternalTransportBuild(versionPtr); |
| if (len == -1) |
| { |
| return -1; |
| } |
| versionPtr += len; |
| accLen += len; |
| assert(accLen < kVoiceEngineVersionMaxMessageSize); |
| #endif |
| #ifdef WEBRTC_VOE_EXTERNAL_REC_AND_PLAYOUT |
| len = AddExternalRecAndPlayoutBuild(versionPtr); |
| if (len == -1) |
| { |
| return -1; |
| } |
| versionPtr += len; |
| accLen += len; |
| assert(accLen < kVoiceEngineVersionMaxMessageSize); |
| #endif |
| |
| memcpy(version, versionBuf, accLen); |
| version[accLen] = '\0'; |
| |
| // to avoid the truncation in the trace, split the string into parts |
| char partOfVersion[256]; |
| WEBRTC_TRACE(kTraceStateInfo, kTraceVoice, |
| VoEId(_shared->instance_id(), -1), "GetVersion() =>"); |
| for (int partStart = 0; partStart < accLen;) |
| { |
| memset(partOfVersion, 0, sizeof(partOfVersion)); |
| int partEnd = partStart + 180; |
| while (version[partEnd] != '\n' && version[partEnd] != '\0') |
| { |
| partEnd--; |
| } |
| if (partEnd < accLen) |
| { |
| memcpy(partOfVersion, &version[partStart], partEnd - partStart); |
| } |
| else |
| { |
| memcpy(partOfVersion, &version[partStart], accLen - partStart); |
| } |
| partStart = partEnd; |
| WEBRTC_TRACE(kTraceStateInfo, kTraceVoice, |
| VoEId(_shared->instance_id(), -1), "%s", partOfVersion); |
| } |
| |
| return 0; |
| } |
| |
| int32_t VoEBaseImpl::AddBuildInfo(char* str) const |
| { |
| return sprintf(str, "Build: svn:%s %s\n", WEBRTC_SVNREVISION, BUILDINFO); |
| } |
| |
| int32_t VoEBaseImpl::AddVoEVersion(char* str) const |
| { |
| return sprintf(str, "VoiceEngine 4.1.0\n"); |
| } |
| |
| #ifdef WEBRTC_EXTERNAL_TRANSPORT |
| int32_t VoEBaseImpl::AddExternalTransportBuild(char* str) const |
| { |
| return sprintf(str, "External transport build\n"); |
| } |
| #endif |
| |
| #ifdef WEBRTC_VOE_EXTERNAL_REC_AND_PLAYOUT |
| int32_t VoEBaseImpl::AddExternalRecAndPlayoutBuild(char* str) const |
| { |
| return sprintf(str, "External recording and playout build\n"); |
| } |
| #endif |
| |
| int VoEBaseImpl::LastError() |
| { |
| WEBRTC_TRACE(kTraceApiCall, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "LastError()"); |
| return (_shared->statistics().LastError()); |
| } |
| |
| |
| int VoEBaseImpl::SetNetEQPlayoutMode(int channel, NetEqModes mode) |
| { |
| WEBRTC_TRACE(kTraceApiCall, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "SetNetEQPlayoutMode(channel=%i, mode=%i)", channel, mode); |
| if (!_shared->statistics().Initialized()) |
| { |
| _shared->SetLastError(VE_NOT_INITED, kTraceError); |
| return -1; |
| } |
| voe::ScopedChannel sc(_shared->channel_manager(), channel); |
| voe::Channel* channelPtr = sc.ChannelPtr(); |
| if (channelPtr == NULL) |
| { |
| _shared->SetLastError(VE_CHANNEL_NOT_VALID, kTraceError, |
| "SetNetEQPlayoutMode() failed to locate channel"); |
| return -1; |
| } |
| return channelPtr->SetNetEQPlayoutMode(mode); |
| } |
| |
| int VoEBaseImpl::GetNetEQPlayoutMode(int channel, NetEqModes& mode) |
| { |
| WEBRTC_TRACE(kTraceApiCall, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "GetNetEQPlayoutMode(channel=%i, mode=?)", channel); |
| if (!_shared->statistics().Initialized()) |
| { |
| _shared->SetLastError(VE_NOT_INITED, kTraceError); |
| return -1; |
| } |
| voe::ScopedChannel sc(_shared->channel_manager(), channel); |
| voe::Channel* channelPtr = sc.ChannelPtr(); |
| if (channelPtr == NULL) |
| { |
| _shared->SetLastError(VE_CHANNEL_NOT_VALID, kTraceError, |
| "GetNetEQPlayoutMode() failed to locate channel"); |
| return -1; |
| } |
| return channelPtr->GetNetEQPlayoutMode(mode); |
| } |
| |
| int VoEBaseImpl::SetOnHoldStatus(int channel, bool enable, OnHoldModes mode) |
| { |
| WEBRTC_TRACE(kTraceApiCall, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "SetOnHoldStatus(channel=%d, enable=%d, mode=%d)", channel, |
| enable, mode); |
| if (!_shared->statistics().Initialized()) |
| { |
| _shared->SetLastError(VE_NOT_INITED, kTraceError); |
| return -1; |
| } |
| voe::ScopedChannel sc(_shared->channel_manager(), channel); |
| voe::Channel* channelPtr = sc.ChannelPtr(); |
| if (channelPtr == NULL) |
| { |
| _shared->SetLastError(VE_CHANNEL_NOT_VALID, kTraceError, |
| "SetOnHoldStatus() failed to locate channel"); |
| return -1; |
| } |
| return channelPtr->SetOnHoldStatus(enable, mode); |
| } |
| |
| int VoEBaseImpl::GetOnHoldStatus(int channel, bool& enabled, OnHoldModes& mode) |
| { |
| WEBRTC_TRACE(kTraceApiCall, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "GetOnHoldStatus(channel=%d, enabled=?, mode=?)", channel); |
| if (!_shared->statistics().Initialized()) |
| { |
| _shared->SetLastError(VE_NOT_INITED, kTraceError); |
| return -1; |
| } |
| voe::ScopedChannel sc(_shared->channel_manager(), channel); |
| voe::Channel* channelPtr = sc.ChannelPtr(); |
| if (channelPtr == NULL) |
| { |
| _shared->SetLastError(VE_CHANNEL_NOT_VALID, kTraceError, |
| "GetOnHoldStatus() failed to locate channel"); |
| return -1; |
| } |
| return channelPtr->GetOnHoldStatus(enabled, mode); |
| } |
| |
| int32_t VoEBaseImpl::StartPlayout() |
| { |
| WEBRTC_TRACE(kTraceInfo, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "VoEBaseImpl::StartPlayout()"); |
| if (_shared->audio_device()->Playing()) |
| { |
| return 0; |
| } |
| if (!_shared->ext_playout()) |
| { |
| if (_shared->audio_device()->InitPlayout() != 0) |
| { |
| WEBRTC_TRACE(kTraceError, kTraceVoice, |
| VoEId(_shared->instance_id(), -1), |
| "StartPlayout() failed to initialize playout"); |
| return -1; |
| } |
| if (_shared->audio_device()->StartPlayout() != 0) |
| { |
| WEBRTC_TRACE(kTraceError, kTraceVoice, |
| VoEId(_shared->instance_id(), -1), |
| "StartPlayout() failed to start playout"); |
| return -1; |
| } |
| } |
| return 0; |
| } |
| |
| int32_t VoEBaseImpl::StopPlayout() |
| { |
| WEBRTC_TRACE(kTraceInfo, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "VoEBaseImpl::StopPlayout()"); |
| |
| int32_t numOfChannels = _shared->channel_manager().NumOfChannels(); |
| if (numOfChannels <= 0) |
| { |
| return 0; |
| } |
| |
| uint16_t nChannelsPlaying(0); |
| int32_t* channelsArray = new int32_t[numOfChannels]; |
| |
| // Get number of playing channels |
| _shared->channel_manager().GetChannelIds(channelsArray, numOfChannels); |
| for (int i = 0; i < numOfChannels; i++) |
| { |
| voe::ScopedChannel sc(_shared->channel_manager(), channelsArray[i]); |
| voe::Channel* chPtr = sc.ChannelPtr(); |
| if (chPtr) |
| { |
| if (chPtr->Playing()) |
| { |
| nChannelsPlaying++; |
| } |
| } |
| } |
| delete[] channelsArray; |
| |
| // Stop audio-device playing if no channel is playing out |
| if (nChannelsPlaying == 0) |
| { |
| if (_shared->audio_device()->StopPlayout() != 0) |
| { |
| _shared->SetLastError(VE_CANNOT_STOP_PLAYOUT, kTraceError, |
| "StopPlayout() failed to stop playout"); |
| return -1; |
| } |
| } |
| return 0; |
| } |
| |
| int32_t VoEBaseImpl::StartSend() |
| { |
| WEBRTC_TRACE(kTraceInfo, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "VoEBaseImpl::StartSend()"); |
| if (_shared->audio_device()->Recording()) |
| { |
| return 0; |
| } |
| if (!_shared->ext_recording()) |
| { |
| if (_shared->audio_device()->InitRecording() != 0) |
| { |
| WEBRTC_TRACE(kTraceError, kTraceVoice, |
| VoEId(_shared->instance_id(), -1), |
| "StartSend() failed to initialize recording"); |
| return -1; |
| } |
| if (_shared->audio_device()->StartRecording() != 0) |
| { |
| WEBRTC_TRACE(kTraceError, kTraceVoice, |
| VoEId(_shared->instance_id(), -1), |
| "StartSend() failed to start recording"); |
| return -1; |
| } |
| } |
| |
| return 0; |
| } |
| |
| int32_t VoEBaseImpl::StopSend() |
| { |
| WEBRTC_TRACE(kTraceInfo, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "VoEBaseImpl::StopSend()"); |
| |
| if (_shared->NumOfSendingChannels() == 0 && |
| !_shared->transmit_mixer()->IsRecordingMic()) |
| { |
| // Stop audio-device recording if no channel is recording |
| if (_shared->audio_device()->StopRecording() != 0) |
| { |
| _shared->SetLastError(VE_CANNOT_STOP_RECORDING, kTraceError, |
| "StopSend() failed to stop recording"); |
| return -1; |
| } |
| _shared->transmit_mixer()->StopSend(); |
| } |
| |
| return 0; |
| } |
| |
| int32_t VoEBaseImpl::TerminateInternal() |
| { |
| WEBRTC_TRACE(kTraceInfo, kTraceVoice, VoEId(_shared->instance_id(), -1), |
| "VoEBaseImpl::TerminateInternal()"); |
| |
| // Delete any remaining channel objects |
| int32_t numOfChannels = _shared->channel_manager().NumOfChannels(); |
| if (numOfChannels > 0) |
| { |
| int32_t* channelsArray = new int32_t[numOfChannels]; |
| _shared->channel_manager().GetChannelIds(channelsArray, numOfChannels); |
| for (int i = 0; i < numOfChannels; i++) |
| { |
| DeleteChannel(channelsArray[i]); |
| } |
| delete[] channelsArray; |
| } |
| |
| if (_shared->process_thread()) |
| { |
| if (_shared->audio_device()) |
| { |
| if (_shared->process_thread()-> |
| DeRegisterModule(_shared->audio_device()) != 0) |
| { |
| _shared->SetLastError(VE_THREAD_ERROR, kTraceError, |
| "TerminateInternal() failed to deregister ADM"); |
| } |
| } |
| if (_shared->process_thread()->Stop() != 0) |
| { |
| _shared->SetLastError(VE_THREAD_ERROR, kTraceError, |
| "TerminateInternal() failed to stop module process thread"); |
| } |
| } |
| |
| if (_shared->audio_device()) |
| { |
| if (_shared->audio_device()->StopPlayout() != 0) |
| { |
| _shared->SetLastError(VE_SOUNDCARD_ERROR, kTraceWarning, |
| "TerminateInternal() failed to stop playout"); |
| } |
| if (_shared->audio_device()->StopRecording() != 0) |
| { |
| _shared->SetLastError(VE_SOUNDCARD_ERROR, kTraceWarning, |
| "TerminateInternal() failed to stop recording"); |
| } |
| if (_shared->audio_device()->RegisterEventObserver(NULL) != 0) { |
| _shared->SetLastError(VE_AUDIO_DEVICE_MODULE_ERROR, kTraceWarning, |
| "TerminateInternal() failed to de-register event observer " |
| "for the ADM"); |
| } |
| if (_shared->audio_device()->RegisterAudioCallback(NULL) != 0) { |
| _shared->SetLastError(VE_AUDIO_DEVICE_MODULE_ERROR, kTraceWarning, |
| "TerminateInternal() failed to de-register audio callback " |
| "for the ADM"); |
| } |
| if (_shared->audio_device()->Terminate() != 0) |
| { |
| _shared->SetLastError(VE_AUDIO_DEVICE_MODULE_ERROR, kTraceError, |
| "TerminateInternal() failed to terminate the ADM"); |
| } |
| _shared->set_audio_device(NULL); |
| } |
| |
| if (_shared->audio_processing()) { |
| _shared->set_audio_processing(NULL); |
| } |
| |
| return _shared->statistics().SetUnInitialized(); |
| } |
| |
| } // namespace webrtc |