/* * Copyright (c) 2012 The WebRTC project authors. All Rights Reserved. * * Use of this source code is governed by a BSD-style license * that can be found in the LICENSE file in the root of the source * tree. An additional intellectual property rights grant can be found * in the file PATENTS. All contributing project authors may * be found in the AUTHORS file in the root of the source tree. */ #include "webrtc/modules/audio_processing/echo_control_mobile_impl.h" #include #include #include "webrtc/modules/audio_processing/aecm/echo_control_mobile.h" #include "webrtc/modules/audio_processing/audio_buffer.h" #include "webrtc/system_wrappers/include/logging.h" namespace webrtc { typedef void Handle; namespace { int16_t MapSetting(EchoControlMobile::RoutingMode mode) { switch (mode) { case EchoControlMobile::kQuietEarpieceOrHeadset: return 0; case EchoControlMobile::kEarpiece: return 1; case EchoControlMobile::kLoudEarpiece: return 2; case EchoControlMobile::kSpeakerphone: return 3; case EchoControlMobile::kLoudSpeakerphone: return 4; } assert(false); return -1; } AudioProcessing::Error MapError(int err) { switch (err) { case AECM_UNSUPPORTED_FUNCTION_ERROR: return AudioProcessing::kUnsupportedFunctionError; case AECM_NULL_POINTER_ERROR: return AudioProcessing::kNullPointerError; case AECM_BAD_PARAMETER_ERROR: return AudioProcessing::kBadParameterError; case AECM_BAD_PARAMETER_WARNING: return AudioProcessing::kBadStreamParameterWarning; default: // AECM_UNSPECIFIED_ERROR // AECM_UNINITIALIZED_ERROR return AudioProcessing::kUnspecifiedError; } } // Maximum length that a frame of samples can have. static const size_t kMaxAllowedValuesOfSamplesPerFrame = 160; // Maximum number of frames to buffer in the render queue. // TODO(peah): Decrease this once we properly handle hugely unbalanced // reverse and forward call numbers. static const size_t kMaxNumFramesToBuffer = 100; } // namespace size_t EchoControlMobile::echo_path_size_bytes() { return WebRtcAecm_echo_path_size_bytes(); } EchoControlMobileImpl::EchoControlMobileImpl(const AudioProcessing* apm, rtc::CriticalSection* crit_render, rtc::CriticalSection* crit_capture) : ProcessingComponent(), apm_(apm), crit_render_(crit_render), crit_capture_(crit_capture), routing_mode_(kSpeakerphone), comfort_noise_enabled_(true), external_echo_path_(NULL), render_queue_element_max_size_(0) { RTC_DCHECK(apm); RTC_DCHECK(crit_render); RTC_DCHECK(crit_capture); } EchoControlMobileImpl::~EchoControlMobileImpl() { if (external_echo_path_ != NULL) { delete [] external_echo_path_; external_echo_path_ = NULL; } } int EchoControlMobileImpl::ProcessRenderAudio(const AudioBuffer* audio) { rtc::CritScope cs_render(crit_render_); if (!is_component_enabled()) { return AudioProcessing::kNoError; } assert(audio->num_frames_per_band() <= 160); assert(audio->num_channels() == apm_->num_reverse_channels()); int err = AudioProcessing::kNoError; // The ordering convention must be followed to pass to the correct AECM. size_t handle_index = 0; render_queue_buffer_.clear(); for (size_t i = 0; i < apm_->num_output_channels(); i++) { for (size_t j = 0; j < audio->num_channels(); j++) { Handle* my_handle = static_cast(handle(handle_index)); err = WebRtcAecm_GetBufferFarendError( my_handle, audio->split_bands_const(j)[kBand0To8kHz], audio->num_frames_per_band()); if (err != AudioProcessing::kNoError) return MapError(err); // TODO(ajm): warning possible?); // Buffer the samples in the render queue. render_queue_buffer_.insert(render_queue_buffer_.end(), audio->split_bands_const(j)[kBand0To8kHz], (audio->split_bands_const(j)[kBand0To8kHz] + audio->num_frames_per_band())); handle_index++; } } // Insert the samples into the queue. if (!render_signal_queue_->Insert(&render_queue_buffer_)) { // The data queue is full and needs to be emptied. ReadQueuedRenderData(); // Retry the insert (should always work). RTC_DCHECK_EQ(render_signal_queue_->Insert(&render_queue_buffer_), true); } return AudioProcessing::kNoError; } // Read chunks of data that were received and queued on the render side from // a queue. All the data chunks are buffered into the farend signal of the AEC. void EchoControlMobileImpl::ReadQueuedRenderData() { rtc::CritScope cs_capture(crit_capture_); if (!is_component_enabled()) { return; } while (render_signal_queue_->Remove(&capture_queue_buffer_)) { size_t handle_index = 0; size_t buffer_index = 0; const size_t num_frames_per_band = capture_queue_buffer_.size() / (apm_->num_output_channels() * apm_->num_reverse_channels()); for (size_t i = 0; i < apm_->num_output_channels(); i++) { for (size_t j = 0; j < apm_->num_reverse_channels(); j++) { Handle* my_handle = static_cast(handle(handle_index)); WebRtcAecm_BufferFarend(my_handle, &capture_queue_buffer_[buffer_index], num_frames_per_band); buffer_index += num_frames_per_band; handle_index++; } } } } int EchoControlMobileImpl::ProcessCaptureAudio(AudioBuffer* audio) { rtc::CritScope cs_capture(crit_capture_); if (!is_component_enabled()) { return AudioProcessing::kNoError; } if (!apm_->was_stream_delay_set()) { return AudioProcessing::kStreamParameterNotSetError; } assert(audio->num_frames_per_band() <= 160); assert(audio->num_channels() == apm_->num_output_channels()); int err = AudioProcessing::kNoError; // The ordering convention must be followed to pass to the correct AECM. size_t handle_index = 0; for (size_t i = 0; i < audio->num_channels(); i++) { // TODO(ajm): improve how this works, possibly inside AECM. // This is kind of hacked up. const int16_t* noisy = audio->low_pass_reference(i); const int16_t* clean = audio->split_bands_const(i)[kBand0To8kHz]; if (noisy == NULL) { noisy = clean; clean = NULL; } for (size_t j = 0; j < apm_->num_reverse_channels(); j++) { Handle* my_handle = static_cast(handle(handle_index)); err = WebRtcAecm_Process( my_handle, noisy, clean, audio->split_bands(i)[kBand0To8kHz], audio->num_frames_per_band(), apm_->stream_delay_ms()); if (err != AudioProcessing::kNoError) return MapError(err); handle_index++; } } return AudioProcessing::kNoError; } int EchoControlMobileImpl::Enable(bool enable) { // Ensure AEC and AECM are not both enabled. rtc::CritScope cs_render(crit_render_); rtc::CritScope cs_capture(crit_capture_); // The is_enabled call is safe from a deadlock perspective // as both locks are allready held in the correct order. if (enable && apm_->echo_cancellation()->is_enabled()) { return AudioProcessing::kBadParameterError; } return EnableComponent(enable); } bool EchoControlMobileImpl::is_enabled() const { rtc::CritScope cs(crit_capture_); return is_component_enabled(); } int EchoControlMobileImpl::set_routing_mode(RoutingMode mode) { if (MapSetting(mode) == -1) { return AudioProcessing::kBadParameterError; } { rtc::CritScope cs(crit_capture_); routing_mode_ = mode; } return Configure(); } EchoControlMobile::RoutingMode EchoControlMobileImpl::routing_mode() const { rtc::CritScope cs(crit_capture_); return routing_mode_; } int EchoControlMobileImpl::enable_comfort_noise(bool enable) { { rtc::CritScope cs(crit_capture_); comfort_noise_enabled_ = enable; } return Configure(); } bool EchoControlMobileImpl::is_comfort_noise_enabled() const { rtc::CritScope cs(crit_capture_); return comfort_noise_enabled_; } int EchoControlMobileImpl::SetEchoPath(const void* echo_path, size_t size_bytes) { { rtc::CritScope cs_render(crit_render_); rtc::CritScope cs_capture(crit_capture_); if (echo_path == NULL) { return AudioProcessing::kNullPointerError; } if (size_bytes != echo_path_size_bytes()) { // Size mismatch return AudioProcessing::kBadParameterError; } if (external_echo_path_ == NULL) { external_echo_path_ = new unsigned char[size_bytes]; } memcpy(external_echo_path_, echo_path, size_bytes); } return Initialize(); } int EchoControlMobileImpl::GetEchoPath(void* echo_path, size_t size_bytes) const { rtc::CritScope cs(crit_capture_); if (echo_path == NULL) { return AudioProcessing::kNullPointerError; } if (size_bytes != echo_path_size_bytes()) { // Size mismatch return AudioProcessing::kBadParameterError; } if (!is_component_enabled()) { return AudioProcessing::kNotEnabledError; } // Get the echo path from the first channel Handle* my_handle = static_cast(handle(0)); int32_t err = WebRtcAecm_GetEchoPath(my_handle, echo_path, size_bytes); if (err != 0) return MapError(err); return AudioProcessing::kNoError; } int EchoControlMobileImpl::Initialize() { { rtc::CritScope cs_capture(crit_capture_); if (!is_component_enabled()) { return AudioProcessing::kNoError; } } if (apm_->proc_sample_rate_hz() > AudioProcessing::kSampleRate16kHz) { LOG(LS_ERROR) << "AECM only supports 16 kHz or lower sample rates"; return AudioProcessing::kBadSampleRateError; } int err = ProcessingComponent::Initialize(); if (err != AudioProcessing::kNoError) { return err; } AllocateRenderQueue(); return AudioProcessing::kNoError; } void EchoControlMobileImpl::AllocateRenderQueue() { const size_t new_render_queue_element_max_size = std::max( static_cast(1), kMaxAllowedValuesOfSamplesPerFrame * num_handles_required()); rtc::CritScope cs_render(crit_render_); rtc::CritScope cs_capture(crit_capture_); // Reallocate the queue if the queue item size is too small to fit the // data to put in the queue. if (render_queue_element_max_size_ < new_render_queue_element_max_size) { render_queue_element_max_size_ = new_render_queue_element_max_size; std::vector template_queue_element(render_queue_element_max_size_); render_signal_queue_.reset( new SwapQueue, RenderQueueItemVerifier>( kMaxNumFramesToBuffer, template_queue_element, RenderQueueItemVerifier(render_queue_element_max_size_))); render_queue_buffer_.resize(render_queue_element_max_size_); capture_queue_buffer_.resize(render_queue_element_max_size_); } else { render_signal_queue_->Clear(); } } void* EchoControlMobileImpl::CreateHandle() const { return WebRtcAecm_Create(); } void EchoControlMobileImpl::DestroyHandle(void* handle) const { // This method is only called in a non-concurrent manner during APM // destruction. WebRtcAecm_Free(static_cast(handle)); } int EchoControlMobileImpl::InitializeHandle(void* handle) const { rtc::CritScope cs_render(crit_render_); rtc::CritScope cs_capture(crit_capture_); assert(handle != NULL); Handle* my_handle = static_cast(handle); if (WebRtcAecm_Init(my_handle, apm_->proc_sample_rate_hz()) != 0) { return GetHandleError(my_handle); } if (external_echo_path_ != NULL) { if (WebRtcAecm_InitEchoPath(my_handle, external_echo_path_, echo_path_size_bytes()) != 0) { return GetHandleError(my_handle); } } return AudioProcessing::kNoError; } int EchoControlMobileImpl::ConfigureHandle(void* handle) const { rtc::CritScope cs_render(crit_render_); rtc::CritScope cs_capture(crit_capture_); AecmConfig config; config.cngMode = comfort_noise_enabled_; config.echoMode = MapSetting(routing_mode_); return WebRtcAecm_set_config(static_cast(handle), config); } size_t EchoControlMobileImpl::num_handles_required() const { // Not locked as it only relies on APM public API which is threadsafe. return apm_->num_output_channels() * apm_->num_reverse_channels(); } int EchoControlMobileImpl::GetHandleError(void* handle) const { // Not locked as it does not rely on anything in the state. assert(handle != NULL); return AudioProcessing::kUnspecifiedError; } } // namespace webrtc