| /* |
| * Copyright (c) 2012 The WebRTC project authors. All Rights Reserved. |
| * |
| * Use of this source code is governed by a BSD-style license |
| * that can be found in the LICENSE file in the root of the source |
| * tree. An additional intellectual property rights grant can be found |
| * in the file PATENTS. All contributing project authors may |
| * be found in the AUTHORS file in the root of the source tree. |
| */ |
| |
| #include "webrtc/modules/audio_coding/codecs/opus/interface/opus_interface.h" |
| |
| #include <stdlib.h> |
| #include <string.h> |
| |
| #include "opus.h" |
| |
| #include "webrtc/common_audio/signal_processing/resample_by_2_internal.h" |
| #include "webrtc/common_audio/signal_processing/include/signal_processing_library.h" |
| |
| enum { |
| /* Maximum supported frame size in WebRTC is 60 ms. */ |
| kWebRtcOpusMaxEncodeFrameSizeMs = 60, |
| |
| /* The format allows up to 120 ms frames. Since we don't control the other |
| * side, we must allow for packets of that size. NetEq is currently limited |
| * to 60 ms on the receive side. */ |
| kWebRtcOpusMaxDecodeFrameSizeMs = 120, |
| |
| /* Maximum sample count per channel is 48 kHz * maximum frame size in |
| * milliseconds. */ |
| kWebRtcOpusMaxFrameSizePerChannel = 48 * kWebRtcOpusMaxDecodeFrameSizeMs, |
| |
| /* Maximum sample count per frame is 48 kHz * maximum frame size in |
| * milliseconds * maximum number of channels. */ |
| kWebRtcOpusMaxFrameSize = kWebRtcOpusMaxFrameSizePerChannel * 2, |
| |
| /* Number of samples in resampler state. */ |
| kWebRtcOpusStateSize = 7, |
| }; |
| |
| struct WebRtcOpusEncInst { |
| OpusEncoder* encoder; |
| }; |
| |
| int16_t WebRtcOpus_EncoderCreate(OpusEncInst** inst, int32_t channels) { |
| OpusEncInst* state; |
| if (inst != NULL) { |
| state = (OpusEncInst*) calloc(1, sizeof(OpusEncInst)); |
| if (state) { |
| int error; |
| /* Default to VoIP application for mono, and AUDIO for stereo. */ |
| int application = |
| (channels == 1) ? OPUS_APPLICATION_VOIP : OPUS_APPLICATION_AUDIO; |
| |
| state->encoder = opus_encoder_create(48000, channels, application, |
| &error); |
| if (error == OPUS_OK && state->encoder != NULL) { |
| *inst = state; |
| return 0; |
| } |
| free(state); |
| } |
| } |
| return -1; |
| } |
| |
| int16_t WebRtcOpus_EncoderFree(OpusEncInst* inst) { |
| if (inst) { |
| opus_encoder_destroy(inst->encoder); |
| free(inst); |
| return 0; |
| } else { |
| return -1; |
| } |
| } |
| |
| int16_t WebRtcOpus_Encode(OpusEncInst* inst, int16_t* audio_in, int16_t samples, |
| int16_t length_encoded_buffer, uint8_t* encoded) { |
| opus_int16* audio = (opus_int16*) audio_in; |
| unsigned char* coded = encoded; |
| int res; |
| |
| if (samples > 48 * kWebRtcOpusMaxEncodeFrameSizeMs) { |
| return -1; |
| } |
| |
| res = opus_encode(inst->encoder, audio, samples, coded, |
| length_encoded_buffer); |
| |
| if (res > 0) { |
| return res; |
| } |
| return -1; |
| } |
| |
| int16_t WebRtcOpus_SetBitRate(OpusEncInst* inst, int32_t rate) { |
| if (inst) { |
| return opus_encoder_ctl(inst->encoder, OPUS_SET_BITRATE(rate)); |
| } else { |
| return -1; |
| } |
| } |
| |
| struct WebRtcOpusDecInst { |
| int16_t state_48_32_left[8]; |
| int16_t state_48_32_right[8]; |
| OpusDecoder* decoder_left; |
| OpusDecoder* decoder_right; |
| int channels; |
| }; |
| |
| int16_t WebRtcOpus_DecoderCreate(OpusDecInst** inst, int channels) { |
| int error_l; |
| int error_r; |
| OpusDecInst* state; |
| |
| if (inst != NULL) { |
| /* Create Opus decoder state. */ |
| state = (OpusDecInst*) calloc(1, sizeof(OpusDecInst)); |
| if (state == NULL) { |
| return -1; |
| } |
| |
| /* Create new memory for left and right channel, always at 48000 Hz. */ |
| state->decoder_left = opus_decoder_create(48000, channels, &error_l); |
| state->decoder_right = opus_decoder_create(48000, channels, &error_r); |
| if (error_l == OPUS_OK && error_r == OPUS_OK && state->decoder_left != NULL |
| && state->decoder_right != NULL) { |
| /* Creation of memory all ok. */ |
| state->channels = channels; |
| *inst = state; |
| return 0; |
| } |
| |
| /* If memory allocation was unsuccessful, free the entire state. */ |
| if (state->decoder_left) { |
| opus_decoder_destroy(state->decoder_left); |
| } |
| if (state->decoder_right) { |
| opus_decoder_destroy(state->decoder_right); |
| } |
| free(state); |
| } |
| return -1; |
| } |
| |
| int16_t WebRtcOpus_DecoderFree(OpusDecInst* inst) { |
| if (inst) { |
| opus_decoder_destroy(inst->decoder_left); |
| opus_decoder_destroy(inst->decoder_right); |
| free(inst); |
| return 0; |
| } else { |
| return -1; |
| } |
| } |
| |
| int WebRtcOpus_DecoderChannels(OpusDecInst* inst) { |
| return inst->channels; |
| } |
| |
| int16_t WebRtcOpus_DecoderInitNew(OpusDecInst* inst) { |
| int error = opus_decoder_ctl(inst->decoder_left, OPUS_RESET_STATE); |
| if (error == OPUS_OK) { |
| memset(inst->state_48_32_left, 0, sizeof(inst->state_48_32_left)); |
| memset(inst->state_48_32_right, 0, sizeof(inst->state_48_32_right)); |
| return 0; |
| } |
| return -1; |
| } |
| |
| int16_t WebRtcOpus_DecoderInit(OpusDecInst* inst) { |
| int error = opus_decoder_ctl(inst->decoder_left, OPUS_RESET_STATE); |
| if (error == OPUS_OK) { |
| memset(inst->state_48_32_left, 0, sizeof(inst->state_48_32_left)); |
| return 0; |
| } |
| return -1; |
| } |
| |
| int16_t WebRtcOpus_DecoderInitSlave(OpusDecInst* inst) { |
| int error = opus_decoder_ctl(inst->decoder_right, OPUS_RESET_STATE); |
| if (error == OPUS_OK) { |
| memset(inst->state_48_32_right, 0, sizeof(inst->state_48_32_right)); |
| return 0; |
| } |
| return -1; |
| } |
| |
| static int DecodeNative(OpusDecoder* inst, const int16_t* encoded, |
| int16_t encoded_bytes, int16_t* decoded, |
| int16_t* audio_type) { |
| unsigned char* coded = (unsigned char*) encoded; |
| opus_int16* audio = (opus_int16*) decoded; |
| |
| int res = opus_decode(inst, coded, encoded_bytes, audio, |
| kWebRtcOpusMaxFrameSizePerChannel, 0); |
| /* TODO(tlegrand): set to DTX for zero-length packets? */ |
| *audio_type = 0; |
| |
| if (res > 0) { |
| return res; |
| } |
| return -1; |
| } |
| |
| /* Resample from 48 to 32 kHz. Length of state is assumed to be |
| * kWebRtcOpusStateSize (7). |
| */ |
| static int WebRtcOpus_Resample48to32(const int16_t* samples_in, int length, |
| int16_t* state, int16_t* samples_out) { |
| int i; |
| int blocks; |
| int16_t output_samples; |
| int32_t buffer32[kWebRtcOpusMaxFrameSize + kWebRtcOpusStateSize]; |
| |
| /* Resample from 48 kHz to 32 kHz. */ |
| for (i = 0; i < kWebRtcOpusStateSize; i++) { |
| buffer32[i] = state[i]; |
| state[i] = samples_in[length - kWebRtcOpusStateSize + i]; |
| } |
| for (i = 0; i < length; i++) { |
| buffer32[kWebRtcOpusStateSize + i] = samples_in[i]; |
| } |
| /* Resampling 3 samples to 2. Function divides the input in |blocks| number |
| * of 3-sample groups, and output is |blocks| number of 2-sample groups. |
| * When this is removed, the compensation in WebRtcOpus_DurationEst should be |
| * removed too. */ |
| blocks = length / 3; |
| WebRtcSpl_Resample48khzTo32khz(buffer32, buffer32, blocks); |
| output_samples = (int16_t) (blocks * 2); |
| WebRtcSpl_VectorBitShiftW32ToW16(samples_out, output_samples, buffer32, 15); |
| |
| return output_samples; |
| } |
| |
| int16_t WebRtcOpus_DecodeNew(OpusDecInst* inst, const uint8_t* encoded, |
| int16_t encoded_bytes, int16_t* decoded, |
| int16_t* audio_type) { |
| /* |buffer16_left| and |buffer_out| are big enough for 120 ms (the largest |
| * Opus packet size) of stereo audio at 48 kHz, while |buffer16_right| only |
| * need to be big enough for maximum size of one of the channels. */ |
| int16_t buffer16_left[kWebRtcOpusMaxFrameSize]; |
| int16_t buffer16_right[kWebRtcOpusMaxFrameSizePerChannel]; |
| int16_t buffer_out[kWebRtcOpusMaxFrameSize]; |
| int16_t* coded = (int16_t*) encoded; |
| int decoded_samples; |
| int resampled_samples; |
| int i; |
| |
| /* If mono case, just do a regular call to the decoder. |
| * If stereo, we need to de-interleave the stereo output in to blocks with |
| * left and right channel. Each block is resampled to 32 kHz, and then |
| * interleaved again. */ |
| |
| /* Decode to temporarily to |buffer16_left|. */ |
| decoded_samples = DecodeNative(inst->decoder_left, coded, encoded_bytes, |
| buffer16_left, audio_type); |
| if (decoded_samples < 0) { |
| return -1; |
| } |
| |
| /* De-interleave if stereo. */ |
| if (inst->channels == 2) { |
| /* The parameter |decoded_samples| holds the number of samples pairs, in |
| * case of stereo. Number of samples in |buffer16_left| equals |
| * |decoded_samples| times 2. */ |
| for (i = 0; i < decoded_samples; i++) { |
| /* Take every second sample, starting at the first sample. */ |
| buffer16_left[i] = buffer16_left[i * 2]; |
| buffer16_right[i] = buffer16_left[i * 2 + 1]; |
| } |
| |
| /* Resample from 48 kHz to 32 kHz for left channel. */ |
| resampled_samples = WebRtcOpus_Resample48to32(buffer16_left, |
| decoded_samples, |
| inst->state_48_32_left, |
| buffer_out); |
| |
| /* Add samples interleaved to output vector. */ |
| for (i = 0; i < resampled_samples; i++) { |
| decoded[i * 2] = buffer_out[i]; |
| } |
| |
| /* Resample from 48 kHz to 32 kHz for right channel. */ |
| resampled_samples = WebRtcOpus_Resample48to32(buffer16_right, |
| decoded_samples, |
| inst->state_48_32_right, |
| buffer_out); |
| |
| /* Add samples interleaved to output vector. */ |
| for (i = 0; i < decoded_samples; i++) { |
| decoded[i * 2 + 1] = buffer_out[i]; |
| } |
| } else { |
| /* Resample from 48 kHz to 32 kHz for left channel. */ |
| resampled_samples = WebRtcOpus_Resample48to32(buffer16_left, |
| decoded_samples, |
| inst->state_48_32_left, |
| decoded); |
| } |
| return resampled_samples; |
| } |
| |
| |
| int16_t WebRtcOpus_Decode(OpusDecInst* inst, const int16_t* encoded, |
| int16_t encoded_bytes, int16_t* decoded, |
| int16_t* audio_type) { |
| /* |buffer16| is big enough for 120 ms (the largestOpus packet size) of |
| * stereo audio at 48 kHz. */ |
| int16_t buffer16[kWebRtcOpusMaxFrameSize]; |
| int decoded_samples; |
| int16_t output_samples; |
| int i; |
| |
| /* If mono case, just do a regular call to the decoder. |
| * If stereo, call to WebRtcOpus_Decode() gives left channel as output, and |
| * calls to WebRtcOpus_Decode_slave() give right channel as output. |
| * This is to make stereo work with the current setup of NetEQ, which |
| * requires two calls to the decoder to produce stereo. */ |
| |
| /* Decode to a temporary buffer. */ |
| decoded_samples = DecodeNative(inst->decoder_left, encoded, encoded_bytes, |
| buffer16, audio_type); |
| if (decoded_samples < 0) { |
| return -1; |
| } |
| if (inst->channels == 2) { |
| /* The parameter |decoded_samples| holds the number of samples pairs, in |
| * case of stereo. Number of samples in |buffer16| equals |decoded_samples| |
| * times 2. */ |
| for (i = 0; i < decoded_samples; i++) { |
| /* Take every second sample, starting at the first sample. This gives |
| * the left channel. */ |
| buffer16[i] = buffer16[i * 2]; |
| } |
| } |
| |
| /* Resample from 48 kHz to 32 kHz. */ |
| output_samples = WebRtcOpus_Resample48to32(buffer16, decoded_samples, |
| inst->state_48_32_left, decoded); |
| |
| return output_samples; |
| } |
| |
| int16_t WebRtcOpus_DecodeSlave(OpusDecInst* inst, const int16_t* encoded, |
| int16_t encoded_bytes, int16_t* decoded, |
| int16_t* audio_type) { |
| /* |buffer16| is big enough for 120 ms (the largestOpus packet size) of |
| * stereo audio at 48 kHz. */ |
| int16_t buffer16[kWebRtcOpusMaxFrameSize]; |
| int decoded_samples; |
| int16_t output_samples; |
| int i; |
| |
| /* Decode to a temporary buffer. */ |
| decoded_samples = DecodeNative(inst->decoder_right, encoded, encoded_bytes, |
| buffer16, audio_type); |
| if (decoded_samples < 0) { |
| return -1; |
| } |
| if (inst->channels == 2) { |
| /* The parameter |decoded_samples| holds the number of samples pairs, in |
| * case of stereo. Number of samples in |buffer16| equals |decoded_samples| |
| * times 2. */ |
| for (i = 0; i < decoded_samples; i++) { |
| /* Take every second sample, starting at the second sample. This gives |
| * the right channel. */ |
| buffer16[i] = buffer16[i * 2 + 1]; |
| } |
| } else { |
| /* Decode slave should never be called for mono packets. */ |
| return -1; |
| } |
| /* Resample from 48 kHz to 32 kHz. */ |
| output_samples = WebRtcOpus_Resample48to32(buffer16, decoded_samples, |
| inst->state_48_32_right, decoded); |
| |
| return output_samples; |
| } |
| |
| int16_t WebRtcOpus_DecodePlc(OpusDecInst* inst, int16_t* decoded, |
| int16_t number_of_lost_frames) { |
| /* TODO(tlegrand): We can pass NULL to opus_decode to activate packet |
| * loss concealment, but I don't know how many samples |
| * number_of_lost_frames corresponds to. */ |
| return -1; |
| } |
| |
| int WebRtcOpus_DurationEst(OpusDecInst* inst, |
| const uint8_t* payload, |
| int payload_length_bytes) |
| { |
| int frames, samples; |
| frames = opus_packet_get_nb_frames(payload, payload_length_bytes); |
| if (frames < 0) { |
| /* Invalid payload data. */ |
| return 0; |
| } |
| samples = frames * opus_packet_get_samples_per_frame(payload, 48000); |
| if (samples < 120 || samples > 5760) { |
| /* Invalid payload duration. */ |
| return 0; |
| } |
| /* Compensate for the down-sampling from 48 kHz to 32 kHz. |
| * This should be removed when the resampling in WebRtcOpus_Decode is |
| * removed. */ |
| samples = samples * 2 / 3; |
| return samples; |
| } |