Measures the duration of each packet loss concealment (a.k.a. expand) event that is not followed by a merge operation. Having decoded and played packet m−1, the next expected packet is m. If packet m arrives after some time of packet loss concealment, we have a delayed packet outage event. However, if instead packet n>m arrives, we have a lost packet outage event. In NetEq, the two outage types results in different operations. Both types start with expand operations to generate audio to play while the buffer is empty. When a lost packet outage happens, the expand operation(s) are followed by one merge operation. For delayed packet outages, merge is not done, and the expand operations are immediately followed by normal operations. This change also includes unit tests for the new statistics. BUG=webrtc:4915, chromium:488124 R=minyue@webrtc.org Review URL: https://codereview.webrtc.org/1290113002 . Cr-Commit-Position: refs/heads/master@{#9725}
160 lines
5.0 KiB
C++
160 lines
5.0 KiB
C++
/*
|
|
* Copyright (c) 2012 The WebRTC project authors. All Rights Reserved.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license
|
|
* that can be found in the LICENSE file in the root of the source
|
|
* tree. An additional intellectual property rights grant can be found
|
|
* in the file PATENTS. All contributing project authors may
|
|
* be found in the AUTHORS file in the root of the source tree.
|
|
*/
|
|
|
|
#ifndef WEBRTC_MODULES_AUDIO_CODING_NETEQ_EXPAND_H_
|
|
#define WEBRTC_MODULES_AUDIO_CODING_NETEQ_EXPAND_H_
|
|
|
|
#include <assert.h>
|
|
|
|
#include "webrtc/base/constructormagic.h"
|
|
#include "webrtc/base/scoped_ptr.h"
|
|
#include "webrtc/modules/audio_coding/neteq/audio_multi_vector.h"
|
|
#include "webrtc/typedefs.h"
|
|
|
|
namespace webrtc {
|
|
|
|
// Forward declarations.
|
|
class BackgroundNoise;
|
|
class RandomVector;
|
|
class StatisticsCalculator;
|
|
class SyncBuffer;
|
|
|
|
// This class handles extrapolation of audio data from the sync_buffer to
|
|
// produce packet-loss concealment.
|
|
// TODO(hlundin): Refactor this class to divide the long methods into shorter
|
|
// ones.
|
|
class Expand {
|
|
public:
|
|
Expand(BackgroundNoise* background_noise,
|
|
SyncBuffer* sync_buffer,
|
|
RandomVector* random_vector,
|
|
StatisticsCalculator* statistics,
|
|
int fs,
|
|
size_t num_channels);
|
|
|
|
virtual ~Expand();
|
|
|
|
// Resets the object.
|
|
virtual void Reset();
|
|
|
|
// The main method to produce concealment data. The data is appended to the
|
|
// end of |output|.
|
|
virtual int Process(AudioMultiVector* output);
|
|
|
|
// Prepare the object to do extra expansion during normal operation following
|
|
// a period of expands.
|
|
virtual void SetParametersForNormalAfterExpand();
|
|
|
|
// Prepare the object to do extra expansion during merge operation following
|
|
// a period of expands.
|
|
virtual void SetParametersForMergeAfterExpand();
|
|
|
|
// Returns the mute factor for |channel|.
|
|
int16_t MuteFactor(size_t channel) {
|
|
assert(channel < num_channels_);
|
|
return channel_parameters_[channel].mute_factor;
|
|
}
|
|
|
|
// Accessors and mutators.
|
|
virtual size_t overlap_length() const;
|
|
int16_t max_lag() const { return max_lag_; }
|
|
|
|
protected:
|
|
static const int kMaxConsecutiveExpands = 200;
|
|
void GenerateRandomVector(int16_t seed_increment,
|
|
size_t length,
|
|
int16_t* random_vector);
|
|
|
|
void GenerateBackgroundNoise(int16_t* random_vector,
|
|
size_t channel,
|
|
int mute_slope,
|
|
bool too_many_expands,
|
|
size_t num_noise_samples,
|
|
int16_t* buffer);
|
|
|
|
// Initializes member variables at the beginning of an expand period.
|
|
void InitializeForAnExpandPeriod();
|
|
|
|
bool TooManyExpands();
|
|
|
|
// Analyzes the signal history in |sync_buffer_|, and set up all parameters
|
|
// necessary to produce concealment data.
|
|
void AnalyzeSignal(int16_t* random_vector);
|
|
|
|
RandomVector* const random_vector_;
|
|
SyncBuffer* const sync_buffer_;
|
|
bool first_expand_;
|
|
const int fs_hz_;
|
|
const size_t num_channels_;
|
|
int consecutive_expands_;
|
|
|
|
private:
|
|
static const int kUnvoicedLpcOrder = 6;
|
|
static const int kNumCorrelationCandidates = 3;
|
|
static const int kDistortionLength = 20;
|
|
static const int kLpcAnalysisLength = 160;
|
|
static const int kMaxSampleRate = 48000;
|
|
static const int kNumLags = 3;
|
|
|
|
struct ChannelParameters {
|
|
ChannelParameters();
|
|
int16_t mute_factor;
|
|
int16_t ar_filter[kUnvoicedLpcOrder + 1];
|
|
int16_t ar_filter_state[kUnvoicedLpcOrder];
|
|
int16_t ar_gain;
|
|
int16_t ar_gain_scale;
|
|
int16_t voice_mix_factor; /* Q14 */
|
|
int16_t current_voice_mix_factor; /* Q14 */
|
|
AudioVector expand_vector0;
|
|
AudioVector expand_vector1;
|
|
bool onset;
|
|
int mute_slope; /* Q20 */
|
|
};
|
|
|
|
// Calculate the auto-correlation of |input|, with length |input_length|
|
|
// samples. The correlation is calculated from a downsampled version of
|
|
// |input|, and is written to |output|. The scale factor is written to
|
|
// |output_scale|.
|
|
void Correlation(const int16_t* input,
|
|
size_t input_length,
|
|
int16_t* output,
|
|
int* output_scale) const;
|
|
|
|
void UpdateLagIndex();
|
|
|
|
BackgroundNoise* const background_noise_;
|
|
StatisticsCalculator* const statistics_;
|
|
const size_t overlap_length_;
|
|
int16_t max_lag_;
|
|
size_t expand_lags_[kNumLags];
|
|
int lag_index_direction_;
|
|
int current_lag_index_;
|
|
bool stop_muting_;
|
|
size_t expand_duration_samples_;
|
|
rtc::scoped_ptr<ChannelParameters[]> channel_parameters_;
|
|
|
|
DISALLOW_COPY_AND_ASSIGN(Expand);
|
|
};
|
|
|
|
struct ExpandFactory {
|
|
ExpandFactory() {}
|
|
virtual ~ExpandFactory() {}
|
|
|
|
virtual Expand* Create(BackgroundNoise* background_noise,
|
|
SyncBuffer* sync_buffer,
|
|
RandomVector* random_vector,
|
|
StatisticsCalculator* statistics,
|
|
int fs,
|
|
size_t num_channels) const;
|
|
};
|
|
|
|
} // namespace webrtc
|
|
#endif // WEBRTC_MODULES_AUDIO_CODING_NETEQ_EXPAND_H_
|