This is the first step in order to move bandwidth estimation closer to the network. The goal is to have RTP header parsing and bandwidth estimation before voice and video engine, and have a joint estimate for audio and video. Moving bandwidth estimation before the RTP module is also required for RTX. TEST=vie_auto_test, voe_auto_test, trybots. BUG=1811 R=andresp@webrtc.org, henrika@webrtc.org, mflodman@webrtc.org Review URL: https://webrtc-codereview.appspot.com/1545004 git-svn-id: http://webrtc.googlecode.com/svn/trunk@4129 4adac7df-926f-26a2-2b94-8c16560cd09d
383 lines
14 KiB
C++
383 lines
14 KiB
C++
/*
|
|
* Copyright (c) 2012 The WebRTC project authors. All Rights Reserved.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license
|
|
* that can be found in the LICENSE file in the root of the source
|
|
* tree. An additional intellectual property rights grant can be found
|
|
* in the file PATENTS. All contributing project authors may
|
|
* be found in the AUTHORS file in the root of the source tree.
|
|
*/
|
|
|
|
#include "webrtc/modules/rtp_rtcp/source/rtp_receiver_audio.h"
|
|
|
|
#include <math.h> // pow()
|
|
|
|
#include <cassert> // assert
|
|
#include <cstring> // memcpy()
|
|
|
|
#include "webrtc/system_wrappers/interface/critical_section_wrapper.h"
|
|
#include "webrtc/system_wrappers/interface/trace.h"
|
|
#include "webrtc/system_wrappers/interface/trace_event.h"
|
|
|
|
namespace webrtc {
|
|
RTPReceiverAudio::RTPReceiverAudio(const int32_t id,
|
|
RtpData* data_callback,
|
|
RtpAudioFeedback* incoming_messages_callback)
|
|
: RTPReceiverStrategy(data_callback),
|
|
id_(id),
|
|
critical_section_rtp_receiver_audio_(
|
|
CriticalSectionWrapper::CreateCriticalSection()),
|
|
last_received_frequency_(8000),
|
|
telephone_event_forward_to_decoder_(false),
|
|
telephone_event_payload_type_(-1),
|
|
cng_nb_payload_type_(-1),
|
|
cng_wb_payload_type_(-1),
|
|
cng_swb_payload_type_(-1),
|
|
cng_fb_payload_type_(-1),
|
|
cng_payload_type_(-1),
|
|
g722_payload_type_(-1),
|
|
last_received_g722_(false),
|
|
cb_audio_feedback_(incoming_messages_callback) {
|
|
last_payload_.Audio.channels = 1;
|
|
}
|
|
|
|
uint32_t RTPReceiverAudio::AudioFrequency() const {
|
|
CriticalSectionScoped lock(critical_section_rtp_receiver_audio_.get());
|
|
if (last_received_g722_) {
|
|
return 8000;
|
|
}
|
|
return last_received_frequency_;
|
|
}
|
|
|
|
// Outband TelephoneEvent(DTMF) detection
|
|
int RTPReceiverAudio::SetTelephoneEventForwardToDecoder(
|
|
bool forward_to_decoder) {
|
|
CriticalSectionScoped lock(critical_section_rtp_receiver_audio_.get());
|
|
telephone_event_forward_to_decoder_ = forward_to_decoder;
|
|
return 0;
|
|
}
|
|
|
|
// Is forwarding of outband telephone events turned on/off?
|
|
bool RTPReceiverAudio::TelephoneEventForwardToDecoder() const {
|
|
CriticalSectionScoped lock(critical_section_rtp_receiver_audio_.get());
|
|
return telephone_event_forward_to_decoder_;
|
|
}
|
|
|
|
bool RTPReceiverAudio::TelephoneEventPayloadType(
|
|
const int8_t payload_type) const {
|
|
CriticalSectionScoped lock(critical_section_rtp_receiver_audio_.get());
|
|
return (telephone_event_payload_type_ == payload_type) ? true : false;
|
|
}
|
|
|
|
bool RTPReceiverAudio::CNGPayloadType(const int8_t payload_type,
|
|
uint32_t* frequency,
|
|
bool* cng_payload_type_has_changed) {
|
|
CriticalSectionScoped lock(critical_section_rtp_receiver_audio_.get());
|
|
*cng_payload_type_has_changed = false;
|
|
|
|
// We can have four CNG on 8000Hz, 16000Hz, 32000Hz and 48000Hz.
|
|
if (cng_nb_payload_type_ == payload_type) {
|
|
*frequency = 8000;
|
|
if (cng_payload_type_ != -1 && cng_payload_type_ != cng_nb_payload_type_)
|
|
*cng_payload_type_has_changed = true;
|
|
|
|
cng_payload_type_ = cng_nb_payload_type_;
|
|
return true;
|
|
} else if (cng_wb_payload_type_ == payload_type) {
|
|
// if last received codec is G.722 we must use frequency 8000
|
|
if (last_received_g722_) {
|
|
*frequency = 8000;
|
|
} else {
|
|
*frequency = 16000;
|
|
}
|
|
if (cng_payload_type_ != -1 && cng_payload_type_ != cng_wb_payload_type_)
|
|
*cng_payload_type_has_changed = true;
|
|
cng_payload_type_ = cng_wb_payload_type_;
|
|
return true;
|
|
} else if (cng_swb_payload_type_ == payload_type) {
|
|
*frequency = 32000;
|
|
if ((cng_payload_type_ != -1) &&
|
|
(cng_payload_type_ != cng_swb_payload_type_))
|
|
*cng_payload_type_has_changed = true;
|
|
cng_payload_type_ = cng_swb_payload_type_;
|
|
return true;
|
|
} else if (cng_fb_payload_type_ == payload_type) {
|
|
*frequency = 48000;
|
|
if (cng_payload_type_ != -1 && cng_payload_type_ != cng_fb_payload_type_)
|
|
*cng_payload_type_has_changed = true;
|
|
cng_payload_type_ = cng_fb_payload_type_;
|
|
return true;
|
|
} else {
|
|
// not CNG
|
|
if (g722_payload_type_ == payload_type) {
|
|
last_received_g722_ = true;
|
|
} else {
|
|
last_received_g722_ = false;
|
|
}
|
|
}
|
|
return false;
|
|
}
|
|
|
|
bool RTPReceiverAudio::ShouldReportCsrcChanges(
|
|
uint8_t payload_type) const {
|
|
// Don't do this for DTMF packets, otherwise it's fine.
|
|
return !TelephoneEventPayloadType(payload_type);
|
|
}
|
|
|
|
// - Sample based or frame based codecs based on RFC 3551
|
|
// -
|
|
// - NOTE! There is one error in the RFC, stating G.722 uses 8 bits/samples.
|
|
// - The correct rate is 4 bits/sample.
|
|
// -
|
|
// - name of sampling default
|
|
// - encoding sample/frame bits/sample rate ms/frame ms/packet
|
|
// -
|
|
// - Sample based audio codecs
|
|
// - DVI4 sample 4 var. 20
|
|
// - G722 sample 4 16,000 20
|
|
// - G726-40 sample 5 8,000 20
|
|
// - G726-32 sample 4 8,000 20
|
|
// - G726-24 sample 3 8,000 20
|
|
// - G726-16 sample 2 8,000 20
|
|
// - L8 sample 8 var. 20
|
|
// - L16 sample 16 var. 20
|
|
// - PCMA sample 8 var. 20
|
|
// - PCMU sample 8 var. 20
|
|
// -
|
|
// - Frame based audio codecs
|
|
// - G723 frame N/A 8,000 30 30
|
|
// - G728 frame N/A 8,000 2.5 20
|
|
// - G729 frame N/A 8,000 10 20
|
|
// - G729D frame N/A 8,000 10 20
|
|
// - G729E frame N/A 8,000 10 20
|
|
// - GSM frame N/A 8,000 20 20
|
|
// - GSM-EFR frame N/A 8,000 20 20
|
|
// - LPC frame N/A 8,000 20 20
|
|
// - MPA frame N/A var. var.
|
|
// -
|
|
// - G7221 frame N/A
|
|
int32_t RTPReceiverAudio::OnNewPayloadTypeCreated(
|
|
const char payload_name[RTP_PAYLOAD_NAME_SIZE],
|
|
const int8_t payload_type,
|
|
const uint32_t frequency) {
|
|
CriticalSectionScoped lock(critical_section_rtp_receiver_audio_.get());
|
|
|
|
if (ModuleRTPUtility::StringCompare(payload_name, "telephone-event", 15)) {
|
|
telephone_event_payload_type_ = payload_type;
|
|
}
|
|
if (ModuleRTPUtility::StringCompare(payload_name, "cn", 2)) {
|
|
// we can have three CNG on 8000Hz, 16000Hz and 32000Hz
|
|
if (frequency == 8000) {
|
|
cng_nb_payload_type_ = payload_type;
|
|
} else if (frequency == 16000) {
|
|
cng_wb_payload_type_ = payload_type;
|
|
} else if (frequency == 32000) {
|
|
cng_swb_payload_type_ = payload_type;
|
|
} else if (frequency == 48000) {
|
|
cng_fb_payload_type_ = payload_type;
|
|
} else {
|
|
assert(false);
|
|
return -1;
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int32_t RTPReceiverAudio::ParseRtpPacket(
|
|
WebRtcRTPHeader* rtp_header,
|
|
const ModuleRTPUtility::PayloadUnion& specific_payload,
|
|
const bool is_red,
|
|
const uint8_t* packet,
|
|
const uint16_t packet_length,
|
|
const int64_t timestamp_ms,
|
|
const bool is_first_packet) {
|
|
TRACE_EVENT2("webrtc_rtp", "Audio::ParseRtp",
|
|
"seqnum", rtp_header->header.sequenceNumber,
|
|
"timestamp", rtp_header->header.timestamp);
|
|
rtp_header->type.Audio.numEnergy = rtp_header->header.numCSRCs;
|
|
const uint8_t* payload_data =
|
|
ModuleRTPUtility::GetPayloadData(rtp_header->header, packet);
|
|
const uint16_t payload_data_length =
|
|
ModuleRTPUtility::GetPayloadDataLength(rtp_header->header, packet_length);
|
|
|
|
return ParseAudioCodecSpecific(rtp_header,
|
|
payload_data,
|
|
payload_data_length,
|
|
specific_payload.Audio,
|
|
is_red);
|
|
}
|
|
|
|
int32_t RTPReceiverAudio::GetFrequencyHz() const {
|
|
return AudioFrequency();
|
|
}
|
|
|
|
RTPAliveType RTPReceiverAudio::ProcessDeadOrAlive(
|
|
uint16_t last_payload_length) const {
|
|
|
|
// Our CNG is 9 bytes; if it's a likely CNG the receiver needs to check
|
|
// kRtpNoRtp against NetEq speech_type kOutputPLCtoCNG.
|
|
if (last_payload_length < 10) { // our CNG is 9 bytes
|
|
return kRtpNoRtp;
|
|
} else {
|
|
return kRtpDead;
|
|
}
|
|
}
|
|
|
|
void RTPReceiverAudio::CheckPayloadChanged(
|
|
const int8_t payload_type,
|
|
ModuleRTPUtility::PayloadUnion* specific_payload,
|
|
bool* should_reset_statistics,
|
|
bool* should_discard_changes) {
|
|
*should_discard_changes = false;
|
|
*should_reset_statistics = false;
|
|
|
|
if (TelephoneEventPayloadType(payload_type)) {
|
|
// Don't do callbacks for DTMF packets.
|
|
*should_discard_changes = true;
|
|
return;
|
|
}
|
|
// frequency is updated for CNG
|
|
bool cng_payload_type_has_changed = false;
|
|
bool is_cng_payload_type = CNGPayloadType(payload_type,
|
|
&specific_payload->Audio.frequency,
|
|
&cng_payload_type_has_changed);
|
|
|
|
*should_reset_statistics = cng_payload_type_has_changed;
|
|
|
|
if (is_cng_payload_type) {
|
|
// Don't do callbacks for DTMF packets.
|
|
*should_discard_changes = true;
|
|
return;
|
|
}
|
|
}
|
|
|
|
int32_t RTPReceiverAudio::InvokeOnInitializeDecoder(
|
|
RtpFeedback* callback,
|
|
const int32_t id,
|
|
const int8_t payload_type,
|
|
const char payload_name[RTP_PAYLOAD_NAME_SIZE],
|
|
const ModuleRTPUtility::PayloadUnion& specific_payload) const {
|
|
if (-1 == callback->OnInitializeDecoder(id,
|
|
payload_type,
|
|
payload_name,
|
|
specific_payload.Audio.frequency,
|
|
specific_payload.Audio.channels,
|
|
specific_payload.Audio.rate)) {
|
|
WEBRTC_TRACE(kTraceError,
|
|
kTraceRtpRtcp,
|
|
id,
|
|
"Failed to create video decoder for payload type:%d",
|
|
payload_type);
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
// We are not allowed to have any critsects when calling data_callback.
|
|
int32_t RTPReceiverAudio::ParseAudioCodecSpecific(
|
|
WebRtcRTPHeader* rtp_header,
|
|
const uint8_t* payload_data,
|
|
const uint16_t payload_length,
|
|
const ModuleRTPUtility::AudioPayload& audio_specific,
|
|
const bool is_red) {
|
|
|
|
if (payload_length == 0) {
|
|
return 0;
|
|
}
|
|
|
|
bool telephone_event_packet =
|
|
TelephoneEventPayloadType(rtp_header->header.payloadType);
|
|
if (telephone_event_packet) {
|
|
CriticalSectionScoped lock(critical_section_rtp_receiver_audio_.get());
|
|
|
|
// RFC 4733 2.3
|
|
// 0 1 2 3
|
|
// 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9 0 1
|
|
// +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
|
|
// | event |E|R| volume | duration |
|
|
// +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
|
|
//
|
|
if (payload_length % 4 != 0) {
|
|
return -1;
|
|
}
|
|
uint8_t number_of_events = payload_length / 4;
|
|
|
|
// sanity
|
|
if (number_of_events >= MAX_NUMBER_OF_PARALLEL_TELEPHONE_EVENTS) {
|
|
number_of_events = MAX_NUMBER_OF_PARALLEL_TELEPHONE_EVENTS;
|
|
}
|
|
for (int n = 0; n < number_of_events; ++n) {
|
|
bool end = (payload_data[(4 * n) + 1] & 0x80) ? true : false;
|
|
|
|
std::set<uint8_t>::iterator event =
|
|
telephone_event_reported_.find(payload_data[4 * n]);
|
|
|
|
if (event != telephone_event_reported_.end()) {
|
|
// we have already seen this event
|
|
if (end) {
|
|
telephone_event_reported_.erase(payload_data[4 * n]);
|
|
}
|
|
} else {
|
|
if (end) {
|
|
// don't add if it's a end of a tone
|
|
} else {
|
|
telephone_event_reported_.insert(payload_data[4 * n]);
|
|
}
|
|
}
|
|
}
|
|
|
|
// RFC 4733 2.5.1.3 & 2.5.2.3 Long-Duration Events
|
|
// should not be a problem since we don't care about the duration
|
|
|
|
// RFC 4733 See 2.5.1.5. & 2.5.2.4. Multiple Events in a Packet
|
|
}
|
|
|
|
{
|
|
CriticalSectionScoped lock(critical_section_rtp_receiver_audio_.get());
|
|
|
|
if (!telephone_event_packet) {
|
|
last_received_frequency_ = audio_specific.frequency;
|
|
}
|
|
|
|
// Check if this is a CNG packet, receiver might want to know
|
|
uint32_t ignored;
|
|
bool also_ignored;
|
|
if (CNGPayloadType(rtp_header->header.payloadType,
|
|
&ignored,
|
|
&also_ignored)) {
|
|
rtp_header->type.Audio.isCNG = true;
|
|
rtp_header->frameType = kAudioFrameCN;
|
|
} else {
|
|
rtp_header->frameType = kAudioFrameSpeech;
|
|
rtp_header->type.Audio.isCNG = false;
|
|
}
|
|
|
|
// check if it's a DTMF event, hence something we can playout
|
|
if (telephone_event_packet) {
|
|
if (!telephone_event_forward_to_decoder_) {
|
|
// don't forward event to decoder
|
|
return 0;
|
|
}
|
|
std::set<uint8_t>::iterator first =
|
|
telephone_event_reported_.begin();
|
|
if (first != telephone_event_reported_.end() && *first > 15) {
|
|
// don't forward non DTMF events
|
|
return 0;
|
|
}
|
|
}
|
|
}
|
|
if (is_red && !(payload_data[0] & 0x80)) {
|
|
// we recive only one frame packed in a RED packet remove the RED wrapper
|
|
rtp_header->header.payloadType = payload_data[0];
|
|
|
|
// only one frame in the RED strip the one byte to help NetEq
|
|
return data_callback_->OnReceivedPayloadData(
|
|
payload_data + 1, payload_length - 1, rtp_header);
|
|
}
|
|
|
|
rtp_header->type.Audio.channel = audio_specific.channels;
|
|
return data_callback_->OnReceivedPayloadData(
|
|
payload_data, payload_length, rtp_header);
|
|
}
|
|
} // namespace webrtc
|