Together with Stefan's http://review.webrtc.org/1960004/. R=mallinath@webrtc.org Review URL: https://webrtc-codereview.appspot.com/2048004 git-svn-id: http://webrtc.googlecode.com/svn/trunk@4556 4adac7df-926f-26a2-2b94-8c16560cd09d
205 lines
7.3 KiB
C++
205 lines
7.3 KiB
C++
/*
|
|
* Copyright (c) 2012 The WebRTC project authors. All Rights Reserved.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license
|
|
* that can be found in the LICENSE file in the root of the source
|
|
* tree. An additional intellectual property rights grant can be found
|
|
* in the file PATENTS. All contributing project authors may
|
|
* be found in the AUTHORS file in the root of the source tree.
|
|
*/
|
|
|
|
#include "webrtc/video_engine/vie_sync_module.h"
|
|
|
|
#include "webrtc/modules/rtp_rtcp/interface/rtp_receiver.h"
|
|
#include "webrtc/modules/rtp_rtcp/interface/rtp_rtcp.h"
|
|
#include "webrtc/modules/video_coding/main/interface/video_coding.h"
|
|
#include "webrtc/system_wrappers/interface/critical_section_wrapper.h"
|
|
#include "webrtc/system_wrappers/interface/trace.h"
|
|
#include "webrtc/system_wrappers/interface/trace_event.h"
|
|
#include "webrtc/video_engine/stream_synchronization.h"
|
|
#include "webrtc/video_engine/vie_channel.h"
|
|
#include "webrtc/voice_engine/include/voe_video_sync.h"
|
|
|
|
namespace webrtc {
|
|
|
|
enum { kSyncInterval = 1000};
|
|
|
|
int UpdateMeasurements(StreamSynchronization::Measurements* stream,
|
|
const RtpRtcp& rtp_rtcp, const RtpReceiver& receiver) {
|
|
stream->latest_timestamp = receiver.Timestamp();
|
|
stream->latest_receive_time_ms = receiver.LastReceivedTimeMs();
|
|
synchronization::RtcpMeasurement measurement;
|
|
if (0 != rtp_rtcp.RemoteNTP(&measurement.ntp_secs,
|
|
&measurement.ntp_frac,
|
|
NULL,
|
|
NULL,
|
|
&measurement.rtp_timestamp)) {
|
|
return -1;
|
|
}
|
|
if (measurement.ntp_secs == 0 && measurement.ntp_frac == 0) {
|
|
return -1;
|
|
}
|
|
for (synchronization::RtcpList::iterator it = stream->rtcp.begin();
|
|
it != stream->rtcp.end(); ++it) {
|
|
if (measurement.ntp_secs == (*it).ntp_secs &&
|
|
measurement.ntp_frac == (*it).ntp_frac) {
|
|
// This RTCP has already been added to the list.
|
|
return 0;
|
|
}
|
|
}
|
|
// We need two RTCP SR reports to map between RTP and NTP. More than two will
|
|
// not improve the mapping.
|
|
if (stream->rtcp.size() == 2) {
|
|
stream->rtcp.pop_back();
|
|
}
|
|
stream->rtcp.push_front(measurement);
|
|
return 0;
|
|
}
|
|
|
|
ViESyncModule::ViESyncModule(VideoCodingModule* vcm,
|
|
ViEChannel* vie_channel)
|
|
: data_cs_(CriticalSectionWrapper::CreateCriticalSection()),
|
|
vcm_(vcm),
|
|
vie_channel_(vie_channel),
|
|
video_receiver_(NULL),
|
|
video_rtp_rtcp_(NULL),
|
|
voe_channel_id_(-1),
|
|
voe_sync_interface_(NULL),
|
|
last_sync_time_(TickTime::Now()),
|
|
sync_() {
|
|
}
|
|
|
|
ViESyncModule::~ViESyncModule() {
|
|
}
|
|
|
|
int ViESyncModule::ConfigureSync(int voe_channel_id,
|
|
VoEVideoSync* voe_sync_interface,
|
|
RtpRtcp* video_rtcp_module,
|
|
RtpReceiver* video_receiver) {
|
|
CriticalSectionScoped cs(data_cs_.get());
|
|
voe_channel_id_ = voe_channel_id;
|
|
voe_sync_interface_ = voe_sync_interface;
|
|
video_receiver_ = video_receiver;
|
|
video_rtp_rtcp_ = video_rtcp_module;
|
|
sync_.reset(new StreamSynchronization(voe_channel_id, vie_channel_->Id()));
|
|
|
|
if (!voe_sync_interface) {
|
|
voe_channel_id_ = -1;
|
|
if (voe_channel_id >= 0) {
|
|
// Trying to set a voice channel but no interface exist.
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int ViESyncModule::VoiceChannel() {
|
|
return voe_channel_id_;
|
|
}
|
|
|
|
int32_t ViESyncModule::TimeUntilNextProcess() {
|
|
return static_cast<int32_t>(kSyncInterval -
|
|
(TickTime::Now() - last_sync_time_).Milliseconds());
|
|
}
|
|
|
|
int32_t ViESyncModule::Process() {
|
|
CriticalSectionScoped cs(data_cs_.get());
|
|
last_sync_time_ = TickTime::Now();
|
|
|
|
const int current_video_delay_ms = vcm_->Delay();
|
|
WEBRTC_TRACE(webrtc::kTraceInfo, webrtc::kTraceVideo, vie_channel_->Id(),
|
|
"Video delay (JB + decoder) is %d ms", current_video_delay_ms);
|
|
|
|
if (voe_channel_id_ == -1) {
|
|
return 0;
|
|
}
|
|
assert(video_rtp_rtcp_ && voe_sync_interface_);
|
|
assert(sync_.get());
|
|
|
|
int audio_jitter_buffer_delay_ms = 0;
|
|
int playout_buffer_delay_ms = 0;
|
|
if (voe_sync_interface_->GetDelayEstimate(voe_channel_id_,
|
|
&audio_jitter_buffer_delay_ms,
|
|
&playout_buffer_delay_ms) != 0) {
|
|
// Could not get VoE delay value, probably not a valid channel Id or
|
|
// the channel have not received enough packets.
|
|
WEBRTC_TRACE(webrtc::kTraceStream, webrtc::kTraceVideo, vie_channel_->Id(),
|
|
"%s: VE_GetDelayEstimate error for voice_channel %d",
|
|
__FUNCTION__, voe_channel_id_);
|
|
return 0;
|
|
}
|
|
const int current_audio_delay_ms = audio_jitter_buffer_delay_ms +
|
|
playout_buffer_delay_ms;
|
|
|
|
RtpRtcp* voice_rtp_rtcp = NULL;
|
|
RtpReceiver* voice_receiver = NULL;
|
|
if (0 != voe_sync_interface_->GetRtpRtcp(voe_channel_id_, &voice_rtp_rtcp,
|
|
&voice_receiver)) {
|
|
return 0;
|
|
}
|
|
assert(voice_rtp_rtcp);
|
|
assert(voice_receiver);
|
|
|
|
if (UpdateMeasurements(&video_measurement_, *video_rtp_rtcp_,
|
|
*video_receiver_) != 0) {
|
|
return 0;
|
|
}
|
|
|
|
if (UpdateMeasurements(&audio_measurement_, *voice_rtp_rtcp,
|
|
*voice_receiver) != 0) {
|
|
return 0;
|
|
}
|
|
|
|
int relative_delay_ms;
|
|
// Calculate how much later or earlier the audio stream is compared to video.
|
|
if (!sync_->ComputeRelativeDelay(audio_measurement_, video_measurement_,
|
|
&relative_delay_ms)) {
|
|
return 0;
|
|
}
|
|
|
|
TRACE_COUNTER1("webrtc", "SyncCurrentVideoDelay", current_video_delay_ms);
|
|
TRACE_COUNTER1("webrtc", "SyncCurrentAudioDelay", current_audio_delay_ms);
|
|
TRACE_COUNTER1("webrtc", "SyncRelativeDelay", relative_delay_ms);
|
|
int target_audio_delay_ms = 0;
|
|
int target_video_delay_ms = current_video_delay_ms;
|
|
// Calculate the necessary extra audio delay and desired total video
|
|
// delay to get the streams in sync.
|
|
if (!sync_->ComputeDelays(relative_delay_ms,
|
|
current_audio_delay_ms,
|
|
&target_audio_delay_ms,
|
|
&target_video_delay_ms)) {
|
|
return 0;
|
|
}
|
|
|
|
WEBRTC_TRACE(webrtc::kTraceInfo, webrtc::kTraceVideo, vie_channel_->Id(),
|
|
"Set delay current(a=%d v=%d rel=%d) target(a=%d v=%d)",
|
|
current_audio_delay_ms, current_video_delay_ms,
|
|
relative_delay_ms,
|
|
target_audio_delay_ms, target_video_delay_ms);
|
|
if (voe_sync_interface_->SetMinimumPlayoutDelay(
|
|
voe_channel_id_, target_audio_delay_ms) == -1) {
|
|
WEBRTC_TRACE(webrtc::kTraceDebug, webrtc::kTraceVideo, vie_channel_->Id(),
|
|
"Error setting voice delay");
|
|
}
|
|
vcm_->SetMinimumPlayoutDelay(target_video_delay_ms);
|
|
return 0;
|
|
}
|
|
|
|
int ViESyncModule::SetTargetBufferingDelay(int target_delay_ms) {
|
|
CriticalSectionScoped cs(data_cs_.get());
|
|
if (!voe_sync_interface_) {
|
|
WEBRTC_TRACE(webrtc::kTraceInfo, webrtc::kTraceVideo, vie_channel_->Id(),
|
|
"voe_sync_interface_ NULL, can't set playout delay.");
|
|
return -1;
|
|
}
|
|
sync_->SetTargetBufferingDelay(target_delay_ms);
|
|
// Setting initial playout delay to voice engine (video engine is updated via
|
|
// the VCM interface).
|
|
voe_sync_interface_->SetInitialPlayoutDelay(voe_channel_id_,
|
|
target_delay_ms);
|
|
return 0;
|
|
}
|
|
|
|
} // namespace webrtc
|