R=andrew@webrtc.org Review URL: https://webrtc-codereview.appspot.com/2296006 git-svn-id: http://webrtc.googlecode.com/svn/trunk@4844 4adac7df-926f-26a2-2b94-8c16560cd09d
662 lines
23 KiB
C++
662 lines
23 KiB
C++
/*
|
|
* Copyright (c) 2012 The WebRTC project authors. All Rights Reserved.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license
|
|
* that can be found in the LICENSE file in the root of the source
|
|
* tree. An additional intellectual property rights grant can be found
|
|
* in the file PATENTS. All contributing project authors may
|
|
* be found in the AUTHORS file in the root of the source tree.
|
|
*/
|
|
|
|
#include "webrtc/video_engine/vie_capturer.h"
|
|
|
|
#include "webrtc/common_video/libyuv/include/webrtc_libyuv.h"
|
|
#include "webrtc/modules/interface/module_common_types.h"
|
|
#include "webrtc/modules/utility/interface/process_thread.h"
|
|
#include "webrtc/modules/video_capture/include/video_capture_factory.h"
|
|
#include "webrtc/modules/video_processing/main/interface/video_processing.h"
|
|
#include "webrtc/modules/video_render/include/video_render_defines.h"
|
|
#include "webrtc/system_wrappers/interface/clock.h"
|
|
#include "webrtc/system_wrappers/interface/critical_section_wrapper.h"
|
|
#include "webrtc/system_wrappers/interface/event_wrapper.h"
|
|
#include "webrtc/system_wrappers/interface/thread_wrapper.h"
|
|
#include "webrtc/system_wrappers/interface/trace.h"
|
|
#include "webrtc/system_wrappers/interface/trace_event.h"
|
|
#include "webrtc/video_engine/include/vie_base.h"
|
|
#include "webrtc/video_engine/include/vie_image_process.h"
|
|
#include "webrtc/video_engine/overuse_frame_detector.h"
|
|
#include "webrtc/video_engine/vie_defines.h"
|
|
#include "webrtc/video_engine/vie_encoder.h"
|
|
|
|
namespace webrtc {
|
|
|
|
const int kThreadWaitTimeMs = 100;
|
|
|
|
ViECapturer::ViECapturer(int capture_id,
|
|
int engine_id,
|
|
const Config& config,
|
|
ProcessThread& module_process_thread)
|
|
: ViEFrameProviderBase(capture_id, engine_id),
|
|
capture_cs_(CriticalSectionWrapper::CreateCriticalSection()),
|
|
deliver_cs_(CriticalSectionWrapper::CreateCriticalSection()),
|
|
capture_module_(NULL),
|
|
external_capture_module_(NULL),
|
|
module_process_thread_(module_process_thread),
|
|
capture_id_(capture_id),
|
|
capture_thread_(*ThreadWrapper::CreateThread(ViECaptureThreadFunction,
|
|
this, kHighPriority,
|
|
"ViECaptureThread")),
|
|
capture_event_(*EventWrapper::Create()),
|
|
deliver_event_(*EventWrapper::Create()),
|
|
effect_filter_(NULL),
|
|
image_proc_module_(NULL),
|
|
image_proc_module_ref_counter_(0),
|
|
deflicker_frame_stats_(NULL),
|
|
brightness_frame_stats_(NULL),
|
|
current_brightness_level_(Normal),
|
|
reported_brightness_level_(Normal),
|
|
denoising_enabled_(false),
|
|
observer_cs_(CriticalSectionWrapper::CreateCriticalSection()),
|
|
observer_(NULL),
|
|
overuse_detector_(new OveruseFrameDetector(Clock::GetRealTimeClock(),
|
|
kNormalUseStdDevMs,
|
|
kOveruseStdDevMs)) {
|
|
WEBRTC_TRACE(kTraceMemory, kTraceVideo, ViEId(engine_id, capture_id),
|
|
"ViECapturer::ViECapturer(capture_id: %d, engine_id: %d)",
|
|
capture_id, engine_id);
|
|
unsigned int t_id = 0;
|
|
if (capture_thread_.Start(t_id)) {
|
|
WEBRTC_TRACE(kTraceInfo, kTraceVideo, ViEId(engine_id, capture_id),
|
|
"%s: thread started: %u", __FUNCTION__, t_id);
|
|
} else {
|
|
assert(false);
|
|
}
|
|
module_process_thread_.RegisterModule(overuse_detector_.get());
|
|
}
|
|
|
|
ViECapturer::~ViECapturer() {
|
|
WEBRTC_TRACE(kTraceMemory, kTraceVideo, ViEId(engine_id_, capture_id_),
|
|
"ViECapturer::~ViECapturer() - capture_id: %d, engine_id: %d",
|
|
capture_id_, engine_id_);
|
|
module_process_thread_.DeRegisterModule(overuse_detector_.get());
|
|
|
|
// Stop the thread.
|
|
deliver_cs_->Enter();
|
|
capture_cs_->Enter();
|
|
capture_thread_.SetNotAlive();
|
|
capture_event_.Set();
|
|
capture_cs_->Leave();
|
|
deliver_cs_->Leave();
|
|
|
|
// Stop the camera input.
|
|
if (capture_module_) {
|
|
module_process_thread_.DeRegisterModule(capture_module_);
|
|
capture_module_->DeRegisterCaptureDataCallback();
|
|
capture_module_->Release();
|
|
capture_module_ = NULL;
|
|
}
|
|
if (capture_thread_.Stop()) {
|
|
// Thread stopped.
|
|
delete &capture_thread_;
|
|
delete &capture_event_;
|
|
delete &deliver_event_;
|
|
} else {
|
|
assert(false);
|
|
WEBRTC_TRACE(kTraceMemory, kTraceVideoRenderer,
|
|
ViEId(engine_id_, capture_id_),
|
|
"%s: Not able to stop capture thread for device %d, leaking",
|
|
__FUNCTION__, capture_id_);
|
|
}
|
|
|
|
if (image_proc_module_) {
|
|
VideoProcessingModule::Destroy(image_proc_module_);
|
|
}
|
|
if (deflicker_frame_stats_) {
|
|
delete deflicker_frame_stats_;
|
|
deflicker_frame_stats_ = NULL;
|
|
}
|
|
delete brightness_frame_stats_;
|
|
}
|
|
|
|
ViECapturer* ViECapturer::CreateViECapture(
|
|
int capture_id,
|
|
int engine_id,
|
|
const Config& config,
|
|
VideoCaptureModule* capture_module,
|
|
ProcessThread& module_process_thread) {
|
|
ViECapturer* capture = new ViECapturer(capture_id, engine_id, config,
|
|
module_process_thread);
|
|
if (!capture || capture->Init(capture_module) != 0) {
|
|
delete capture;
|
|
capture = NULL;
|
|
}
|
|
return capture;
|
|
}
|
|
|
|
int32_t ViECapturer::Init(VideoCaptureModule* capture_module) {
|
|
assert(capture_module_ == NULL);
|
|
capture_module_ = capture_module;
|
|
capture_module_->RegisterCaptureDataCallback(*this);
|
|
capture_module_->AddRef();
|
|
if (module_process_thread_.RegisterModule(capture_module_) != 0) {
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
ViECapturer* ViECapturer::CreateViECapture(
|
|
int capture_id,
|
|
int engine_id,
|
|
const Config& config,
|
|
const char* device_unique_idUTF8,
|
|
const uint32_t device_unique_idUTF8Length,
|
|
ProcessThread& module_process_thread) {
|
|
ViECapturer* capture = new ViECapturer(capture_id, engine_id, config,
|
|
module_process_thread);
|
|
if (!capture ||
|
|
capture->Init(device_unique_idUTF8, device_unique_idUTF8Length) != 0) {
|
|
delete capture;
|
|
capture = NULL;
|
|
}
|
|
return capture;
|
|
}
|
|
|
|
int32_t ViECapturer::Init(const char* device_unique_idUTF8,
|
|
uint32_t device_unique_idUTF8Length) {
|
|
assert(capture_module_ == NULL);
|
|
if (device_unique_idUTF8 == NULL) {
|
|
capture_module_ = VideoCaptureFactory::Create(
|
|
ViEModuleId(engine_id_, capture_id_), external_capture_module_);
|
|
} else {
|
|
capture_module_ = VideoCaptureFactory::Create(
|
|
ViEModuleId(engine_id_, capture_id_), device_unique_idUTF8);
|
|
}
|
|
if (!capture_module_) {
|
|
return -1;
|
|
}
|
|
capture_module_->AddRef();
|
|
capture_module_->RegisterCaptureDataCallback(*this);
|
|
if (module_process_thread_.RegisterModule(capture_module_) != 0) {
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int ViECapturer::FrameCallbackChanged() {
|
|
if (Started() && !CaptureCapabilityFixed()) {
|
|
// Reconfigure the camera if a new size is required and the capture device
|
|
// does not provide encoded frames.
|
|
int best_width;
|
|
int best_height;
|
|
int best_frame_rate;
|
|
VideoCaptureCapability capture_settings;
|
|
capture_module_->CaptureSettings(capture_settings);
|
|
GetBestFormat(&best_width, &best_height, &best_frame_rate);
|
|
if (best_width != 0 && best_height != 0 && best_frame_rate != 0) {
|
|
if (best_width != capture_settings.width ||
|
|
best_height != capture_settings.height ||
|
|
best_frame_rate != capture_settings.maxFPS ||
|
|
capture_settings.codecType != kVideoCodecUnknown) {
|
|
Stop();
|
|
Start(requested_capability_);
|
|
}
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int32_t ViECapturer::Start(const CaptureCapability& capture_capability) {
|
|
WEBRTC_TRACE(kTraceInfo, kTraceVideo, ViEId(engine_id_, capture_id_), "%s",
|
|
__FUNCTION__);
|
|
int width;
|
|
int height;
|
|
int frame_rate;
|
|
VideoCaptureCapability capability;
|
|
requested_capability_ = capture_capability;
|
|
|
|
if (!CaptureCapabilityFixed()) {
|
|
// Ask the observers for best size.
|
|
GetBestFormat(&width, &height, &frame_rate);
|
|
if (width == 0) {
|
|
width = kViECaptureDefaultWidth;
|
|
}
|
|
if (height == 0) {
|
|
height = kViECaptureDefaultHeight;
|
|
}
|
|
if (frame_rate == 0) {
|
|
frame_rate = kViECaptureDefaultFramerate;
|
|
}
|
|
capability.height = height;
|
|
capability.width = width;
|
|
capability.maxFPS = frame_rate;
|
|
capability.rawType = kVideoI420;
|
|
capability.codecType = kVideoCodecUnknown;
|
|
} else {
|
|
// Width, height and type specified with call to Start, not set by
|
|
// observers.
|
|
capability.width = requested_capability_.width;
|
|
capability.height = requested_capability_.height;
|
|
capability.maxFPS = requested_capability_.maxFPS;
|
|
capability.rawType = requested_capability_.rawType;
|
|
capability.interlaced = requested_capability_.interlaced;
|
|
}
|
|
return capture_module_->StartCapture(capability);
|
|
}
|
|
|
|
int32_t ViECapturer::Stop() {
|
|
WEBRTC_TRACE(kTraceInfo, kTraceVideo, ViEId(engine_id_, capture_id_), "%s",
|
|
__FUNCTION__);
|
|
requested_capability_ = CaptureCapability();
|
|
return capture_module_->StopCapture();
|
|
}
|
|
|
|
bool ViECapturer::Started() {
|
|
WEBRTC_TRACE(kTraceInfo, kTraceVideo, ViEId(engine_id_, capture_id_), "%s",
|
|
__FUNCTION__);
|
|
return capture_module_->CaptureStarted();
|
|
}
|
|
|
|
const char* ViECapturer::CurrentDeviceName() const {
|
|
return capture_module_->CurrentDeviceName();
|
|
}
|
|
|
|
void ViECapturer::RegisterCpuOveruseObserver(CpuOveruseObserver* observer) {
|
|
overuse_detector_->SetObserver(observer);
|
|
}
|
|
|
|
int32_t ViECapturer::SetCaptureDelay(int32_t delay_ms) {
|
|
return capture_module_->SetCaptureDelay(delay_ms);
|
|
}
|
|
|
|
int32_t ViECapturer::SetRotateCapturedFrames(
|
|
const RotateCapturedFrame rotation) {
|
|
VideoCaptureRotation converted_rotation = kCameraRotate0;
|
|
switch (rotation) {
|
|
case RotateCapturedFrame_0:
|
|
converted_rotation = kCameraRotate0;
|
|
break;
|
|
case RotateCapturedFrame_90:
|
|
converted_rotation = kCameraRotate90;
|
|
break;
|
|
case RotateCapturedFrame_180:
|
|
converted_rotation = kCameraRotate180;
|
|
break;
|
|
case RotateCapturedFrame_270:
|
|
converted_rotation = kCameraRotate270;
|
|
break;
|
|
}
|
|
return capture_module_->SetCaptureRotation(converted_rotation);
|
|
}
|
|
|
|
int ViECapturer::IncomingFrame(unsigned char* video_frame,
|
|
unsigned int video_frame_length,
|
|
uint16_t width,
|
|
uint16_t height,
|
|
RawVideoType video_type,
|
|
unsigned long long capture_time) { // NOLINT
|
|
WEBRTC_TRACE(kTraceInfo, kTraceVideo, ViEId(engine_id_, capture_id_),
|
|
"ExternalCapture::IncomingFrame width %d, height %d, "
|
|
"capture_time %u", width, height, capture_time);
|
|
|
|
if (!external_capture_module_) {
|
|
return -1;
|
|
}
|
|
VideoCaptureCapability capability;
|
|
capability.width = width;
|
|
capability.height = height;
|
|
capability.rawType = video_type;
|
|
return external_capture_module_->IncomingFrame(video_frame,
|
|
video_frame_length,
|
|
capability, capture_time);
|
|
}
|
|
|
|
int ViECapturer::IncomingFrameI420(const ViEVideoFrameI420& video_frame,
|
|
unsigned long long capture_time) { // NOLINT
|
|
WEBRTC_TRACE(kTraceInfo, kTraceVideo, ViEId(engine_id_, capture_id_),
|
|
"ExternalCapture::IncomingFrame width %d, height %d, "
|
|
" capture_time %u", video_frame.width, video_frame.height,
|
|
capture_time);
|
|
|
|
if (!external_capture_module_) {
|
|
return -1;
|
|
}
|
|
|
|
VideoFrameI420 frame;
|
|
frame.width = video_frame.width;
|
|
frame.height = video_frame.height;
|
|
frame.y_plane = video_frame.y_plane;
|
|
frame.u_plane = video_frame.u_plane;
|
|
frame.v_plane = video_frame.v_plane;
|
|
frame.y_pitch = video_frame.y_pitch;
|
|
frame.u_pitch = video_frame.u_pitch;
|
|
frame.v_pitch = video_frame.v_pitch;
|
|
|
|
return external_capture_module_->IncomingFrameI420(frame, capture_time);
|
|
}
|
|
|
|
void ViECapturer::OnIncomingCapturedFrame(const int32_t capture_id,
|
|
I420VideoFrame& video_frame) {
|
|
WEBRTC_TRACE(kTraceStream, kTraceVideo, ViEId(engine_id_, capture_id_),
|
|
"%s(capture_id: %d)", __FUNCTION__, capture_id);
|
|
CriticalSectionScoped cs(capture_cs_.get());
|
|
// Make sure we render this frame earlier since we know the render time set
|
|
// is slightly off since it's being set when the frame has been received from
|
|
// the camera, and not when the camera actually captured the frame.
|
|
video_frame.set_render_time_ms(video_frame.render_time_ms() - FrameDelay());
|
|
|
|
TRACE_EVENT_ASYNC_BEGIN1("webrtc", "Video", video_frame.render_time_ms(),
|
|
"render_time", video_frame.render_time_ms());
|
|
|
|
captured_frame_.SwapFrame(&video_frame);
|
|
capture_event_.Set();
|
|
overuse_detector_->FrameCaptured(captured_frame_.width(),
|
|
captured_frame_.height());
|
|
return;
|
|
}
|
|
|
|
void ViECapturer::OnCaptureDelayChanged(const int32_t id,
|
|
const int32_t delay) {
|
|
WEBRTC_TRACE(kTraceStream, kTraceVideo, ViEId(engine_id_, capture_id_),
|
|
"%s(capture_id: %d) delay %d", __FUNCTION__, capture_id_,
|
|
delay);
|
|
|
|
// Deliver the network delay to all registered callbacks.
|
|
ViEFrameProviderBase::SetFrameDelay(delay);
|
|
}
|
|
|
|
int32_t ViECapturer::RegisterEffectFilter(
|
|
ViEEffectFilter* effect_filter) {
|
|
CriticalSectionScoped cs(deliver_cs_.get());
|
|
|
|
if (!effect_filter) {
|
|
if (!effect_filter_) {
|
|
WEBRTC_TRACE(kTraceError, kTraceVideo, ViEId(engine_id_, capture_id_),
|
|
"%s: no effect filter added for capture device %d",
|
|
__FUNCTION__, capture_id_);
|
|
return -1;
|
|
}
|
|
WEBRTC_TRACE(kTraceInfo, kTraceVideo, ViEId(engine_id_, capture_id_),
|
|
"%s: deregister effect filter for device %d", __FUNCTION__,
|
|
capture_id_);
|
|
} else {
|
|
if (effect_filter_) {
|
|
WEBRTC_TRACE(kTraceError, kTraceVideo, ViEId(engine_id_, capture_id_),
|
|
"%s: effect filter already added for capture device %d",
|
|
__FUNCTION__, capture_id_);
|
|
return -1;
|
|
}
|
|
WEBRTC_TRACE(kTraceInfo, kTraceVideo, ViEId(engine_id_, capture_id_),
|
|
"%s: register effect filter for device %d", __FUNCTION__,
|
|
capture_id_);
|
|
}
|
|
effect_filter_ = effect_filter;
|
|
return 0;
|
|
}
|
|
|
|
int32_t ViECapturer::IncImageProcRefCount() {
|
|
if (!image_proc_module_) {
|
|
assert(image_proc_module_ref_counter_ == 0);
|
|
image_proc_module_ = VideoProcessingModule::Create(
|
|
ViEModuleId(engine_id_, capture_id_));
|
|
if (!image_proc_module_) {
|
|
WEBRTC_TRACE(kTraceError, kTraceVideo, ViEId(engine_id_, capture_id_),
|
|
"%s: could not create video processing module",
|
|
__FUNCTION__);
|
|
return -1;
|
|
}
|
|
}
|
|
image_proc_module_ref_counter_++;
|
|
return 0;
|
|
}
|
|
|
|
int32_t ViECapturer::DecImageProcRefCount() {
|
|
image_proc_module_ref_counter_--;
|
|
if (image_proc_module_ref_counter_ == 0) {
|
|
// Destroy module.
|
|
VideoProcessingModule::Destroy(image_proc_module_);
|
|
image_proc_module_ = NULL;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int32_t ViECapturer::EnableDenoising(bool enable) {
|
|
WEBRTC_TRACE(kTraceInfo, kTraceVideo, ViEId(engine_id_, capture_id_),
|
|
"%s(capture_device_id: %d, enable: %d)", __FUNCTION__,
|
|
capture_id_, enable);
|
|
|
|
CriticalSectionScoped cs(deliver_cs_.get());
|
|
if (enable) {
|
|
if (denoising_enabled_) {
|
|
// Already enabled, nothing need to be done.
|
|
return 0;
|
|
}
|
|
denoising_enabled_ = true;
|
|
if (IncImageProcRefCount() != 0) {
|
|
return -1;
|
|
}
|
|
} else {
|
|
if (denoising_enabled_ == false) {
|
|
// Already disabled, nothing need to be done.
|
|
return 0;
|
|
}
|
|
denoising_enabled_ = false;
|
|
DecImageProcRefCount();
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int32_t ViECapturer::EnableDeflickering(bool enable) {
|
|
WEBRTC_TRACE(kTraceInfo, kTraceVideo, ViEId(engine_id_, capture_id_),
|
|
"%s(capture_device_id: %d, enable: %d)", __FUNCTION__,
|
|
capture_id_, enable);
|
|
|
|
CriticalSectionScoped cs(deliver_cs_.get());
|
|
if (enable) {
|
|
if (deflicker_frame_stats_) {
|
|
WEBRTC_TRACE(kTraceError, kTraceVideo, ViEId(engine_id_, capture_id_),
|
|
"%s: deflickering already enabled", __FUNCTION__);
|
|
return -1;
|
|
}
|
|
if (IncImageProcRefCount() != 0) {
|
|
return -1;
|
|
}
|
|
deflicker_frame_stats_ = new VideoProcessingModule::FrameStats();
|
|
} else {
|
|
if (deflicker_frame_stats_ == NULL) {
|
|
WEBRTC_TRACE(kTraceError, kTraceVideo, ViEId(engine_id_, capture_id_),
|
|
"%s: deflickering not enabled", __FUNCTION__);
|
|
return -1;
|
|
}
|
|
DecImageProcRefCount();
|
|
delete deflicker_frame_stats_;
|
|
deflicker_frame_stats_ = NULL;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int32_t ViECapturer::EnableBrightnessAlarm(bool enable) {
|
|
WEBRTC_TRACE(kTraceInfo, kTraceVideo, ViEId(engine_id_, capture_id_),
|
|
"%s(capture_device_id: %d, enable: %d)", __FUNCTION__,
|
|
capture_id_, enable);
|
|
|
|
CriticalSectionScoped cs(deliver_cs_.get());
|
|
if (enable) {
|
|
if (brightness_frame_stats_) {
|
|
WEBRTC_TRACE(kTraceError, kTraceVideo, ViEId(engine_id_, capture_id_),
|
|
"%s: BrightnessAlarm already enabled", __FUNCTION__);
|
|
return -1;
|
|
}
|
|
if (IncImageProcRefCount() != 0) {
|
|
return -1;
|
|
}
|
|
brightness_frame_stats_ = new VideoProcessingModule::FrameStats();
|
|
} else {
|
|
DecImageProcRefCount();
|
|
if (brightness_frame_stats_ == NULL) {
|
|
WEBRTC_TRACE(kTraceError, kTraceVideo, ViEId(engine_id_, capture_id_),
|
|
"%s: deflickering not enabled", __FUNCTION__);
|
|
return -1;
|
|
}
|
|
delete brightness_frame_stats_;
|
|
brightness_frame_stats_ = NULL;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
bool ViECapturer::ViECaptureThreadFunction(void* obj) {
|
|
return static_cast<ViECapturer*>(obj)->ViECaptureProcess();
|
|
}
|
|
|
|
bool ViECapturer::ViECaptureProcess() {
|
|
if (capture_event_.Wait(kThreadWaitTimeMs) == kEventSignaled) {
|
|
deliver_cs_->Enter();
|
|
if (SwapCapturedAndDeliverFrameIfAvailable()) {
|
|
DeliverI420Frame(&deliver_frame_);
|
|
}
|
|
deliver_cs_->Leave();
|
|
if (current_brightness_level_ != reported_brightness_level_) {
|
|
CriticalSectionScoped cs(observer_cs_.get());
|
|
if (observer_) {
|
|
observer_->BrightnessAlarm(id_, current_brightness_level_);
|
|
reported_brightness_level_ = current_brightness_level_;
|
|
}
|
|
}
|
|
}
|
|
// We're done!
|
|
return true;
|
|
}
|
|
|
|
void ViECapturer::DeliverI420Frame(I420VideoFrame* video_frame) {
|
|
// Apply image enhancement and effect filter.
|
|
if (deflicker_frame_stats_) {
|
|
if (image_proc_module_->GetFrameStats(deflicker_frame_stats_,
|
|
*video_frame) == 0) {
|
|
image_proc_module_->Deflickering(video_frame, deflicker_frame_stats_);
|
|
} else {
|
|
WEBRTC_TRACE(kTraceStream, kTraceVideo, ViEId(engine_id_, capture_id_),
|
|
"%s: could not get frame stats for captured frame",
|
|
__FUNCTION__);
|
|
}
|
|
}
|
|
if (denoising_enabled_) {
|
|
image_proc_module_->Denoising(video_frame);
|
|
}
|
|
if (brightness_frame_stats_) {
|
|
if (image_proc_module_->GetFrameStats(brightness_frame_stats_,
|
|
*video_frame) == 0) {
|
|
int32_t brightness = image_proc_module_->BrightnessDetection(
|
|
*video_frame, *brightness_frame_stats_);
|
|
|
|
switch (brightness) {
|
|
case VideoProcessingModule::kNoWarning:
|
|
current_brightness_level_ = Normal;
|
|
break;
|
|
case VideoProcessingModule::kDarkWarning:
|
|
current_brightness_level_ = Dark;
|
|
break;
|
|
case VideoProcessingModule::kBrightWarning:
|
|
current_brightness_level_ = Bright;
|
|
break;
|
|
default:
|
|
WEBRTC_TRACE(kTraceInfo, kTraceVideo, ViEId(engine_id_, capture_id_),
|
|
"%s: Brightness detection failed", __FUNCTION__);
|
|
}
|
|
}
|
|
}
|
|
if (effect_filter_) {
|
|
unsigned int length = CalcBufferSize(kI420,
|
|
video_frame->width(),
|
|
video_frame->height());
|
|
scoped_array<uint8_t> video_buffer(new uint8_t[length]);
|
|
ExtractBuffer(*video_frame, length, video_buffer.get());
|
|
effect_filter_->Transform(length, video_buffer.get(),
|
|
video_frame->timestamp(), video_frame->width(),
|
|
video_frame->height());
|
|
}
|
|
// Deliver the captured frame to all observers (channels, renderer or file).
|
|
ViEFrameProviderBase::DeliverFrame(video_frame);
|
|
}
|
|
|
|
int ViECapturer::DeregisterFrameCallback(
|
|
const ViEFrameCallback* callbackObject) {
|
|
return ViEFrameProviderBase::DeregisterFrameCallback(callbackObject);
|
|
}
|
|
|
|
bool ViECapturer::IsFrameCallbackRegistered(
|
|
const ViEFrameCallback* callbackObject) {
|
|
CriticalSectionScoped cs(provider_cs_.get());
|
|
return ViEFrameProviderBase::IsFrameCallbackRegistered(callbackObject);
|
|
}
|
|
|
|
bool ViECapturer::CaptureCapabilityFixed() {
|
|
return requested_capability_.width != 0 &&
|
|
requested_capability_.height != 0 &&
|
|
requested_capability_.maxFPS != 0;
|
|
}
|
|
|
|
int32_t ViECapturer::RegisterObserver(ViECaptureObserver* observer) {
|
|
if (observer_) {
|
|
WEBRTC_TRACE(kTraceError, kTraceVideo, ViEId(engine_id_, capture_id_),
|
|
"%s Observer already registered", __FUNCTION__, capture_id_);
|
|
return -1;
|
|
}
|
|
if (capture_module_->RegisterCaptureCallback(*this) != 0) {
|
|
return -1;
|
|
}
|
|
capture_module_->EnableFrameRateCallback(true);
|
|
capture_module_->EnableNoPictureAlarm(true);
|
|
observer_ = observer;
|
|
return 0;
|
|
}
|
|
|
|
int32_t ViECapturer::DeRegisterObserver() {
|
|
CriticalSectionScoped cs(observer_cs_.get());
|
|
if (!observer_) {
|
|
return 0;
|
|
}
|
|
capture_module_->EnableFrameRateCallback(false);
|
|
capture_module_->EnableNoPictureAlarm(false);
|
|
capture_module_->DeRegisterCaptureCallback();
|
|
observer_ = NULL;
|
|
return 0;
|
|
}
|
|
|
|
bool ViECapturer::IsObserverRegistered() {
|
|
CriticalSectionScoped cs(observer_cs_.get());
|
|
return observer_ != NULL;
|
|
}
|
|
|
|
void ViECapturer::OnCaptureFrameRate(const int32_t id,
|
|
const uint32_t frame_rate) {
|
|
WEBRTC_TRACE(kTraceStream, kTraceVideo, ViEId(engine_id_, capture_id_),
|
|
"OnCaptureFrameRate %d", frame_rate);
|
|
|
|
CriticalSectionScoped cs(observer_cs_.get());
|
|
observer_->CapturedFrameRate(id_, static_cast<uint8_t>(frame_rate));
|
|
}
|
|
|
|
void ViECapturer::OnNoPictureAlarm(const int32_t id,
|
|
const VideoCaptureAlarm alarm) {
|
|
WEBRTC_TRACE(kTraceStream, kTraceVideo, ViEId(engine_id_, capture_id_),
|
|
"OnNoPictureAlarm %d", alarm);
|
|
|
|
CriticalSectionScoped cs(observer_cs_.get());
|
|
CaptureAlarm vie_alarm = (alarm == Raised) ? AlarmRaised : AlarmCleared;
|
|
observer_->NoPictureAlarm(id, vie_alarm);
|
|
}
|
|
|
|
bool ViECapturer::SwapCapturedAndDeliverFrameIfAvailable() {
|
|
CriticalSectionScoped cs(capture_cs_.get());
|
|
if (captured_frame_.IsZeroSize())
|
|
return false;
|
|
|
|
deliver_frame_.SwapFrame(&captured_frame_);
|
|
captured_frame_.ResetSize();
|
|
return true;
|
|
}
|
|
|
|
} // namespace webrtc
|