blob: 349153ddaf8722ee81c0d3372a6944efdde3f045 [file] [log] [blame]
/*
* Copyright (c) 2011 The WebRTC project authors. All Rights Reserved.
*
* Use of this source code is governed by a BSD-style license
* that can be found in the LICENSE file in the root of the source
* tree. An additional intellectual property rights grant can be found
* in the file PATENTS. All contributing project authors may
* be found in the AUTHORS file in the root of the source tree.
*/
#include "video_capture_impl.h"
#include "common_video/libyuv/include/libyuv.h"
#include "critical_section_wrapper.h"
#include "module_common_types.h"
#include "ref_count.h"
#include "tick_util.h"
#include "trace.h"
#include "video_capture_config.h"
#ifdef WEBRTC_ANDROID
#include "video_capture_android.h" // Need inclusion here to set Java environment.
#endif
namespace webrtc
{
namespace videocapturemodule
{
VideoCaptureModule* VideoCaptureImpl::Create(
const WebRtc_Word32 id,
VideoCaptureExternal*& externalCapture)
{
RefCountImpl<VideoCaptureImpl>* implementation =
new RefCountImpl<VideoCaptureImpl>(id);
externalCapture = implementation;
return implementation;
}
const WebRtc_UWord8* VideoCaptureImpl::CurrentDeviceName() const
{
return _deviceUniqueId;
}
WebRtc_Word32 VideoCaptureImpl::ChangeUniqueId(const WebRtc_Word32 id)
{
_id = id;
return 0;
}
// returns the number of milliseconds until the module want a worker thread to call Process
WebRtc_Word32 VideoCaptureImpl::TimeUntilNextProcess()
{
TickTime timeNow = TickTime::Now();
WebRtc_Word32 timeToNormalProcess = kProcessInterval
- (WebRtc_Word32)((TickTime::Now() - _lastProcessTime).Milliseconds());
WebRtc_Word32 timeToStartImage = timeToNormalProcess;
if (_startImageFrameIntervall)
{
timeToStartImage = _startImageFrameIntervall
- (WebRtc_Word32)((timeNow - _lastSentStartImageTime).Milliseconds());
if (timeToStartImage < 0)
{
timeToStartImage = 0;
}
}
return (timeToStartImage < timeToNormalProcess)
? timeToStartImage : timeToNormalProcess;
}
// Process any pending tasks such as timeouts
WebRtc_Word32 VideoCaptureImpl::Process()
{
CriticalSectionScoped cs(_callBackCs);
const TickTime now = TickTime::Now();
_lastProcessTime = TickTime::Now();
// Handle No picture alarm
if (_lastProcessFrameCount.Ticks() == _incomingFrameTimes[0].Ticks() &&
_captureAlarm != Raised)
{
if (_noPictureAlarmCallBack && _captureCallBack)
{
_captureAlarm = Raised;
_captureCallBack->OnNoPictureAlarm(_id, _captureAlarm);
}
}
else if (_lastProcessFrameCount.Ticks() != _incomingFrameTimes[0].Ticks() &&
_captureAlarm != Cleared)
{
if (_noPictureAlarmCallBack && _captureCallBack)
{
_captureAlarm = Cleared;
_captureCallBack->OnNoPictureAlarm(_id, _captureAlarm);
}
}
// Handle frame rate callback
if ((now - _lastFrameRateCallbackTime).Milliseconds()
> kFrameRateCallbackInterval)
{
if (_frameRateCallBack && _captureCallBack)
{
const WebRtc_UWord32 frameRate = CalculateFrameRate(now);
_captureCallBack->OnCaptureFrameRate(_id, frameRate);
}
_lastFrameRateCallbackTime = now; // Can be set by EnableFrameRateCallback
}
_lastProcessFrameCount = _incomingFrameTimes[0];
// Handle start image frame rates.
if (_startImageFrameIntervall
&& (now - _lastSentStartImageTime).Milliseconds() >= _startImageFrameIntervall)
{
_lastSentStartImageTime = now;
if (_dataCallBack)
{
_captureFrame.CopyFrame(_startImage);
_captureFrame.SetRenderTime(TickTime::MillisecondTimestamp());
_dataCallBack->OnIncomingCapturedFrame(_id, _captureFrame,
kVideoCodecUnknown);
}
}
return 0;
}
VideoCaptureImpl::VideoCaptureImpl(const WebRtc_Word32 id)
: _id(id), _deviceUniqueId(NULL), _apiCs(*CriticalSectionWrapper::CreateCriticalSection()),
_captureDelay(0), _requestedCapability(),
_callBackCs(*CriticalSectionWrapper::CreateCriticalSection()),
_lastProcessTime(TickTime::Now()),
_lastFrameRateCallbackTime(TickTime::Now()), _frameRateCallBack(false),
_noPictureAlarmCallBack(false), _captureAlarm(Cleared), _setCaptureDelay(0),
_dataCallBack(NULL), _captureCallBack(NULL), _startImageFrameIntervall(0),
_lastProcessFrameCount(TickTime::Now()), _rotateFrame(kRotateNone),
last_capture_time_(TickTime::MillisecondTimestamp())
{
_requestedCapability.width = kDefaultWidth;
_requestedCapability.height = kDefaultHeight;
_requestedCapability.maxFPS = 30;
_requestedCapability.rawType = kVideoI420;
_requestedCapability.codecType = kVideoCodecUnknown;
memset(_incomingFrameTimes, 0, sizeof(_incomingFrameTimes));
}
VideoCaptureImpl::~VideoCaptureImpl()
{
DeRegisterCaptureDataCallback();
DeRegisterCaptureCallback();
delete &_callBackCs;
delete &_apiCs;
if (_deviceUniqueId)
delete[] _deviceUniqueId;
}
WebRtc_Word32 VideoCaptureImpl::RegisterCaptureDataCallback(
VideoCaptureDataCallback& dataCallBack)
{
WEBRTC_TRACE(webrtc::kTraceModuleCall, webrtc::kTraceVideoCapture, _id,
"RegisterCaptureDataCallback");
CriticalSectionScoped cs(_apiCs);
CriticalSectionScoped cs2(_callBackCs);
_dataCallBack = &dataCallBack;
return 0;
}
WebRtc_Word32 VideoCaptureImpl::DeRegisterCaptureDataCallback()
{
WEBRTC_TRACE(webrtc::kTraceModuleCall, webrtc::kTraceVideoCapture, _id,
"DeRegisterCaptureDataCallback");
CriticalSectionScoped cs(_apiCs);
CriticalSectionScoped cs2(_callBackCs);
_dataCallBack = NULL;
return 0;
}
WebRtc_Word32 VideoCaptureImpl::RegisterCaptureCallback(VideoCaptureFeedBack& callBack)
{
WEBRTC_TRACE(webrtc::kTraceModuleCall, webrtc::kTraceVideoCapture, _id, "RegisterCaptureCallback %x");
CriticalSectionScoped cs(_apiCs);
CriticalSectionScoped cs2(_callBackCs);
_captureCallBack = &callBack;
return 0;
}
WebRtc_Word32 VideoCaptureImpl::DeRegisterCaptureCallback()
{
WEBRTC_TRACE(webrtc::kTraceModuleCall, webrtc::kTraceVideoCapture, _id, "DeRegisterCaptureCallback");
CriticalSectionScoped cs(_apiCs);
CriticalSectionScoped cs2(_callBackCs);
_captureCallBack = NULL;
return 0;
}
WebRtc_Word32 VideoCaptureImpl::SetCaptureDelay(WebRtc_Word32 delayMS)
{
WEBRTC_TRACE(webrtc::kTraceModuleCall, webrtc::kTraceVideoCapture, _id, "SetCaptureDelay %d",
(int) delayMS);
CriticalSectionScoped cs(_apiCs);
_captureDelay = delayMS;
return 0;
}
WebRtc_Word32 VideoCaptureImpl::CaptureDelay()
{
WEBRTC_TRACE(webrtc::kTraceModuleCall, webrtc::kTraceVideoCapture, _id, "CaptureDelay %d",
(int) _captureDelay);
CriticalSectionScoped cs(_apiCs);
return _setCaptureDelay;
}
WebRtc_Word32 VideoCaptureImpl::DeliverCapturedFrame(VideoFrame& captureFrame,
WebRtc_Word32 width, WebRtc_Word32 height, WebRtc_Word64 capture_time,
VideoCodecType codec_type) {
UpdateFrameCount();// frame count used for local frame rate callback.
_startImageFrameIntervall = 0; // prevent the start image to be displayed.
const bool callOnCaptureDelayChanged = _setCaptureDelay != _captureDelay;
// Capture delay changed
if (_setCaptureDelay != _captureDelay) {
_setCaptureDelay = _captureDelay;
}
// Set the capture time
if (capture_time != 0) {
captureFrame.SetRenderTime(capture_time);
}
else {
captureFrame.SetRenderTime(TickTime::MillisecondTimestamp());
}
if (captureFrame.RenderTimeMs() == last_capture_time_) {
// We don't allow the same capture time for two frames, drop this one.
return -1;
}
last_capture_time_ = captureFrame.RenderTimeMs();
captureFrame.SetHeight(height);
captureFrame.SetWidth(width);
if (_dataCallBack) {
if (callOnCaptureDelayChanged) {
_dataCallBack->OnCaptureDelayChanged(_id, _captureDelay);
}
_dataCallBack->OnIncomingCapturedFrame(_id, captureFrame, codec_type);
}
return 0;
}
WebRtc_Word32 VideoCaptureImpl::IncomingFrame(WebRtc_UWord8* videoFrame,
WebRtc_Word32 videoFrameLength,
const VideoCaptureCapability& frameInfo,
WebRtc_Word64 captureTime/*=0*/)
{
WEBRTC_TRACE(webrtc::kTraceStream, webrtc::kTraceVideoCapture, _id,
"IncomingFrame width %d, height %d", (int) frameInfo.width,
(int) frameInfo.height);
TickTime startProcessTime = TickTime::Now();
CriticalSectionScoped cs(_callBackCs);
const WebRtc_Word32 width = frameInfo.width;
const WebRtc_Word32 height = frameInfo.height;
if (frameInfo.codecType == kVideoCodecUnknown) // None encoded. Convert to I420.
{
const VideoType commonVideoType =
RawVideoTypeToCommonVideoVideoType(frameInfo.rawType);
int size = CalcBufferSize(commonVideoType, width, height);
if (size != videoFrameLength)
{
WEBRTC_TRACE(webrtc::kTraceError, webrtc::kTraceVideoCapture, _id,
"Wrong incoming frame length.");
return -1;
}
// Allocate I420 buffer
int requiredLength = CalcBufferSize(kI420, width, height);
_captureFrame.VerifyAndAllocate(requiredLength);
if (!_captureFrame.Buffer())
{
WEBRTC_TRACE(webrtc::kTraceError, webrtc::kTraceVideoCapture, _id,
"Failed to allocate frame buffer.");
return -1;
}
memset(_captureFrame.Buffer(), 0, _captureFrame.Size());
int dstStride = width; // Keeping stride = width for I420 destination.
const int conversionResult = ConvertToI420(commonVideoType,
videoFrame,
0, 0, // No cropping
width, height,
0, // Ignored for non-JPG.
width, height, dstStride,
_rotateFrame,
_captureFrame.Buffer());
if (conversionResult < 0)
{
WEBRTC_TRACE(webrtc::kTraceError, webrtc::kTraceVideoCapture, _id,
"Failed to convert capture frame from type %d to I420",
frameInfo.rawType);
return -1;
}
_captureFrame.SetLength(requiredLength);
}
else // Encoded format
{
if (_captureFrame.CopyFrame(videoFrameLength, videoFrame) != 0)
{
WEBRTC_TRACE(webrtc::kTraceError, webrtc::kTraceVideoCapture, _id,
"Failed to copy captured frame of length %d", (int) videoFrameLength);
}
}
DeliverCapturedFrame(_captureFrame, width, height, captureTime, frameInfo.codecType);
const WebRtc_UWord32 processTime =
(WebRtc_UWord32)(TickTime::Now() - startProcessTime).Milliseconds();
if (processTime > 10) // If the process time is too long MJPG will not work well.
{
WEBRTC_TRACE(webrtc::kTraceWarning, webrtc::kTraceVideoCapture, _id,
"Too long processing time of Incoming frame: %ums",
(unsigned int) processTime);
}
return 0;
}
WebRtc_Word32 VideoCaptureImpl::IncomingFrameI420(
const VideoFrameI420& video_frame, WebRtc_Word64 captureTime) {
CriticalSectionScoped cs(_callBackCs);
// Allocate I420 buffer
int frame_size = CalcBufferSize(kI420,
video_frame.width,
video_frame.height);
_captureFrame.VerifyAndAllocate(frame_size);
if (!_captureFrame.Buffer()) {
WEBRTC_TRACE(webrtc::kTraceError, webrtc::kTraceVideoCapture, _id,
"Failed to allocate frame buffer.");
return -1;
}
// Copy planes to the _captureFrame
int y_width = video_frame.width;
int uv_width = video_frame.width / 2;
int y_rows = video_frame.height;
int uv_rows = video_frame.height / 2; // I420
unsigned char* current_pointer = _captureFrame.Buffer();
unsigned char* y_plane = video_frame.y_plane;
unsigned char* u_plane = video_frame.u_plane;
unsigned char* v_plane = video_frame.v_plane;
// Copy Y
for (int i = 0; i < y_rows; ++i) {
memcpy(current_pointer, y_plane, y_width);
current_pointer += video_frame.y_pitch;
y_plane += video_frame.y_pitch;
}
// Copy U
for (int i = 0; i < uv_rows; ++i) {
memcpy(current_pointer, u_plane, uv_width);
current_pointer += video_frame.u_pitch;
u_plane += video_frame.u_pitch;
}
// Copy V
for (int i = 0; i < uv_rows; ++i) {
memcpy(current_pointer, v_plane, uv_width);
current_pointer += video_frame.v_pitch;
v_plane += video_frame.v_pitch;
}
_captureFrame.SetLength(frame_size);
DeliverCapturedFrame(_captureFrame,
video_frame.width,
video_frame.height,
captureTime,
kVideoCodecUnknown);
return 0;
}
WebRtc_Word32 VideoCaptureImpl::SetCaptureRotation(VideoCaptureRotation rotation)
{
CriticalSectionScoped cs(_apiCs);
CriticalSectionScoped cs2(_callBackCs);
switch (rotation)
{
case kCameraRotate0:
_rotateFrame = kRotateNone;
break;
case kCameraRotate90:
_rotateFrame = kRotate90;
break;
case kCameraRotate180:
_rotateFrame = kRotate180;
break;
case kCameraRotate270:
_rotateFrame = kRotate270;
break;
}
return 0;
}
WebRtc_Word32 VideoCaptureImpl::StartSendImage(const VideoFrame& videoFrame,
WebRtc_Word32 frameRate)
{
WEBRTC_TRACE(webrtc::kTraceModuleCall, webrtc::kTraceVideoCapture, _id,
"StartSendImage, frameRate %d", (int) frameRate);
CriticalSectionScoped cs(_apiCs);
CriticalSectionScoped cs2(_callBackCs);
if (frameRate < 1 || frameRate > kMaxFrameRate)
{
WEBRTC_TRACE(webrtc::kTraceError, webrtc::kTraceVideoCapture, _id,
"StartSendImage Invalid parameter. frameRate %d", (int) frameRate);
return -1;;
}
_startImage.CopyFrame(videoFrame);
_startImageFrameIntervall = 1000 / frameRate;
_lastSentStartImageTime = TickTime::Now();
return 0;
}
WebRtc_Word32 VideoCaptureImpl::StopSendImage()
{
CriticalSectionScoped cs(_apiCs);
CriticalSectionScoped cs2(_callBackCs);
_startImageFrameIntervall = 0;
return 0;
}
WebRtc_Word32 VideoCaptureImpl::EnableFrameRateCallback(const bool enable)
{
CriticalSectionScoped cs(_apiCs);
CriticalSectionScoped cs2(_callBackCs);
_frameRateCallBack = enable;
if (enable)
{
_lastFrameRateCallbackTime = TickTime::Now();
}
return 0;
}
WebRtc_Word32 VideoCaptureImpl::EnableNoPictureAlarm(const bool enable)
{
CriticalSectionScoped cs(_apiCs);
CriticalSectionScoped cs2(_callBackCs);
_noPictureAlarmCallBack = enable;
return 0;
}
void VideoCaptureImpl::UpdateFrameCount()
{
if (_incomingFrameTimes[0].MicrosecondTimestamp() == 0)
{
// first no shift
}
else
{
// shift
for (int i = (kFrameRateCountHistorySize - 2); i >= 0; i--)
{
_incomingFrameTimes[i + 1] = _incomingFrameTimes[i];
}
}
_incomingFrameTimes[0] = TickTime::Now();
}
WebRtc_UWord32 VideoCaptureImpl::CalculateFrameRate(const TickTime& now)
{
WebRtc_Word32 num = 0;
WebRtc_Word32 nrOfFrames = 0;
for (num = 1; num < (kFrameRateCountHistorySize - 1); num++)
{
if (_incomingFrameTimes[num].Ticks() <= 0
|| (now - _incomingFrameTimes[num]).Milliseconds() > kFrameRateHistoryWindowMs) // don't use data older than 2sec
{
break;
}
else
{
nrOfFrames++;
}
}
if (num > 1)
{
WebRtc_Word64 diff = (now - _incomingFrameTimes[num - 1]).Milliseconds();
if (diff > 0)
{
return WebRtc_UWord32((nrOfFrames * 1000.0f / diff) + 0.5f);
}
}
return nrOfFrames;
}
} // namespace videocapturemodule
} // namespace webrtc