RTCOpenGLVideoRenderer is currently shared between RTCEAGLVideoView and RTCNSGLVideoView, and contains some iOS specific code and some Mac specific code. We plan on adding an interface for injectable OpenGL ES shaders to RTCEAGLVideoView, and it's easier if RTCEAGLVideoView and RTCNSGLVideoView have separate code paths. This CL removes RTCOpenGLVideoRenderer and inlines the necessary code in RTCEAGLVideoView and RTCNSGLVideoView. BUG=webrtc:7473 Review-Url: https://codereview.webrtc.org/2812613003 Cr-Commit-Position: refs/heads/master@{#17693}
174 lines
4.7 KiB
Objective-C
174 lines
4.7 KiB
Objective-C
/*
|
|
* Copyright 2015 The WebRTC project authors. All Rights Reserved.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license
|
|
* that can be found in the LICENSE file in the root of the source
|
|
* tree. An additional intellectual property rights grant can be found
|
|
* in the file PATENTS. All contributing project authors may
|
|
* be found in the AUTHORS file in the root of the source tree.
|
|
*/
|
|
|
|
#import <Foundation/Foundation.h>
|
|
|
|
#if !TARGET_OS_IPHONE
|
|
|
|
#import "WebRTC/RTCNSGLVideoView.h"
|
|
|
|
#import <AppKit/NSOpenGL.h>
|
|
#import <CoreVideo/CVDisplayLink.h>
|
|
#import <OpenGL/gl3.h>
|
|
|
|
#import "RTCShader+Private.h"
|
|
#import "WebRTC/RTCLogging.h"
|
|
#import "WebRTC/RTCVideoFrame.h"
|
|
|
|
@interface RTCNSGLVideoView ()
|
|
// |videoFrame| is set when we receive a frame from a worker thread and is read
|
|
// from the display link callback so atomicity is required.
|
|
@property(atomic, strong) RTCVideoFrame *videoFrame;
|
|
@property(atomic, strong) id<RTCShader> i420Shader;
|
|
|
|
- (void)drawFrame;
|
|
@end
|
|
|
|
static CVReturn OnDisplayLinkFired(CVDisplayLinkRef displayLink,
|
|
const CVTimeStamp *now,
|
|
const CVTimeStamp *outputTime,
|
|
CVOptionFlags flagsIn,
|
|
CVOptionFlags *flagsOut,
|
|
void *displayLinkContext) {
|
|
RTCNSGLVideoView *view = (__bridge RTCNSGLVideoView *)displayLinkContext;
|
|
[view drawFrame];
|
|
return kCVReturnSuccess;
|
|
}
|
|
|
|
@implementation RTCNSGLVideoView {
|
|
CVDisplayLinkRef _displayLink;
|
|
RTCVideoFrame *_lastDrawnFrame;
|
|
}
|
|
|
|
@synthesize delegate = _delegate;
|
|
@synthesize videoFrame = _videoFrame;
|
|
@synthesize i420Shader = _i420Shader;
|
|
|
|
- (void)dealloc {
|
|
[self teardownDisplayLink];
|
|
}
|
|
|
|
- (void)drawRect:(NSRect)rect {
|
|
[self drawFrame];
|
|
}
|
|
|
|
- (void)reshape {
|
|
[super reshape];
|
|
NSRect frame = [self frame];
|
|
CGLLockContext([[self openGLContext] CGLContextObj]);
|
|
glViewport(0, 0, frame.size.width, frame.size.height);
|
|
CGLUnlockContext([[self openGLContext] CGLContextObj]);
|
|
}
|
|
|
|
- (void)lockFocus {
|
|
NSOpenGLContext *context = [self openGLContext];
|
|
[super lockFocus];
|
|
if ([context view] != self) {
|
|
[context setView:self];
|
|
}
|
|
[context makeCurrentContext];
|
|
}
|
|
|
|
- (void)prepareOpenGL {
|
|
[super prepareOpenGL];
|
|
[self ensureGLContext];
|
|
glDisable(GL_DITHER);
|
|
[self setupDisplayLink];
|
|
}
|
|
|
|
- (void)clearGLContext {
|
|
[self ensureGLContext];
|
|
self.i420Shader = nil;
|
|
[super clearGLContext];
|
|
}
|
|
|
|
#pragma mark - RTCVideoRenderer
|
|
|
|
// These methods may be called on non-main thread.
|
|
- (void)setSize:(CGSize)size {
|
|
dispatch_async(dispatch_get_main_queue(), ^{
|
|
[self.delegate videoView:self didChangeVideoSize:size];
|
|
});
|
|
}
|
|
|
|
- (void)renderFrame:(RTCVideoFrame *)frame {
|
|
self.videoFrame = frame;
|
|
}
|
|
|
|
#pragma mark - Private
|
|
|
|
- (void)drawFrame {
|
|
RTCVideoFrame *frame = self.videoFrame;
|
|
if (!frame || frame == _lastDrawnFrame) {
|
|
return;
|
|
}
|
|
// This method may be called from CVDisplayLink callback which isn't on the
|
|
// main thread so we have to lock the GL context before drawing.
|
|
NSOpenGLContext *context = [self openGLContext];
|
|
CGLLockContext([context CGLContextObj]);
|
|
|
|
[self ensureGLContext];
|
|
glClear(GL_COLOR_BUFFER_BIT);
|
|
|
|
// Rendering native CVPixelBuffer is not supported on OS X.
|
|
frame = [frame newI420VideoFrame];
|
|
if (!self.i420Shader) {
|
|
self.i420Shader = [[RTCI420Shader alloc] initWithContext:context];
|
|
}
|
|
if (self.i420Shader && [self.i420Shader drawFrame:frame]) {
|
|
[context flushBuffer];
|
|
_lastDrawnFrame = frame;
|
|
} else {
|
|
RTCLog(@"Failed to draw frame.");
|
|
}
|
|
CGLUnlockContext([context CGLContextObj]);
|
|
}
|
|
|
|
- (void)setupDisplayLink {
|
|
if (_displayLink) {
|
|
return;
|
|
}
|
|
// Synchronize buffer swaps with vertical refresh rate.
|
|
GLint swapInt = 1;
|
|
[[self openGLContext] setValues:&swapInt forParameter:NSOpenGLCPSwapInterval];
|
|
|
|
// Create display link.
|
|
CVDisplayLinkCreateWithActiveCGDisplays(&_displayLink);
|
|
CVDisplayLinkSetOutputCallback(_displayLink,
|
|
&OnDisplayLinkFired,
|
|
(__bridge void *)self);
|
|
// Set the display link for the current renderer.
|
|
CGLContextObj cglContext = [[self openGLContext] CGLContextObj];
|
|
CGLPixelFormatObj cglPixelFormat = [[self pixelFormat] CGLPixelFormatObj];
|
|
CVDisplayLinkSetCurrentCGDisplayFromOpenGLContext(
|
|
_displayLink, cglContext, cglPixelFormat);
|
|
CVDisplayLinkStart(_displayLink);
|
|
}
|
|
|
|
- (void)teardownDisplayLink {
|
|
if (!_displayLink) {
|
|
return;
|
|
}
|
|
CVDisplayLinkRelease(_displayLink);
|
|
_displayLink = NULL;
|
|
}
|
|
|
|
- (void)ensureGLContext {
|
|
NSOpenGLContext* context = [self openGLContext];
|
|
NSAssert(context, @"context shouldn't be nil");
|
|
if ([NSOpenGLContext currentContext] != context) {
|
|
[context makeCurrentContext];
|
|
}
|
|
}
|
|
|
|
@end
|
|
|
|
#endif // !TARGET_OS_IPHONE
|