mirror of
https://repo.dec05eba.com/gpu-screen-recorder
synced 2026-05-05 06:20:43 +09:00
Name capture/encoder variable self
This commit is contained in:
@@ -16,6 +16,7 @@ src = [
|
|||||||
'src/encoder/video/video.c',
|
'src/encoder/video/video.c',
|
||||||
'src/encoder/video/cuda.c',
|
'src/encoder/video/cuda.c',
|
||||||
'src/encoder/video/vaapi.c',
|
'src/encoder/video/vaapi.c',
|
||||||
|
'src/encoder/video/vulkan.c',
|
||||||
'src/encoder/video/software.c',
|
'src/encoder/video/software.c',
|
||||||
'src/egl.c',
|
'src/egl.c',
|
||||||
'src/cuda.c',
|
'src/cuda.c',
|
||||||
|
|||||||
@@ -102,7 +102,7 @@ static void set_func_ptr(void **dst, void *src) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
static bool gsr_capture_nvfbc_load_library(gsr_capture *cap) {
|
static bool gsr_capture_nvfbc_load_library(gsr_capture *cap) {
|
||||||
gsr_capture_nvfbc *cap_nvfbc = cap->priv;
|
gsr_capture_nvfbc *self = cap->priv;
|
||||||
|
|
||||||
dlerror(); /* clear */
|
dlerror(); /* clear */
|
||||||
void *lib = dlopen("libnvidia-fbc.so.1", RTLD_LAZY);
|
void *lib = dlopen("libnvidia-fbc.so.1", RTLD_LAZY);
|
||||||
@@ -111,23 +111,23 @@ static bool gsr_capture_nvfbc_load_library(gsr_capture *cap) {
|
|||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
set_func_ptr((void**)&cap_nvfbc->nv_fbc_create_instance, dlsym(lib, "NvFBCCreateInstance"));
|
set_func_ptr((void**)&self->nv_fbc_create_instance, dlsym(lib, "NvFBCCreateInstance"));
|
||||||
if(!cap_nvfbc->nv_fbc_create_instance) {
|
if(!self->nv_fbc_create_instance) {
|
||||||
fprintf(stderr, "gsr error: unable to resolve symbol 'NvFBCCreateInstance'\n");
|
fprintf(stderr, "gsr error: unable to resolve symbol 'NvFBCCreateInstance'\n");
|
||||||
dlclose(lib);
|
dlclose(lib);
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
memset(&cap_nvfbc->nv_fbc_function_list, 0, sizeof(cap_nvfbc->nv_fbc_function_list));
|
memset(&self->nv_fbc_function_list, 0, sizeof(self->nv_fbc_function_list));
|
||||||
cap_nvfbc->nv_fbc_function_list.dwVersion = NVFBC_VERSION;
|
self->nv_fbc_function_list.dwVersion = NVFBC_VERSION;
|
||||||
NVFBCSTATUS status = cap_nvfbc->nv_fbc_create_instance(&cap_nvfbc->nv_fbc_function_list);
|
NVFBCSTATUS status = self->nv_fbc_create_instance(&self->nv_fbc_function_list);
|
||||||
if(status != NVFBC_SUCCESS) {
|
if(status != NVFBC_SUCCESS) {
|
||||||
fprintf(stderr, "gsr error: failed to create NvFBC instance (status: %d)\n", status);
|
fprintf(stderr, "gsr error: failed to create NvFBC instance (status: %d)\n", status);
|
||||||
dlclose(lib);
|
dlclose(lib);
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
cap_nvfbc->library = lib;
|
self->library = lib;
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -153,64 +153,64 @@ static void set_vertical_sync_enabled(gsr_egl *egl, int enabled) {
|
|||||||
fprintf(stderr, "gsr warning: setting vertical sync failed\n");
|
fprintf(stderr, "gsr warning: setting vertical sync failed\n");
|
||||||
}
|
}
|
||||||
|
|
||||||
static void gsr_capture_nvfbc_destroy_session(gsr_capture_nvfbc *cap_nvfbc) {
|
static void gsr_capture_nvfbc_destroy_session(gsr_capture_nvfbc *self) {
|
||||||
if(cap_nvfbc->fbc_handle_created && cap_nvfbc->capture_session_created) {
|
if(self->fbc_handle_created && self->capture_session_created) {
|
||||||
NVFBC_DESTROY_CAPTURE_SESSION_PARAMS destroy_capture_params;
|
NVFBC_DESTROY_CAPTURE_SESSION_PARAMS destroy_capture_params;
|
||||||
memset(&destroy_capture_params, 0, sizeof(destroy_capture_params));
|
memset(&destroy_capture_params, 0, sizeof(destroy_capture_params));
|
||||||
destroy_capture_params.dwVersion = NVFBC_DESTROY_CAPTURE_SESSION_PARAMS_VER;
|
destroy_capture_params.dwVersion = NVFBC_DESTROY_CAPTURE_SESSION_PARAMS_VER;
|
||||||
cap_nvfbc->nv_fbc_function_list.nvFBCDestroyCaptureSession(cap_nvfbc->nv_fbc_handle, &destroy_capture_params);
|
self->nv_fbc_function_list.nvFBCDestroyCaptureSession(self->nv_fbc_handle, &destroy_capture_params);
|
||||||
cap_nvfbc->capture_session_created = false;
|
self->capture_session_created = false;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static void gsr_capture_nvfbc_destroy_handle(gsr_capture_nvfbc *cap_nvfbc) {
|
static void gsr_capture_nvfbc_destroy_handle(gsr_capture_nvfbc *self) {
|
||||||
if(cap_nvfbc->fbc_handle_created) {
|
if(self->fbc_handle_created) {
|
||||||
NVFBC_DESTROY_HANDLE_PARAMS destroy_params;
|
NVFBC_DESTROY_HANDLE_PARAMS destroy_params;
|
||||||
memset(&destroy_params, 0, sizeof(destroy_params));
|
memset(&destroy_params, 0, sizeof(destroy_params));
|
||||||
destroy_params.dwVersion = NVFBC_DESTROY_HANDLE_PARAMS_VER;
|
destroy_params.dwVersion = NVFBC_DESTROY_HANDLE_PARAMS_VER;
|
||||||
cap_nvfbc->nv_fbc_function_list.nvFBCDestroyHandle(cap_nvfbc->nv_fbc_handle, &destroy_params);
|
self->nv_fbc_function_list.nvFBCDestroyHandle(self->nv_fbc_handle, &destroy_params);
|
||||||
cap_nvfbc->fbc_handle_created = false;
|
self->fbc_handle_created = false;
|
||||||
cap_nvfbc->nv_fbc_handle = 0;
|
self->nv_fbc_handle = 0;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static void gsr_capture_nvfbc_destroy_session_and_handle(gsr_capture_nvfbc *cap_nvfbc) {
|
static void gsr_capture_nvfbc_destroy_session_and_handle(gsr_capture_nvfbc *self) {
|
||||||
gsr_capture_nvfbc_destroy_session(cap_nvfbc);
|
gsr_capture_nvfbc_destroy_session(self);
|
||||||
gsr_capture_nvfbc_destroy_handle(cap_nvfbc);
|
gsr_capture_nvfbc_destroy_handle(self);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int gsr_capture_nvfbc_setup_handle(gsr_capture_nvfbc *cap_nvfbc) {
|
static int gsr_capture_nvfbc_setup_handle(gsr_capture_nvfbc *self) {
|
||||||
NVFBCSTATUS status;
|
NVFBCSTATUS status;
|
||||||
|
|
||||||
NVFBC_CREATE_HANDLE_PARAMS create_params;
|
NVFBC_CREATE_HANDLE_PARAMS create_params;
|
||||||
memset(&create_params, 0, sizeof(create_params));
|
memset(&create_params, 0, sizeof(create_params));
|
||||||
create_params.dwVersion = NVFBC_CREATE_HANDLE_PARAMS_VER;
|
create_params.dwVersion = NVFBC_CREATE_HANDLE_PARAMS_VER;
|
||||||
create_params.bExternallyManagedContext = NVFBC_TRUE;
|
create_params.bExternallyManagedContext = NVFBC_TRUE;
|
||||||
create_params.glxCtx = cap_nvfbc->params.egl->glx_context;
|
create_params.glxCtx = self->params.egl->glx_context;
|
||||||
create_params.glxFBConfig = cap_nvfbc->params.egl->glx_fb_config;
|
create_params.glxFBConfig = self->params.egl->glx_fb_config;
|
||||||
|
|
||||||
status = cap_nvfbc->nv_fbc_function_list.nvFBCCreateHandle(&cap_nvfbc->nv_fbc_handle, &create_params);
|
status = self->nv_fbc_function_list.nvFBCCreateHandle(&self->nv_fbc_handle, &create_params);
|
||||||
if(status != NVFBC_SUCCESS) {
|
if(status != NVFBC_SUCCESS) {
|
||||||
// Reverse engineering for interoperability
|
// Reverse engineering for interoperability
|
||||||
const uint8_t enable_key[] = { 0xac, 0x10, 0xc9, 0x2e, 0xa5, 0xe6, 0x87, 0x4f, 0x8f, 0x4b, 0xf4, 0x61, 0xf8, 0x56, 0x27, 0xe9 };
|
const uint8_t enable_key[] = { 0xac, 0x10, 0xc9, 0x2e, 0xa5, 0xe6, 0x87, 0x4f, 0x8f, 0x4b, 0xf4, 0x61, 0xf8, 0x56, 0x27, 0xe9 };
|
||||||
create_params.privateData = enable_key;
|
create_params.privateData = enable_key;
|
||||||
create_params.privateDataSize = 16;
|
create_params.privateDataSize = 16;
|
||||||
|
|
||||||
status = cap_nvfbc->nv_fbc_function_list.nvFBCCreateHandle(&cap_nvfbc->nv_fbc_handle, &create_params);
|
status = self->nv_fbc_function_list.nvFBCCreateHandle(&self->nv_fbc_handle, &create_params);
|
||||||
if(status != NVFBC_SUCCESS) {
|
if(status != NVFBC_SUCCESS) {
|
||||||
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: %s\n", cap_nvfbc->nv_fbc_function_list.nvFBCGetLastErrorStr(cap_nvfbc->nv_fbc_handle));
|
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: %s\n", self->nv_fbc_function_list.nvFBCGetLastErrorStr(self->nv_fbc_handle));
|
||||||
goto error_cleanup;
|
goto error_cleanup;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
cap_nvfbc->fbc_handle_created = true;
|
self->fbc_handle_created = true;
|
||||||
|
|
||||||
NVFBC_GET_STATUS_PARAMS status_params;
|
NVFBC_GET_STATUS_PARAMS status_params;
|
||||||
memset(&status_params, 0, sizeof(status_params));
|
memset(&status_params, 0, sizeof(status_params));
|
||||||
status_params.dwVersion = NVFBC_GET_STATUS_PARAMS_VER;
|
status_params.dwVersion = NVFBC_GET_STATUS_PARAMS_VER;
|
||||||
|
|
||||||
status = cap_nvfbc->nv_fbc_function_list.nvFBCGetStatus(cap_nvfbc->nv_fbc_handle, &status_params);
|
status = self->nv_fbc_function_list.nvFBCGetStatus(self->nv_fbc_handle, &status_params);
|
||||||
if(status != NVFBC_SUCCESS) {
|
if(status != NVFBC_SUCCESS) {
|
||||||
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: %s\n", cap_nvfbc->nv_fbc_function_list.nvFBCGetLastErrorStr(cap_nvfbc->nv_fbc_handle));
|
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: %s\n", self->nv_fbc_function_list.nvFBCGetLastErrorStr(self->nv_fbc_handle));
|
||||||
goto error_cleanup;
|
goto error_cleanup;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -219,10 +219,10 @@ static int gsr_capture_nvfbc_setup_handle(gsr_capture_nvfbc *cap_nvfbc) {
|
|||||||
goto error_cleanup;
|
goto error_cleanup;
|
||||||
}
|
}
|
||||||
|
|
||||||
cap_nvfbc->tracking_width = XWidthOfScreen(DefaultScreenOfDisplay(cap_nvfbc->params.egl->x11.dpy));
|
self->tracking_width = XWidthOfScreen(DefaultScreenOfDisplay(self->params.egl->x11.dpy));
|
||||||
cap_nvfbc->tracking_height = XHeightOfScreen(DefaultScreenOfDisplay(cap_nvfbc->params.egl->x11.dpy));
|
self->tracking_height = XHeightOfScreen(DefaultScreenOfDisplay(self->params.egl->x11.dpy));
|
||||||
cap_nvfbc->tracking_type = strcmp(cap_nvfbc->params.display_to_capture, "screen") == 0 ? NVFBC_TRACKING_SCREEN : NVFBC_TRACKING_OUTPUT;
|
self->tracking_type = strcmp(self->params.display_to_capture, "screen") == 0 ? NVFBC_TRACKING_SCREEN : NVFBC_TRACKING_OUTPUT;
|
||||||
if(cap_nvfbc->tracking_type == NVFBC_TRACKING_OUTPUT) {
|
if(self->tracking_type == NVFBC_TRACKING_OUTPUT) {
|
||||||
if(!status_params.bXRandRAvailable) {
|
if(!status_params.bXRandRAvailable) {
|
||||||
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: the xrandr extension is not available\n");
|
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: the xrandr extension is not available\n");
|
||||||
goto error_cleanup;
|
goto error_cleanup;
|
||||||
@@ -233,9 +233,9 @@ static int gsr_capture_nvfbc_setup_handle(gsr_capture_nvfbc *cap_nvfbc) {
|
|||||||
goto error_cleanup;
|
goto error_cleanup;
|
||||||
}
|
}
|
||||||
|
|
||||||
cap_nvfbc->output_id = get_output_id_from_display_name(status_params.outputs, status_params.dwOutputNum, cap_nvfbc->params.display_to_capture, &cap_nvfbc->tracking_width, &cap_nvfbc->tracking_height);
|
self->output_id = get_output_id_from_display_name(status_params.outputs, status_params.dwOutputNum, self->params.display_to_capture, &self->tracking_width, &self->tracking_height);
|
||||||
if(cap_nvfbc->output_id == 0) {
|
if(self->output_id == 0) {
|
||||||
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: display '%s' not found\n", cap_nvfbc->params.display_to_capture);
|
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: display '%s' not found\n", self->params.display_to_capture);
|
||||||
goto error_cleanup;
|
goto error_cleanup;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -243,83 +243,83 @@ static int gsr_capture_nvfbc_setup_handle(gsr_capture_nvfbc *cap_nvfbc) {
|
|||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
error_cleanup:
|
error_cleanup:
|
||||||
gsr_capture_nvfbc_destroy_session_and_handle(cap_nvfbc);
|
gsr_capture_nvfbc_destroy_session_and_handle(self);
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int gsr_capture_nvfbc_setup_session(gsr_capture_nvfbc *cap_nvfbc) {
|
static int gsr_capture_nvfbc_setup_session(gsr_capture_nvfbc *self) {
|
||||||
NVFBC_CREATE_CAPTURE_SESSION_PARAMS create_capture_params;
|
NVFBC_CREATE_CAPTURE_SESSION_PARAMS create_capture_params;
|
||||||
memset(&create_capture_params, 0, sizeof(create_capture_params));
|
memset(&create_capture_params, 0, sizeof(create_capture_params));
|
||||||
create_capture_params.dwVersion = NVFBC_CREATE_CAPTURE_SESSION_PARAMS_VER;
|
create_capture_params.dwVersion = NVFBC_CREATE_CAPTURE_SESSION_PARAMS_VER;
|
||||||
create_capture_params.eCaptureType = NVFBC_CAPTURE_TO_GL;
|
create_capture_params.eCaptureType = NVFBC_CAPTURE_TO_GL;
|
||||||
create_capture_params.bWithCursor = (!cap_nvfbc->params.direct_capture || cap_nvfbc->supports_direct_cursor) ? NVFBC_TRUE : NVFBC_FALSE;
|
create_capture_params.bWithCursor = (!self->params.direct_capture || self->supports_direct_cursor) ? NVFBC_TRUE : NVFBC_FALSE;
|
||||||
if(!cap_nvfbc->params.record_cursor)
|
if(!self->params.record_cursor)
|
||||||
create_capture_params.bWithCursor = false;
|
create_capture_params.bWithCursor = false;
|
||||||
if(cap_nvfbc->capture_region)
|
if(self->capture_region)
|
||||||
create_capture_params.captureBox = (NVFBC_BOX){ cap_nvfbc->x, cap_nvfbc->y, cap_nvfbc->width, cap_nvfbc->height };
|
create_capture_params.captureBox = (NVFBC_BOX){ self->x, self->y, self->width, self->height };
|
||||||
create_capture_params.eTrackingType = cap_nvfbc->tracking_type;
|
create_capture_params.eTrackingType = self->tracking_type;
|
||||||
create_capture_params.dwSamplingRateMs = (uint32_t)ceilf(1000.0f / (float)cap_nvfbc->params.fps);
|
create_capture_params.dwSamplingRateMs = (uint32_t)ceilf(1000.0f / (float)self->params.fps);
|
||||||
create_capture_params.bAllowDirectCapture = cap_nvfbc->params.direct_capture ? NVFBC_TRUE : NVFBC_FALSE;
|
create_capture_params.bAllowDirectCapture = self->params.direct_capture ? NVFBC_TRUE : NVFBC_FALSE;
|
||||||
create_capture_params.bPushModel = cap_nvfbc->params.direct_capture ? NVFBC_TRUE : NVFBC_FALSE;
|
create_capture_params.bPushModel = self->params.direct_capture ? NVFBC_TRUE : NVFBC_FALSE;
|
||||||
create_capture_params.bDisableAutoModesetRecovery = true;
|
create_capture_params.bDisableAutoModesetRecovery = true;
|
||||||
if(cap_nvfbc->tracking_type == NVFBC_TRACKING_OUTPUT)
|
if(self->tracking_type == NVFBC_TRACKING_OUTPUT)
|
||||||
create_capture_params.dwOutputId = cap_nvfbc->output_id;
|
create_capture_params.dwOutputId = self->output_id;
|
||||||
|
|
||||||
NVFBCSTATUS status = cap_nvfbc->nv_fbc_function_list.nvFBCCreateCaptureSession(cap_nvfbc->nv_fbc_handle, &create_capture_params);
|
NVFBCSTATUS status = self->nv_fbc_function_list.nvFBCCreateCaptureSession(self->nv_fbc_handle, &create_capture_params);
|
||||||
if(status != NVFBC_SUCCESS) {
|
if(status != NVFBC_SUCCESS) {
|
||||||
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: %s\n", cap_nvfbc->nv_fbc_function_list.nvFBCGetLastErrorStr(cap_nvfbc->nv_fbc_handle));
|
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: %s\n", self->nv_fbc_function_list.nvFBCGetLastErrorStr(self->nv_fbc_handle));
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
cap_nvfbc->capture_session_created = true;
|
self->capture_session_created = true;
|
||||||
|
|
||||||
memset(&cap_nvfbc->setup_params, 0, sizeof(cap_nvfbc->setup_params));
|
memset(&self->setup_params, 0, sizeof(self->setup_params));
|
||||||
cap_nvfbc->setup_params.dwVersion = NVFBC_TOGL_SETUP_PARAMS_VER;
|
self->setup_params.dwVersion = NVFBC_TOGL_SETUP_PARAMS_VER;
|
||||||
cap_nvfbc->setup_params.eBufferFormat = NVFBC_BUFFER_FORMAT_BGRA;
|
self->setup_params.eBufferFormat = NVFBC_BUFFER_FORMAT_BGRA;
|
||||||
|
|
||||||
status = cap_nvfbc->nv_fbc_function_list.nvFBCToGLSetUp(cap_nvfbc->nv_fbc_handle, &cap_nvfbc->setup_params);
|
status = self->nv_fbc_function_list.nvFBCToGLSetUp(self->nv_fbc_handle, &self->setup_params);
|
||||||
if(status != NVFBC_SUCCESS) {
|
if(status != NVFBC_SUCCESS) {
|
||||||
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: %s\n", cap_nvfbc->nv_fbc_function_list.nvFBCGetLastErrorStr(cap_nvfbc->nv_fbc_handle));
|
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: %s\n", self->nv_fbc_function_list.nvFBCGetLastErrorStr(self->nv_fbc_handle));
|
||||||
gsr_capture_nvfbc_destroy_session(cap_nvfbc);
|
gsr_capture_nvfbc_destroy_session(self);
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void gsr_capture_nvfbc_stop(gsr_capture_nvfbc *cap_nvfbc) {
|
static void gsr_capture_nvfbc_stop(gsr_capture_nvfbc *self) {
|
||||||
gsr_capture_nvfbc_destroy_session_and_handle(cap_nvfbc);
|
gsr_capture_nvfbc_destroy_session_and_handle(self);
|
||||||
if(cap_nvfbc->library) {
|
if(self->library) {
|
||||||
dlclose(cap_nvfbc->library);
|
dlclose(self->library);
|
||||||
cap_nvfbc->library = NULL;
|
self->library = NULL;
|
||||||
}
|
}
|
||||||
if(cap_nvfbc->params.display_to_capture) {
|
if(self->params.display_to_capture) {
|
||||||
free((void*)cap_nvfbc->params.display_to_capture);
|
free((void*)self->params.display_to_capture);
|
||||||
cap_nvfbc->params.display_to_capture = NULL;
|
self->params.display_to_capture = NULL;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static int gsr_capture_nvfbc_start(gsr_capture *cap, AVCodecContext *video_codec_context, AVFrame *frame) {
|
static int gsr_capture_nvfbc_start(gsr_capture *cap, AVCodecContext *video_codec_context, AVFrame *frame) {
|
||||||
gsr_capture_nvfbc *cap_nvfbc = cap->priv;
|
gsr_capture_nvfbc *self = cap->priv;
|
||||||
|
|
||||||
if(!gsr_capture_nvfbc_load_library(cap))
|
if(!gsr_capture_nvfbc_load_library(cap))
|
||||||
return -1;
|
return -1;
|
||||||
|
|
||||||
cap_nvfbc->x = max_int(cap_nvfbc->params.pos.x, 0);
|
self->x = max_int(self->params.pos.x, 0);
|
||||||
cap_nvfbc->y = max_int(cap_nvfbc->params.pos.y, 0);
|
self->y = max_int(self->params.pos.y, 0);
|
||||||
cap_nvfbc->width = max_int(cap_nvfbc->params.size.x, 0);
|
self->width = max_int(self->params.size.x, 0);
|
||||||
cap_nvfbc->height = max_int(cap_nvfbc->params.size.y, 0);
|
self->height = max_int(self->params.size.y, 0);
|
||||||
|
|
||||||
cap_nvfbc->capture_region = (cap_nvfbc->x > 0 || cap_nvfbc->y > 0 || cap_nvfbc->width > 0 || cap_nvfbc->height > 0);
|
self->capture_region = (self->x > 0 || self->y > 0 || self->width > 0 || self->height > 0);
|
||||||
|
|
||||||
cap_nvfbc->supports_direct_cursor = false;
|
self->supports_direct_cursor = false;
|
||||||
int driver_major_version = 0;
|
int driver_major_version = 0;
|
||||||
int driver_minor_version = 0;
|
int driver_minor_version = 0;
|
||||||
if(cap_nvfbc->params.direct_capture && get_driver_version(&driver_major_version, &driver_minor_version)) {
|
if(self->params.direct_capture && get_driver_version(&driver_major_version, &driver_minor_version)) {
|
||||||
fprintf(stderr, "Info: detected nvidia version: %d.%d\n", driver_major_version, driver_minor_version);
|
fprintf(stderr, "Info: detected nvidia version: %d.%d\n", driver_major_version, driver_minor_version);
|
||||||
|
|
||||||
// TODO:
|
// TODO:
|
||||||
if(version_at_least(driver_major_version, driver_minor_version, 515, 57) && version_less_than(driver_major_version, driver_minor_version, 520, 56)) {
|
if(version_at_least(driver_major_version, driver_minor_version, 515, 57) && version_less_than(driver_major_version, driver_minor_version, 520, 56)) {
|
||||||
cap_nvfbc->params.direct_capture = false;
|
self->params.direct_capture = false;
|
||||||
fprintf(stderr, "Warning: \"screen-direct\" has temporary been disabled as it causes stuttering with driver versions >= 515.57 and < 520.56. Please update your driver if possible. Capturing \"screen\" instead.\n");
|
fprintf(stderr, "Warning: \"screen-direct\" has temporary been disabled as it causes stuttering with driver versions >= 515.57 and < 520.56. Please update your driver if possible. Capturing \"screen\" instead.\n");
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -328,63 +328,63 @@ static int gsr_capture_nvfbc_start(gsr_capture *cap, AVCodecContext *video_codec
|
|||||||
/*
|
/*
|
||||||
if(direct_capture) {
|
if(direct_capture) {
|
||||||
if(version_at_least(driver_major_version, driver_minor_version, 515, 57))
|
if(version_at_least(driver_major_version, driver_minor_version, 515, 57))
|
||||||
cap_nvfbc->supports_direct_cursor = true;
|
self->supports_direct_cursor = true;
|
||||||
else
|
else
|
||||||
fprintf(stderr, "Info: capturing \"screen-direct\" but driver version appears to be less than 515.57. Disabling capture of cursor. Please update your driver if you want to capture your cursor or record \"screen\" instead.\n");
|
fprintf(stderr, "Info: capturing \"screen-direct\" but driver version appears to be less than 515.57. Disabling capture of cursor. Please update your driver if you want to capture your cursor or record \"screen\" instead.\n");
|
||||||
}
|
}
|
||||||
*/
|
*/
|
||||||
}
|
}
|
||||||
|
|
||||||
if(gsr_capture_nvfbc_setup_handle(cap_nvfbc) != 0) {
|
if(gsr_capture_nvfbc_setup_handle(self) != 0) {
|
||||||
goto error_cleanup;
|
goto error_cleanup;
|
||||||
}
|
}
|
||||||
|
|
||||||
if(gsr_capture_nvfbc_setup_session(cap_nvfbc) != 0) {
|
if(gsr_capture_nvfbc_setup_session(self) != 0) {
|
||||||
goto error_cleanup;
|
goto error_cleanup;
|
||||||
}
|
}
|
||||||
|
|
||||||
if(cap_nvfbc->capture_region) {
|
if(self->capture_region) {
|
||||||
video_codec_context->width = FFALIGN(cap_nvfbc->width, 2);
|
video_codec_context->width = FFALIGN(self->width, 2);
|
||||||
video_codec_context->height = FFALIGN(cap_nvfbc->height, 2);
|
video_codec_context->height = FFALIGN(self->height, 2);
|
||||||
} else {
|
} else {
|
||||||
video_codec_context->width = FFALIGN(cap_nvfbc->tracking_width, 2);
|
video_codec_context->width = FFALIGN(self->tracking_width, 2);
|
||||||
video_codec_context->height = FFALIGN(cap_nvfbc->tracking_height, 2);
|
video_codec_context->height = FFALIGN(self->tracking_height, 2);
|
||||||
}
|
}
|
||||||
|
|
||||||
frame->width = video_codec_context->width;
|
frame->width = video_codec_context->width;
|
||||||
frame->height = video_codec_context->height;
|
frame->height = video_codec_context->height;
|
||||||
|
|
||||||
/* Disable vsync */
|
/* Disable vsync */
|
||||||
set_vertical_sync_enabled(cap_nvfbc->params.egl, 0);
|
set_vertical_sync_enabled(self->params.egl, 0);
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
error_cleanup:
|
error_cleanup:
|
||||||
gsr_capture_nvfbc_stop(cap_nvfbc);
|
gsr_capture_nvfbc_stop(self);
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int gsr_capture_nvfbc_capture(gsr_capture *cap, AVFrame *frame, gsr_color_conversion *color_conversion) {
|
static int gsr_capture_nvfbc_capture(gsr_capture *cap, AVFrame *frame, gsr_color_conversion *color_conversion) {
|
||||||
gsr_capture_nvfbc *cap_nvfbc = cap->priv;
|
gsr_capture_nvfbc *self = cap->priv;
|
||||||
|
|
||||||
const double nvfbc_recreate_retry_time_seconds = 1.0;
|
const double nvfbc_recreate_retry_time_seconds = 1.0;
|
||||||
if(cap_nvfbc->nvfbc_needs_recreate) {
|
if(self->nvfbc_needs_recreate) {
|
||||||
const double now = clock_get_monotonic_seconds();
|
const double now = clock_get_monotonic_seconds();
|
||||||
if(now - cap_nvfbc->nvfbc_dead_start >= nvfbc_recreate_retry_time_seconds) {
|
if(now - self->nvfbc_dead_start >= nvfbc_recreate_retry_time_seconds) {
|
||||||
cap_nvfbc->nvfbc_dead_start = now;
|
self->nvfbc_dead_start = now;
|
||||||
gsr_capture_nvfbc_destroy_session_and_handle(cap_nvfbc);
|
gsr_capture_nvfbc_destroy_session_and_handle(self);
|
||||||
|
|
||||||
if(gsr_capture_nvfbc_setup_handle(cap_nvfbc) != 0) {
|
if(gsr_capture_nvfbc_setup_handle(self) != 0) {
|
||||||
fprintf(stderr, "gsr error: gsr_capture_nvfbc_capture failed to recreate nvfbc handle, trying again in %f second(s)\n", nvfbc_recreate_retry_time_seconds);
|
fprintf(stderr, "gsr error: gsr_capture_nvfbc_capture failed to recreate nvfbc handle, trying again in %f second(s)\n", nvfbc_recreate_retry_time_seconds);
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
if(gsr_capture_nvfbc_setup_session(cap_nvfbc) != 0) {
|
if(gsr_capture_nvfbc_setup_session(self) != 0) {
|
||||||
fprintf(stderr, "gsr error: gsr_capture_nvfbc_capture failed to recreate nvfbc session, trying again in %f second(s)\n", nvfbc_recreate_retry_time_seconds);
|
fprintf(stderr, "gsr error: gsr_capture_nvfbc_capture failed to recreate nvfbc session, trying again in %f second(s)\n", nvfbc_recreate_retry_time_seconds);
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
cap_nvfbc->nvfbc_needs_recreate = false;
|
self->nvfbc_needs_recreate = false;
|
||||||
} else {
|
} else {
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
@@ -400,24 +400,24 @@ static int gsr_capture_nvfbc_capture(gsr_capture *cap, AVFrame *frame, gsr_color
|
|||||||
grab_params.pFrameGrabInfo = &frame_info;
|
grab_params.pFrameGrabInfo = &frame_info;
|
||||||
grab_params.dwTimeoutMs = 0;
|
grab_params.dwTimeoutMs = 0;
|
||||||
|
|
||||||
NVFBCSTATUS status = cap_nvfbc->nv_fbc_function_list.nvFBCToGLGrabFrame(cap_nvfbc->nv_fbc_handle, &grab_params);
|
NVFBCSTATUS status = self->nv_fbc_function_list.nvFBCToGLGrabFrame(self->nv_fbc_handle, &grab_params);
|
||||||
if(status != NVFBC_SUCCESS) {
|
if(status != NVFBC_SUCCESS) {
|
||||||
fprintf(stderr, "gsr error: gsr_capture_nvfbc_capture failed: %s (%d), recreating session after %f second(s)\n", cap_nvfbc->nv_fbc_function_list.nvFBCGetLastErrorStr(cap_nvfbc->nv_fbc_handle), status, nvfbc_recreate_retry_time_seconds);
|
fprintf(stderr, "gsr error: gsr_capture_nvfbc_capture failed: %s (%d), recreating session after %f second(s)\n", self->nv_fbc_function_list.nvFBCGetLastErrorStr(self->nv_fbc_handle), status, nvfbc_recreate_retry_time_seconds);
|
||||||
cap_nvfbc->nvfbc_needs_recreate = true;
|
self->nvfbc_needs_recreate = true;
|
||||||
cap_nvfbc->nvfbc_dead_start = clock_get_monotonic_seconds();
|
self->nvfbc_dead_start = clock_get_monotonic_seconds();
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
cap_nvfbc->params.egl->glFlush();
|
self->params.egl->glFlush();
|
||||||
cap_nvfbc->params.egl->glFinish();
|
self->params.egl->glFinish();
|
||||||
|
|
||||||
gsr_color_conversion_draw(color_conversion, cap_nvfbc->setup_params.dwTextures[grab_params.dwTextureIndex],
|
gsr_color_conversion_draw(color_conversion, self->setup_params.dwTextures[grab_params.dwTextureIndex],
|
||||||
(vec2i){0, 0}, (vec2i){frame->width, frame->height},
|
(vec2i){0, 0}, (vec2i){frame->width, frame->height},
|
||||||
(vec2i){0, 0}, (vec2i){frame->width, frame->height},
|
(vec2i){0, 0}, (vec2i){frame->width, frame->height},
|
||||||
0.0f, false);
|
0.0f, false);
|
||||||
|
|
||||||
cap_nvfbc->params.egl->glFlush();
|
self->params.egl->glFlush();
|
||||||
cap_nvfbc->params.egl->glFinish();
|
self->params.egl->glFinish();
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
@@ -429,8 +429,9 @@ static gsr_source_color gsr_capture_nvfbc_get_source_color(gsr_capture *cap) {
|
|||||||
|
|
||||||
static void gsr_capture_nvfbc_destroy(gsr_capture *cap, AVCodecContext *video_codec_context) {
|
static void gsr_capture_nvfbc_destroy(gsr_capture *cap, AVCodecContext *video_codec_context) {
|
||||||
(void)video_codec_context;
|
(void)video_codec_context;
|
||||||
gsr_capture_nvfbc *cap_nvfbc = cap->priv;
|
gsr_capture_nvfbc *self = cap->priv;
|
||||||
gsr_capture_nvfbc_stop(cap_nvfbc);
|
gsr_capture_nvfbc_stop(self);
|
||||||
|
free(cap->priv);
|
||||||
free(cap);
|
free(cap);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -414,9 +414,9 @@ static void gsr_capture_portal_clear_damage(gsr_capture *cap) {
|
|||||||
|
|
||||||
static void gsr_capture_portal_destroy(gsr_capture *cap, AVCodecContext *video_codec_context) {
|
static void gsr_capture_portal_destroy(gsr_capture *cap, AVCodecContext *video_codec_context) {
|
||||||
(void)video_codec_context;
|
(void)video_codec_context;
|
||||||
gsr_capture_portal *cap_portal = cap->priv;
|
gsr_capture_portal *self = cap->priv;
|
||||||
if(cap->priv) {
|
if(cap->priv) {
|
||||||
gsr_capture_portal_stop(cap_portal);
|
gsr_capture_portal_stop(self);
|
||||||
free(cap->priv);
|
free(cap->priv);
|
||||||
cap->priv = NULL;
|
cap->priv = NULL;
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -709,7 +709,7 @@ bool gsr_egl_process_event(gsr_egl *self) {
|
|||||||
|
|
||||||
void gsr_egl_swap_buffers(gsr_egl *self) {
|
void gsr_egl_swap_buffers(gsr_egl *self) {
|
||||||
/* This uses less cpu than swap buffer on nvidia */
|
/* This uses less cpu than swap buffer on nvidia */
|
||||||
// TODO:
|
// TODO: Do these and remove swap
|
||||||
//self->glFlush();
|
//self->glFlush();
|
||||||
//self->glFinish();
|
//self->glFinish();
|
||||||
if(self->egl_display) {
|
if(self->egl_display) {
|
||||||
|
|||||||
@@ -356,22 +356,22 @@ static gsr_supported_video_codecs gsr_video_encoder_cuda_get_supported_codecs(gs
|
|||||||
static void gsr_video_encoder_cuda_stop(gsr_video_encoder_cuda *self, AVCodecContext *video_codec_context);
|
static void gsr_video_encoder_cuda_stop(gsr_video_encoder_cuda *self, AVCodecContext *video_codec_context);
|
||||||
|
|
||||||
static bool gsr_video_encoder_cuda_start(gsr_video_encoder *encoder, AVCodecContext *video_codec_context, AVFrame *frame) {
|
static bool gsr_video_encoder_cuda_start(gsr_video_encoder *encoder, AVCodecContext *video_codec_context, AVFrame *frame) {
|
||||||
gsr_video_encoder_cuda *encoder_cuda = encoder->priv;
|
gsr_video_encoder_cuda *self = encoder->priv;
|
||||||
|
|
||||||
const bool overclock = gsr_egl_get_display_server(encoder_cuda->params.egl) == GSR_DISPLAY_SERVER_X11 ? encoder_cuda->params.overclock : false;
|
const bool overclock = gsr_egl_get_display_server(self->params.egl) == GSR_DISPLAY_SERVER_X11 ? self->params.overclock : false;
|
||||||
if(!gsr_cuda_load(&encoder_cuda->cuda, encoder_cuda->params.egl->x11.dpy, overclock)) {
|
if(!gsr_cuda_load(&self->cuda, self->params.egl->x11.dpy, overclock)) {
|
||||||
fprintf(stderr, "gsr error: gsr_video_encoder_cuda_start: failed to load cuda\n");
|
fprintf(stderr, "gsr error: gsr_video_encoder_cuda_start: failed to load cuda\n");
|
||||||
gsr_video_encoder_cuda_stop(encoder_cuda, video_codec_context);
|
gsr_video_encoder_cuda_stop(self, video_codec_context);
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
if(!gsr_video_encoder_cuda_setup_context(encoder_cuda, video_codec_context)) {
|
if(!gsr_video_encoder_cuda_setup_context(self, video_codec_context)) {
|
||||||
gsr_video_encoder_cuda_stop(encoder_cuda, video_codec_context);
|
gsr_video_encoder_cuda_stop(self, video_codec_context);
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
if(!gsr_video_encoder_cuda_setup_textures(encoder_cuda, video_codec_context, frame)) {
|
if(!gsr_video_encoder_cuda_setup_textures(self, video_codec_context, frame)) {
|
||||||
gsr_video_encoder_cuda_stop(encoder_cuda, video_codec_context);
|
gsr_video_encoder_cuda_stop(self, video_codec_context);
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -402,7 +402,7 @@ void gsr_video_encoder_cuda_stop(gsr_video_encoder_cuda *self, AVCodecContext *v
|
|||||||
}
|
}
|
||||||
|
|
||||||
static void gsr_video_encoder_cuda_copy_textures_to_frame(gsr_video_encoder *encoder, AVFrame *frame) {
|
static void gsr_video_encoder_cuda_copy_textures_to_frame(gsr_video_encoder *encoder, AVFrame *frame) {
|
||||||
gsr_video_encoder_cuda *encoder_cuda = encoder->priv;
|
gsr_video_encoder_cuda *self = encoder->priv;
|
||||||
const int div[2] = {1, 2}; // divide UV texture size by 2 because chroma is half size
|
const int div[2] = {1, 2}; // divide UV texture size by 2 because chroma is half size
|
||||||
for(int i = 0; i < 2; ++i) {
|
for(int i = 0; i < 2; ++i) {
|
||||||
CUDA_MEMCPY2D memcpy_struct;
|
CUDA_MEMCPY2D memcpy_struct;
|
||||||
@@ -414,26 +414,26 @@ static void gsr_video_encoder_cuda_copy_textures_to_frame(gsr_video_encoder *enc
|
|||||||
memcpy_struct.dstY = 0;
|
memcpy_struct.dstY = 0;
|
||||||
memcpy_struct.dstMemoryType = CU_MEMORYTYPE_DEVICE;
|
memcpy_struct.dstMemoryType = CU_MEMORYTYPE_DEVICE;
|
||||||
|
|
||||||
memcpy_struct.srcArray = encoder_cuda->mapped_arrays[i];
|
memcpy_struct.srcArray = self->mapped_arrays[i];
|
||||||
memcpy_struct.srcPitch = frame->width / div[i];
|
memcpy_struct.srcPitch = frame->width / div[i];
|
||||||
memcpy_struct.dstDevice = (CUdeviceptr)frame->data[i];
|
memcpy_struct.dstDevice = (CUdeviceptr)frame->data[i];
|
||||||
memcpy_struct.dstPitch = frame->linesize[i];
|
memcpy_struct.dstPitch = frame->linesize[i];
|
||||||
memcpy_struct.WidthInBytes = frame->width * (encoder_cuda->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? 2 : 1);
|
memcpy_struct.WidthInBytes = frame->width * (self->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? 2 : 1);
|
||||||
memcpy_struct.Height = frame->height / div[i];
|
memcpy_struct.Height = frame->height / div[i];
|
||||||
// TODO: Remove this copy if possible
|
// TODO: Remove this copy if possible
|
||||||
encoder_cuda->cuda.cuMemcpy2DAsync_v2(&memcpy_struct, encoder_cuda->cuda_stream);
|
self->cuda.cuMemcpy2DAsync_v2(&memcpy_struct, self->cuda_stream);
|
||||||
}
|
}
|
||||||
|
|
||||||
// TODO: needed?
|
// TODO: needed?
|
||||||
encoder_cuda->cuda.cuStreamSynchronize(encoder_cuda->cuda_stream);
|
self->cuda.cuStreamSynchronize(self->cuda_stream);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void gsr_video_encoder_cuda_get_textures(gsr_video_encoder *encoder, unsigned int *textures, int *num_textures, gsr_destination_color *destination_color) {
|
static void gsr_video_encoder_cuda_get_textures(gsr_video_encoder *encoder, unsigned int *textures, int *num_textures, gsr_destination_color *destination_color) {
|
||||||
gsr_video_encoder_cuda *encoder_cuda = encoder->priv;
|
gsr_video_encoder_cuda *self = encoder->priv;
|
||||||
textures[0] = encoder_cuda->target_textures[0];
|
textures[0] = self->target_textures[0];
|
||||||
textures[1] = encoder_cuda->target_textures[1];
|
textures[1] = self->target_textures[1];
|
||||||
*num_textures = 2;
|
*num_textures = 2;
|
||||||
*destination_color = encoder_cuda->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? GSR_DESTINATION_COLOR_P010 : GSR_DESTINATION_COLOR_NV12;
|
*destination_color = self->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? GSR_DESTINATION_COLOR_P010 : GSR_DESTINATION_COLOR_NV12;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void gsr_video_encoder_cuda_destroy(gsr_video_encoder *encoder, AVCodecContext *video_codec_context) {
|
static void gsr_video_encoder_cuda_destroy(gsr_video_encoder *encoder, AVCodecContext *video_codec_context) {
|
||||||
|
|||||||
@@ -77,7 +77,7 @@ static gsr_supported_video_codecs gsr_video_encoder_software_get_supported_codec
|
|||||||
static void gsr_video_encoder_software_stop(gsr_video_encoder_software *self, AVCodecContext *video_codec_context);
|
static void gsr_video_encoder_software_stop(gsr_video_encoder_software *self, AVCodecContext *video_codec_context);
|
||||||
|
|
||||||
static bool gsr_video_encoder_software_start(gsr_video_encoder *encoder, AVCodecContext *video_codec_context, AVFrame *frame) {
|
static bool gsr_video_encoder_software_start(gsr_video_encoder *encoder, AVCodecContext *video_codec_context, AVFrame *frame) {
|
||||||
gsr_video_encoder_software *encoder_software = encoder->priv;
|
gsr_video_encoder_software *self = encoder->priv;
|
||||||
|
|
||||||
video_codec_context->width = FFALIGN(video_codec_context->width, LINESIZE_ALIGNMENT);
|
video_codec_context->width = FFALIGN(video_codec_context->width, LINESIZE_ALIGNMENT);
|
||||||
video_codec_context->height = FFALIGN(video_codec_context->height, 2);
|
video_codec_context->height = FFALIGN(video_codec_context->height, 2);
|
||||||
@@ -85,8 +85,8 @@ static bool gsr_video_encoder_software_start(gsr_video_encoder *encoder, AVCodec
|
|||||||
frame->width = video_codec_context->width;
|
frame->width = video_codec_context->width;
|
||||||
frame->height = video_codec_context->height;
|
frame->height = video_codec_context->height;
|
||||||
|
|
||||||
if(!gsr_video_encoder_software_setup_textures(encoder_software, video_codec_context, frame)) {
|
if(!gsr_video_encoder_software_setup_textures(self, video_codec_context, frame)) {
|
||||||
gsr_video_encoder_software_stop(encoder_software, video_codec_context);
|
gsr_video_encoder_software_stop(self, video_codec_context);
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -101,28 +101,28 @@ void gsr_video_encoder_software_stop(gsr_video_encoder_software *self, AVCodecCo
|
|||||||
}
|
}
|
||||||
|
|
||||||
static void gsr_video_encoder_software_copy_textures_to_frame(gsr_video_encoder *encoder, AVFrame *frame) {
|
static void gsr_video_encoder_software_copy_textures_to_frame(gsr_video_encoder *encoder, AVFrame *frame) {
|
||||||
gsr_video_encoder_software *encoder_software = encoder->priv;
|
gsr_video_encoder_software *self = encoder->priv;
|
||||||
// TODO: hdr support
|
// TODO: hdr support
|
||||||
const unsigned int formats[2] = { GL_RED, GL_RG };
|
const unsigned int formats[2] = { GL_RED, GL_RG };
|
||||||
for(int i = 0; i < 2; ++i) {
|
for(int i = 0; i < 2; ++i) {
|
||||||
encoder_software->params.egl->glBindTexture(GL_TEXTURE_2D, encoder_software->target_textures[i]);
|
self->params.egl->glBindTexture(GL_TEXTURE_2D, self->target_textures[i]);
|
||||||
// We could use glGetTexSubImage and then we wouldn't have to use a specific linesize (LINESIZE_ALIGNMENT) that adds padding,
|
// We could use glGetTexSubImage and then we wouldn't have to use a specific linesize (LINESIZE_ALIGNMENT) that adds padding,
|
||||||
// but glGetTexSubImage is only available starting from opengl 4.5.
|
// but glGetTexSubImage is only available starting from opengl 4.5.
|
||||||
encoder_software->params.egl->glGetTexImage(GL_TEXTURE_2D, 0, formats[i], GL_UNSIGNED_BYTE, frame->data[i]);
|
self->params.egl->glGetTexImage(GL_TEXTURE_2D, 0, formats[i], GL_UNSIGNED_BYTE, frame->data[i]);
|
||||||
}
|
}
|
||||||
encoder_software->params.egl->glBindTexture(GL_TEXTURE_2D, 0);
|
self->params.egl->glBindTexture(GL_TEXTURE_2D, 0);
|
||||||
// cap_kms->kms.base.egl->eglSwapBuffers(cap_kms->kms.base.egl->egl_display, cap_kms->kms.base.egl->egl_surface);
|
// cap_kms->kms.base.egl->eglSwapBuffers(cap_kms->kms.base.egl->egl_display, cap_kms->kms.base.egl->egl_surface);
|
||||||
|
|
||||||
encoder_software->params.egl->glFlush();
|
self->params.egl->glFlush();
|
||||||
encoder_software->params.egl->glFinish();
|
self->params.egl->glFinish();
|
||||||
}
|
}
|
||||||
|
|
||||||
static void gsr_video_encoder_software_get_textures(gsr_video_encoder *encoder, unsigned int *textures, int *num_textures, gsr_destination_color *destination_color) {
|
static void gsr_video_encoder_software_get_textures(gsr_video_encoder *encoder, unsigned int *textures, int *num_textures, gsr_destination_color *destination_color) {
|
||||||
gsr_video_encoder_software *encoder_software = encoder->priv;
|
gsr_video_encoder_software *self = encoder->priv;
|
||||||
textures[0] = encoder_software->target_textures[0];
|
textures[0] = self->target_textures[0];
|
||||||
textures[1] = encoder_software->target_textures[1];
|
textures[1] = self->target_textures[1];
|
||||||
*num_textures = 2;
|
*num_textures = 2;
|
||||||
*destination_color = encoder_software->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? GSR_DESTINATION_COLOR_P010 : GSR_DESTINATION_COLOR_NV12;
|
*destination_color = self->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? GSR_DESTINATION_COLOR_P010 : GSR_DESTINATION_COLOR_NV12;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void gsr_video_encoder_software_destroy(gsr_video_encoder *encoder, AVCodecContext *video_codec_context) {
|
static void gsr_video_encoder_software_destroy(gsr_video_encoder *encoder, AVCodecContext *video_codec_context) {
|
||||||
|
|||||||
@@ -299,12 +299,12 @@ static bool get_supported_video_codecs(VADisplay va_dpy, gsr_supported_video_cod
|
|||||||
}
|
}
|
||||||
|
|
||||||
static gsr_supported_video_codecs gsr_video_encoder_vaapi_get_supported_codecs(gsr_video_encoder *encoder, bool cleanup) {
|
static gsr_supported_video_codecs gsr_video_encoder_vaapi_get_supported_codecs(gsr_video_encoder *encoder, bool cleanup) {
|
||||||
gsr_video_encoder_vaapi *encoder_vaapi = encoder->priv;
|
gsr_video_encoder_vaapi *self = encoder->priv;
|
||||||
gsr_supported_video_codecs supported_video_codecs = {0};
|
gsr_supported_video_codecs supported_video_codecs = {0};
|
||||||
|
|
||||||
char render_path[128];
|
char render_path[128];
|
||||||
if(!gsr_card_path_get_render_path(encoder_vaapi->params.egl->card_path, render_path)) {
|
if(!gsr_card_path_get_render_path(self->params.egl->card_path, render_path)) {
|
||||||
fprintf(stderr, "gsr error: gsr_video_encoder_vaapi_get_supported_codecs: failed to get /dev/dri/renderDXXX file from %s\n", encoder_vaapi->params.egl->card_path);
|
fprintf(stderr, "gsr error: gsr_video_encoder_vaapi_get_supported_codecs: failed to get /dev/dri/renderDXXX file from %s\n", self->params.egl->card_path);
|
||||||
return supported_video_codecs;
|
return supported_video_codecs;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -329,13 +329,13 @@ static gsr_supported_video_codecs gsr_video_encoder_vaapi_get_supported_codecs(g
|
|||||||
static void gsr_video_encoder_vaapi_stop(gsr_video_encoder_vaapi *self, AVCodecContext *video_codec_context);
|
static void gsr_video_encoder_vaapi_stop(gsr_video_encoder_vaapi *self, AVCodecContext *video_codec_context);
|
||||||
|
|
||||||
static bool gsr_video_encoder_vaapi_start(gsr_video_encoder *encoder, AVCodecContext *video_codec_context, AVFrame *frame) {
|
static bool gsr_video_encoder_vaapi_start(gsr_video_encoder *encoder, AVCodecContext *video_codec_context, AVFrame *frame) {
|
||||||
gsr_video_encoder_vaapi *encoder_vaapi = encoder->priv;
|
gsr_video_encoder_vaapi *self = encoder->priv;
|
||||||
|
|
||||||
if(encoder_vaapi->params.egl->gpu_info.vendor == GSR_GPU_VENDOR_AMD && video_codec_context->codec_id == AV_CODEC_ID_HEVC) {
|
if(self->params.egl->gpu_info.vendor == GSR_GPU_VENDOR_AMD && video_codec_context->codec_id == AV_CODEC_ID_HEVC) {
|
||||||
// TODO: dont do this if using ffmpeg reports that this is not needed (AMD driver bug that was fixed recently)
|
// TODO: dont do this if using ffmpeg reports that this is not needed (AMD driver bug that was fixed recently)
|
||||||
video_codec_context->width = FFALIGN(video_codec_context->width, 64);
|
video_codec_context->width = FFALIGN(video_codec_context->width, 64);
|
||||||
video_codec_context->height = FFALIGN(video_codec_context->height, 16);
|
video_codec_context->height = FFALIGN(video_codec_context->height, 16);
|
||||||
} else if(encoder_vaapi->params.egl->gpu_info.vendor == GSR_GPU_VENDOR_AMD && video_codec_context->codec_id == AV_CODEC_ID_AV1) {
|
} else if(self->params.egl->gpu_info.vendor == GSR_GPU_VENDOR_AMD && video_codec_context->codec_id == AV_CODEC_ID_AV1) {
|
||||||
// TODO: Dont do this for VCN 5 and forward which should fix this hardware bug
|
// TODO: Dont do this for VCN 5 and forward which should fix this hardware bug
|
||||||
video_codec_context->width = FFALIGN(video_codec_context->width, 64);
|
video_codec_context->width = FFALIGN(video_codec_context->width, 64);
|
||||||
// AMD driver has special case handling for 1080 height to set it to 1082 instead of 1088 (1080 aligned to 16).
|
// AMD driver has special case handling for 1080 height to set it to 1082 instead of 1088 (1080 aligned to 16).
|
||||||
@@ -354,13 +354,13 @@ static bool gsr_video_encoder_vaapi_start(gsr_video_encoder *encoder, AVCodecCon
|
|||||||
frame->width = video_codec_context->width;
|
frame->width = video_codec_context->width;
|
||||||
frame->height = video_codec_context->height;
|
frame->height = video_codec_context->height;
|
||||||
|
|
||||||
if(!gsr_video_encoder_vaapi_setup_context(encoder_vaapi, video_codec_context)) {
|
if(!gsr_video_encoder_vaapi_setup_context(self, video_codec_context)) {
|
||||||
gsr_video_encoder_vaapi_stop(encoder_vaapi, video_codec_context);
|
gsr_video_encoder_vaapi_stop(self, video_codec_context);
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
if(!gsr_video_encoder_vaapi_setup_textures(encoder_vaapi, video_codec_context, frame)) {
|
if(!gsr_video_encoder_vaapi_setup_textures(self, video_codec_context, frame)) {
|
||||||
gsr_video_encoder_vaapi_stop(encoder_vaapi, video_codec_context);
|
gsr_video_encoder_vaapi_stop(self, video_codec_context);
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -386,11 +386,11 @@ void gsr_video_encoder_vaapi_stop(gsr_video_encoder_vaapi *self, AVCodecContext
|
|||||||
}
|
}
|
||||||
|
|
||||||
static void gsr_video_encoder_vaapi_get_textures(gsr_video_encoder *encoder, unsigned int *textures, int *num_textures, gsr_destination_color *destination_color) {
|
static void gsr_video_encoder_vaapi_get_textures(gsr_video_encoder *encoder, unsigned int *textures, int *num_textures, gsr_destination_color *destination_color) {
|
||||||
gsr_video_encoder_vaapi *encoder_vaapi = encoder->priv;
|
gsr_video_encoder_vaapi *self = encoder->priv;
|
||||||
textures[0] = encoder_vaapi->target_textures[0];
|
textures[0] = self->target_textures[0];
|
||||||
textures[1] = encoder_vaapi->target_textures[1];
|
textures[1] = self->target_textures[1];
|
||||||
*num_textures = 2;
|
*num_textures = 2;
|
||||||
*destination_color = encoder_vaapi->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? GSR_DESTINATION_COLOR_P010 : GSR_DESTINATION_COLOR_NV12;
|
*destination_color = self->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? GSR_DESTINATION_COLOR_P010 : GSR_DESTINATION_COLOR_NV12;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void gsr_video_encoder_vaapi_destroy(gsr_video_encoder *encoder, AVCodecContext *video_codec_context) {
|
static void gsr_video_encoder_vaapi_destroy(gsr_video_encoder *encoder, AVCodecContext *video_codec_context) {
|
||||||
|
|||||||
@@ -86,11 +86,11 @@ void gsr_video_encoder_vulkan_stop(gsr_video_encoder_vulkan *self, AVCodecContex
|
|||||||
}
|
}
|
||||||
|
|
||||||
static void gsr_video_encoder_vulkan_get_textures(gsr_video_encoder *encoder, unsigned int *textures, int *num_textures, gsr_destination_color *destination_color) {
|
static void gsr_video_encoder_vulkan_get_textures(gsr_video_encoder *encoder, unsigned int *textures, int *num_textures, gsr_destination_color *destination_color) {
|
||||||
gsr_video_encoder_vulkan *encoder_vaapi = encoder->priv;
|
gsr_video_encoder_vulkan *self = encoder->priv;
|
||||||
textures[0] = encoder_vaapi->target_textures[0];
|
textures[0] = self->target_textures[0];
|
||||||
textures[1] = encoder_vaapi->target_textures[1];
|
textures[1] = self->target_textures[1];
|
||||||
*num_textures = 2;
|
*num_textures = 2;
|
||||||
*destination_color = encoder_vaapi->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? GSR_DESTINATION_COLOR_P010 : GSR_DESTINATION_COLOR_NV12;
|
*destination_color = self->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? GSR_DESTINATION_COLOR_P010 : GSR_DESTINATION_COLOR_NV12;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void gsr_video_encoder_vulkan_destroy(gsr_video_encoder *encoder, AVCodecContext *video_codec_context) {
|
static void gsr_video_encoder_vulkan_destroy(gsr_video_encoder *encoder, AVCodecContext *video_codec_context) {
|
||||||
@@ -104,20 +104,20 @@ gsr_video_encoder* gsr_video_encoder_vulkan_create(const gsr_video_encoder_vulka
|
|||||||
if(!encoder)
|
if(!encoder)
|
||||||
return NULL;
|
return NULL;
|
||||||
|
|
||||||
gsr_video_encoder_vulkan *encoder_vaapi = calloc(1, sizeof(gsr_video_encoder_vulkan));
|
gsr_video_encoder_vulkan *encoder_vulkan = calloc(1, sizeof(gsr_video_encoder_vulkan));
|
||||||
if(!encoder_vaapi) {
|
if(!encoder_vulkan) {
|
||||||
free(encoder);
|
free(encoder);
|
||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
encoder_vaapi->params = *params;
|
encoder_vulkan->params = *params;
|
||||||
|
|
||||||
*encoder = (gsr_video_encoder) {
|
*encoder = (gsr_video_encoder) {
|
||||||
.start = gsr_video_encoder_vulkan_start,
|
.start = gsr_video_encoder_vulkan_start,
|
||||||
.copy_textures_to_frame = NULL,
|
.copy_textures_to_frame = NULL,
|
||||||
.get_textures = gsr_video_encoder_vulkan_get_textures,
|
.get_textures = gsr_video_encoder_vulkan_get_textures,
|
||||||
.destroy = gsr_video_encoder_vulkan_destroy,
|
.destroy = gsr_video_encoder_vulkan_destroy,
|
||||||
.priv = encoder_vaapi
|
.priv = encoder_vulkan
|
||||||
};
|
};
|
||||||
|
|
||||||
return encoder;
|
return encoder;
|
||||||
|
|||||||
Reference in New Issue
Block a user