mirror of
https://repo.dec05eba.com/gpu-screen-recorder
synced 2026-04-14 14:32:25 +09:00
Name capture/encoder variable self
This commit is contained in:
@@ -16,6 +16,7 @@ src = [
|
||||
'src/encoder/video/video.c',
|
||||
'src/encoder/video/cuda.c',
|
||||
'src/encoder/video/vaapi.c',
|
||||
'src/encoder/video/vulkan.c',
|
||||
'src/encoder/video/software.c',
|
||||
'src/egl.c',
|
||||
'src/cuda.c',
|
||||
|
||||
@@ -102,7 +102,7 @@ static void set_func_ptr(void **dst, void *src) {
|
||||
}
|
||||
|
||||
static bool gsr_capture_nvfbc_load_library(gsr_capture *cap) {
|
||||
gsr_capture_nvfbc *cap_nvfbc = cap->priv;
|
||||
gsr_capture_nvfbc *self = cap->priv;
|
||||
|
||||
dlerror(); /* clear */
|
||||
void *lib = dlopen("libnvidia-fbc.so.1", RTLD_LAZY);
|
||||
@@ -111,23 +111,23 @@ static bool gsr_capture_nvfbc_load_library(gsr_capture *cap) {
|
||||
return false;
|
||||
}
|
||||
|
||||
set_func_ptr((void**)&cap_nvfbc->nv_fbc_create_instance, dlsym(lib, "NvFBCCreateInstance"));
|
||||
if(!cap_nvfbc->nv_fbc_create_instance) {
|
||||
set_func_ptr((void**)&self->nv_fbc_create_instance, dlsym(lib, "NvFBCCreateInstance"));
|
||||
if(!self->nv_fbc_create_instance) {
|
||||
fprintf(stderr, "gsr error: unable to resolve symbol 'NvFBCCreateInstance'\n");
|
||||
dlclose(lib);
|
||||
return false;
|
||||
}
|
||||
|
||||
memset(&cap_nvfbc->nv_fbc_function_list, 0, sizeof(cap_nvfbc->nv_fbc_function_list));
|
||||
cap_nvfbc->nv_fbc_function_list.dwVersion = NVFBC_VERSION;
|
||||
NVFBCSTATUS status = cap_nvfbc->nv_fbc_create_instance(&cap_nvfbc->nv_fbc_function_list);
|
||||
memset(&self->nv_fbc_function_list, 0, sizeof(self->nv_fbc_function_list));
|
||||
self->nv_fbc_function_list.dwVersion = NVFBC_VERSION;
|
||||
NVFBCSTATUS status = self->nv_fbc_create_instance(&self->nv_fbc_function_list);
|
||||
if(status != NVFBC_SUCCESS) {
|
||||
fprintf(stderr, "gsr error: failed to create NvFBC instance (status: %d)\n", status);
|
||||
dlclose(lib);
|
||||
return false;
|
||||
}
|
||||
|
||||
cap_nvfbc->library = lib;
|
||||
self->library = lib;
|
||||
return true;
|
||||
}
|
||||
|
||||
@@ -153,64 +153,64 @@ static void set_vertical_sync_enabled(gsr_egl *egl, int enabled) {
|
||||
fprintf(stderr, "gsr warning: setting vertical sync failed\n");
|
||||
}
|
||||
|
||||
static void gsr_capture_nvfbc_destroy_session(gsr_capture_nvfbc *cap_nvfbc) {
|
||||
if(cap_nvfbc->fbc_handle_created && cap_nvfbc->capture_session_created) {
|
||||
static void gsr_capture_nvfbc_destroy_session(gsr_capture_nvfbc *self) {
|
||||
if(self->fbc_handle_created && self->capture_session_created) {
|
||||
NVFBC_DESTROY_CAPTURE_SESSION_PARAMS destroy_capture_params;
|
||||
memset(&destroy_capture_params, 0, sizeof(destroy_capture_params));
|
||||
destroy_capture_params.dwVersion = NVFBC_DESTROY_CAPTURE_SESSION_PARAMS_VER;
|
||||
cap_nvfbc->nv_fbc_function_list.nvFBCDestroyCaptureSession(cap_nvfbc->nv_fbc_handle, &destroy_capture_params);
|
||||
cap_nvfbc->capture_session_created = false;
|
||||
self->nv_fbc_function_list.nvFBCDestroyCaptureSession(self->nv_fbc_handle, &destroy_capture_params);
|
||||
self->capture_session_created = false;
|
||||
}
|
||||
}
|
||||
|
||||
static void gsr_capture_nvfbc_destroy_handle(gsr_capture_nvfbc *cap_nvfbc) {
|
||||
if(cap_nvfbc->fbc_handle_created) {
|
||||
static void gsr_capture_nvfbc_destroy_handle(gsr_capture_nvfbc *self) {
|
||||
if(self->fbc_handle_created) {
|
||||
NVFBC_DESTROY_HANDLE_PARAMS destroy_params;
|
||||
memset(&destroy_params, 0, sizeof(destroy_params));
|
||||
destroy_params.dwVersion = NVFBC_DESTROY_HANDLE_PARAMS_VER;
|
||||
cap_nvfbc->nv_fbc_function_list.nvFBCDestroyHandle(cap_nvfbc->nv_fbc_handle, &destroy_params);
|
||||
cap_nvfbc->fbc_handle_created = false;
|
||||
cap_nvfbc->nv_fbc_handle = 0;
|
||||
self->nv_fbc_function_list.nvFBCDestroyHandle(self->nv_fbc_handle, &destroy_params);
|
||||
self->fbc_handle_created = false;
|
||||
self->nv_fbc_handle = 0;
|
||||
}
|
||||
}
|
||||
|
||||
static void gsr_capture_nvfbc_destroy_session_and_handle(gsr_capture_nvfbc *cap_nvfbc) {
|
||||
gsr_capture_nvfbc_destroy_session(cap_nvfbc);
|
||||
gsr_capture_nvfbc_destroy_handle(cap_nvfbc);
|
||||
static void gsr_capture_nvfbc_destroy_session_and_handle(gsr_capture_nvfbc *self) {
|
||||
gsr_capture_nvfbc_destroy_session(self);
|
||||
gsr_capture_nvfbc_destroy_handle(self);
|
||||
}
|
||||
|
||||
static int gsr_capture_nvfbc_setup_handle(gsr_capture_nvfbc *cap_nvfbc) {
|
||||
static int gsr_capture_nvfbc_setup_handle(gsr_capture_nvfbc *self) {
|
||||
NVFBCSTATUS status;
|
||||
|
||||
NVFBC_CREATE_HANDLE_PARAMS create_params;
|
||||
memset(&create_params, 0, sizeof(create_params));
|
||||
create_params.dwVersion = NVFBC_CREATE_HANDLE_PARAMS_VER;
|
||||
create_params.bExternallyManagedContext = NVFBC_TRUE;
|
||||
create_params.glxCtx = cap_nvfbc->params.egl->glx_context;
|
||||
create_params.glxFBConfig = cap_nvfbc->params.egl->glx_fb_config;
|
||||
create_params.glxCtx = self->params.egl->glx_context;
|
||||
create_params.glxFBConfig = self->params.egl->glx_fb_config;
|
||||
|
||||
status = cap_nvfbc->nv_fbc_function_list.nvFBCCreateHandle(&cap_nvfbc->nv_fbc_handle, &create_params);
|
||||
status = self->nv_fbc_function_list.nvFBCCreateHandle(&self->nv_fbc_handle, &create_params);
|
||||
if(status != NVFBC_SUCCESS) {
|
||||
// Reverse engineering for interoperability
|
||||
const uint8_t enable_key[] = { 0xac, 0x10, 0xc9, 0x2e, 0xa5, 0xe6, 0x87, 0x4f, 0x8f, 0x4b, 0xf4, 0x61, 0xf8, 0x56, 0x27, 0xe9 };
|
||||
create_params.privateData = enable_key;
|
||||
create_params.privateDataSize = 16;
|
||||
|
||||
status = cap_nvfbc->nv_fbc_function_list.nvFBCCreateHandle(&cap_nvfbc->nv_fbc_handle, &create_params);
|
||||
status = self->nv_fbc_function_list.nvFBCCreateHandle(&self->nv_fbc_handle, &create_params);
|
||||
if(status != NVFBC_SUCCESS) {
|
||||
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: %s\n", cap_nvfbc->nv_fbc_function_list.nvFBCGetLastErrorStr(cap_nvfbc->nv_fbc_handle));
|
||||
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: %s\n", self->nv_fbc_function_list.nvFBCGetLastErrorStr(self->nv_fbc_handle));
|
||||
goto error_cleanup;
|
||||
}
|
||||
}
|
||||
cap_nvfbc->fbc_handle_created = true;
|
||||
self->fbc_handle_created = true;
|
||||
|
||||
NVFBC_GET_STATUS_PARAMS status_params;
|
||||
memset(&status_params, 0, sizeof(status_params));
|
||||
status_params.dwVersion = NVFBC_GET_STATUS_PARAMS_VER;
|
||||
|
||||
status = cap_nvfbc->nv_fbc_function_list.nvFBCGetStatus(cap_nvfbc->nv_fbc_handle, &status_params);
|
||||
status = self->nv_fbc_function_list.nvFBCGetStatus(self->nv_fbc_handle, &status_params);
|
||||
if(status != NVFBC_SUCCESS) {
|
||||
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: %s\n", cap_nvfbc->nv_fbc_function_list.nvFBCGetLastErrorStr(cap_nvfbc->nv_fbc_handle));
|
||||
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: %s\n", self->nv_fbc_function_list.nvFBCGetLastErrorStr(self->nv_fbc_handle));
|
||||
goto error_cleanup;
|
||||
}
|
||||
|
||||
@@ -219,10 +219,10 @@ static int gsr_capture_nvfbc_setup_handle(gsr_capture_nvfbc *cap_nvfbc) {
|
||||
goto error_cleanup;
|
||||
}
|
||||
|
||||
cap_nvfbc->tracking_width = XWidthOfScreen(DefaultScreenOfDisplay(cap_nvfbc->params.egl->x11.dpy));
|
||||
cap_nvfbc->tracking_height = XHeightOfScreen(DefaultScreenOfDisplay(cap_nvfbc->params.egl->x11.dpy));
|
||||
cap_nvfbc->tracking_type = strcmp(cap_nvfbc->params.display_to_capture, "screen") == 0 ? NVFBC_TRACKING_SCREEN : NVFBC_TRACKING_OUTPUT;
|
||||
if(cap_nvfbc->tracking_type == NVFBC_TRACKING_OUTPUT) {
|
||||
self->tracking_width = XWidthOfScreen(DefaultScreenOfDisplay(self->params.egl->x11.dpy));
|
||||
self->tracking_height = XHeightOfScreen(DefaultScreenOfDisplay(self->params.egl->x11.dpy));
|
||||
self->tracking_type = strcmp(self->params.display_to_capture, "screen") == 0 ? NVFBC_TRACKING_SCREEN : NVFBC_TRACKING_OUTPUT;
|
||||
if(self->tracking_type == NVFBC_TRACKING_OUTPUT) {
|
||||
if(!status_params.bXRandRAvailable) {
|
||||
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: the xrandr extension is not available\n");
|
||||
goto error_cleanup;
|
||||
@@ -233,9 +233,9 @@ static int gsr_capture_nvfbc_setup_handle(gsr_capture_nvfbc *cap_nvfbc) {
|
||||
goto error_cleanup;
|
||||
}
|
||||
|
||||
cap_nvfbc->output_id = get_output_id_from_display_name(status_params.outputs, status_params.dwOutputNum, cap_nvfbc->params.display_to_capture, &cap_nvfbc->tracking_width, &cap_nvfbc->tracking_height);
|
||||
if(cap_nvfbc->output_id == 0) {
|
||||
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: display '%s' not found\n", cap_nvfbc->params.display_to_capture);
|
||||
self->output_id = get_output_id_from_display_name(status_params.outputs, status_params.dwOutputNum, self->params.display_to_capture, &self->tracking_width, &self->tracking_height);
|
||||
if(self->output_id == 0) {
|
||||
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: display '%s' not found\n", self->params.display_to_capture);
|
||||
goto error_cleanup;
|
||||
}
|
||||
}
|
||||
@@ -243,83 +243,83 @@ static int gsr_capture_nvfbc_setup_handle(gsr_capture_nvfbc *cap_nvfbc) {
|
||||
return 0;
|
||||
|
||||
error_cleanup:
|
||||
gsr_capture_nvfbc_destroy_session_and_handle(cap_nvfbc);
|
||||
gsr_capture_nvfbc_destroy_session_and_handle(self);
|
||||
return -1;
|
||||
}
|
||||
|
||||
static int gsr_capture_nvfbc_setup_session(gsr_capture_nvfbc *cap_nvfbc) {
|
||||
static int gsr_capture_nvfbc_setup_session(gsr_capture_nvfbc *self) {
|
||||
NVFBC_CREATE_CAPTURE_SESSION_PARAMS create_capture_params;
|
||||
memset(&create_capture_params, 0, sizeof(create_capture_params));
|
||||
create_capture_params.dwVersion = NVFBC_CREATE_CAPTURE_SESSION_PARAMS_VER;
|
||||
create_capture_params.eCaptureType = NVFBC_CAPTURE_TO_GL;
|
||||
create_capture_params.bWithCursor = (!cap_nvfbc->params.direct_capture || cap_nvfbc->supports_direct_cursor) ? NVFBC_TRUE : NVFBC_FALSE;
|
||||
if(!cap_nvfbc->params.record_cursor)
|
||||
create_capture_params.bWithCursor = (!self->params.direct_capture || self->supports_direct_cursor) ? NVFBC_TRUE : NVFBC_FALSE;
|
||||
if(!self->params.record_cursor)
|
||||
create_capture_params.bWithCursor = false;
|
||||
if(cap_nvfbc->capture_region)
|
||||
create_capture_params.captureBox = (NVFBC_BOX){ cap_nvfbc->x, cap_nvfbc->y, cap_nvfbc->width, cap_nvfbc->height };
|
||||
create_capture_params.eTrackingType = cap_nvfbc->tracking_type;
|
||||
create_capture_params.dwSamplingRateMs = (uint32_t)ceilf(1000.0f / (float)cap_nvfbc->params.fps);
|
||||
create_capture_params.bAllowDirectCapture = cap_nvfbc->params.direct_capture ? NVFBC_TRUE : NVFBC_FALSE;
|
||||
create_capture_params.bPushModel = cap_nvfbc->params.direct_capture ? NVFBC_TRUE : NVFBC_FALSE;
|
||||
if(self->capture_region)
|
||||
create_capture_params.captureBox = (NVFBC_BOX){ self->x, self->y, self->width, self->height };
|
||||
create_capture_params.eTrackingType = self->tracking_type;
|
||||
create_capture_params.dwSamplingRateMs = (uint32_t)ceilf(1000.0f / (float)self->params.fps);
|
||||
create_capture_params.bAllowDirectCapture = self->params.direct_capture ? NVFBC_TRUE : NVFBC_FALSE;
|
||||
create_capture_params.bPushModel = self->params.direct_capture ? NVFBC_TRUE : NVFBC_FALSE;
|
||||
create_capture_params.bDisableAutoModesetRecovery = true;
|
||||
if(cap_nvfbc->tracking_type == NVFBC_TRACKING_OUTPUT)
|
||||
create_capture_params.dwOutputId = cap_nvfbc->output_id;
|
||||
if(self->tracking_type == NVFBC_TRACKING_OUTPUT)
|
||||
create_capture_params.dwOutputId = self->output_id;
|
||||
|
||||
NVFBCSTATUS status = cap_nvfbc->nv_fbc_function_list.nvFBCCreateCaptureSession(cap_nvfbc->nv_fbc_handle, &create_capture_params);
|
||||
NVFBCSTATUS status = self->nv_fbc_function_list.nvFBCCreateCaptureSession(self->nv_fbc_handle, &create_capture_params);
|
||||
if(status != NVFBC_SUCCESS) {
|
||||
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: %s\n", cap_nvfbc->nv_fbc_function_list.nvFBCGetLastErrorStr(cap_nvfbc->nv_fbc_handle));
|
||||
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: %s\n", self->nv_fbc_function_list.nvFBCGetLastErrorStr(self->nv_fbc_handle));
|
||||
return -1;
|
||||
}
|
||||
cap_nvfbc->capture_session_created = true;
|
||||
self->capture_session_created = true;
|
||||
|
||||
memset(&cap_nvfbc->setup_params, 0, sizeof(cap_nvfbc->setup_params));
|
||||
cap_nvfbc->setup_params.dwVersion = NVFBC_TOGL_SETUP_PARAMS_VER;
|
||||
cap_nvfbc->setup_params.eBufferFormat = NVFBC_BUFFER_FORMAT_BGRA;
|
||||
memset(&self->setup_params, 0, sizeof(self->setup_params));
|
||||
self->setup_params.dwVersion = NVFBC_TOGL_SETUP_PARAMS_VER;
|
||||
self->setup_params.eBufferFormat = NVFBC_BUFFER_FORMAT_BGRA;
|
||||
|
||||
status = cap_nvfbc->nv_fbc_function_list.nvFBCToGLSetUp(cap_nvfbc->nv_fbc_handle, &cap_nvfbc->setup_params);
|
||||
status = self->nv_fbc_function_list.nvFBCToGLSetUp(self->nv_fbc_handle, &self->setup_params);
|
||||
if(status != NVFBC_SUCCESS) {
|
||||
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: %s\n", cap_nvfbc->nv_fbc_function_list.nvFBCGetLastErrorStr(cap_nvfbc->nv_fbc_handle));
|
||||
gsr_capture_nvfbc_destroy_session(cap_nvfbc);
|
||||
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: %s\n", self->nv_fbc_function_list.nvFBCGetLastErrorStr(self->nv_fbc_handle));
|
||||
gsr_capture_nvfbc_destroy_session(self);
|
||||
return -1;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void gsr_capture_nvfbc_stop(gsr_capture_nvfbc *cap_nvfbc) {
|
||||
gsr_capture_nvfbc_destroy_session_and_handle(cap_nvfbc);
|
||||
if(cap_nvfbc->library) {
|
||||
dlclose(cap_nvfbc->library);
|
||||
cap_nvfbc->library = NULL;
|
||||
static void gsr_capture_nvfbc_stop(gsr_capture_nvfbc *self) {
|
||||
gsr_capture_nvfbc_destroy_session_and_handle(self);
|
||||
if(self->library) {
|
||||
dlclose(self->library);
|
||||
self->library = NULL;
|
||||
}
|
||||
if(cap_nvfbc->params.display_to_capture) {
|
||||
free((void*)cap_nvfbc->params.display_to_capture);
|
||||
cap_nvfbc->params.display_to_capture = NULL;
|
||||
if(self->params.display_to_capture) {
|
||||
free((void*)self->params.display_to_capture);
|
||||
self->params.display_to_capture = NULL;
|
||||
}
|
||||
}
|
||||
|
||||
static int gsr_capture_nvfbc_start(gsr_capture *cap, AVCodecContext *video_codec_context, AVFrame *frame) {
|
||||
gsr_capture_nvfbc *cap_nvfbc = cap->priv;
|
||||
gsr_capture_nvfbc *self = cap->priv;
|
||||
|
||||
if(!gsr_capture_nvfbc_load_library(cap))
|
||||
return -1;
|
||||
|
||||
cap_nvfbc->x = max_int(cap_nvfbc->params.pos.x, 0);
|
||||
cap_nvfbc->y = max_int(cap_nvfbc->params.pos.y, 0);
|
||||
cap_nvfbc->width = max_int(cap_nvfbc->params.size.x, 0);
|
||||
cap_nvfbc->height = max_int(cap_nvfbc->params.size.y, 0);
|
||||
self->x = max_int(self->params.pos.x, 0);
|
||||
self->y = max_int(self->params.pos.y, 0);
|
||||
self->width = max_int(self->params.size.x, 0);
|
||||
self->height = max_int(self->params.size.y, 0);
|
||||
|
||||
cap_nvfbc->capture_region = (cap_nvfbc->x > 0 || cap_nvfbc->y > 0 || cap_nvfbc->width > 0 || cap_nvfbc->height > 0);
|
||||
self->capture_region = (self->x > 0 || self->y > 0 || self->width > 0 || self->height > 0);
|
||||
|
||||
cap_nvfbc->supports_direct_cursor = false;
|
||||
self->supports_direct_cursor = false;
|
||||
int driver_major_version = 0;
|
||||
int driver_minor_version = 0;
|
||||
if(cap_nvfbc->params.direct_capture && get_driver_version(&driver_major_version, &driver_minor_version)) {
|
||||
if(self->params.direct_capture && get_driver_version(&driver_major_version, &driver_minor_version)) {
|
||||
fprintf(stderr, "Info: detected nvidia version: %d.%d\n", driver_major_version, driver_minor_version);
|
||||
|
||||
// TODO:
|
||||
if(version_at_least(driver_major_version, driver_minor_version, 515, 57) && version_less_than(driver_major_version, driver_minor_version, 520, 56)) {
|
||||
cap_nvfbc->params.direct_capture = false;
|
||||
self->params.direct_capture = false;
|
||||
fprintf(stderr, "Warning: \"screen-direct\" has temporary been disabled as it causes stuttering with driver versions >= 515.57 and < 520.56. Please update your driver if possible. Capturing \"screen\" instead.\n");
|
||||
}
|
||||
|
||||
@@ -328,63 +328,63 @@ static int gsr_capture_nvfbc_start(gsr_capture *cap, AVCodecContext *video_codec
|
||||
/*
|
||||
if(direct_capture) {
|
||||
if(version_at_least(driver_major_version, driver_minor_version, 515, 57))
|
||||
cap_nvfbc->supports_direct_cursor = true;
|
||||
self->supports_direct_cursor = true;
|
||||
else
|
||||
fprintf(stderr, "Info: capturing \"screen-direct\" but driver version appears to be less than 515.57. Disabling capture of cursor. Please update your driver if you want to capture your cursor or record \"screen\" instead.\n");
|
||||
}
|
||||
*/
|
||||
}
|
||||
|
||||
if(gsr_capture_nvfbc_setup_handle(cap_nvfbc) != 0) {
|
||||
if(gsr_capture_nvfbc_setup_handle(self) != 0) {
|
||||
goto error_cleanup;
|
||||
}
|
||||
|
||||
if(gsr_capture_nvfbc_setup_session(cap_nvfbc) != 0) {
|
||||
if(gsr_capture_nvfbc_setup_session(self) != 0) {
|
||||
goto error_cleanup;
|
||||
}
|
||||
|
||||
if(cap_nvfbc->capture_region) {
|
||||
video_codec_context->width = FFALIGN(cap_nvfbc->width, 2);
|
||||
video_codec_context->height = FFALIGN(cap_nvfbc->height, 2);
|
||||
if(self->capture_region) {
|
||||
video_codec_context->width = FFALIGN(self->width, 2);
|
||||
video_codec_context->height = FFALIGN(self->height, 2);
|
||||
} else {
|
||||
video_codec_context->width = FFALIGN(cap_nvfbc->tracking_width, 2);
|
||||
video_codec_context->height = FFALIGN(cap_nvfbc->tracking_height, 2);
|
||||
video_codec_context->width = FFALIGN(self->tracking_width, 2);
|
||||
video_codec_context->height = FFALIGN(self->tracking_height, 2);
|
||||
}
|
||||
|
||||
frame->width = video_codec_context->width;
|
||||
frame->height = video_codec_context->height;
|
||||
|
||||
/* Disable vsync */
|
||||
set_vertical_sync_enabled(cap_nvfbc->params.egl, 0);
|
||||
set_vertical_sync_enabled(self->params.egl, 0);
|
||||
|
||||
return 0;
|
||||
|
||||
error_cleanup:
|
||||
gsr_capture_nvfbc_stop(cap_nvfbc);
|
||||
gsr_capture_nvfbc_stop(self);
|
||||
return -1;
|
||||
}
|
||||
|
||||
static int gsr_capture_nvfbc_capture(gsr_capture *cap, AVFrame *frame, gsr_color_conversion *color_conversion) {
|
||||
gsr_capture_nvfbc *cap_nvfbc = cap->priv;
|
||||
gsr_capture_nvfbc *self = cap->priv;
|
||||
|
||||
const double nvfbc_recreate_retry_time_seconds = 1.0;
|
||||
if(cap_nvfbc->nvfbc_needs_recreate) {
|
||||
if(self->nvfbc_needs_recreate) {
|
||||
const double now = clock_get_monotonic_seconds();
|
||||
if(now - cap_nvfbc->nvfbc_dead_start >= nvfbc_recreate_retry_time_seconds) {
|
||||
cap_nvfbc->nvfbc_dead_start = now;
|
||||
gsr_capture_nvfbc_destroy_session_and_handle(cap_nvfbc);
|
||||
if(now - self->nvfbc_dead_start >= nvfbc_recreate_retry_time_seconds) {
|
||||
self->nvfbc_dead_start = now;
|
||||
gsr_capture_nvfbc_destroy_session_and_handle(self);
|
||||
|
||||
if(gsr_capture_nvfbc_setup_handle(cap_nvfbc) != 0) {
|
||||
if(gsr_capture_nvfbc_setup_handle(self) != 0) {
|
||||
fprintf(stderr, "gsr error: gsr_capture_nvfbc_capture failed to recreate nvfbc handle, trying again in %f second(s)\n", nvfbc_recreate_retry_time_seconds);
|
||||
return -1;
|
||||
}
|
||||
|
||||
if(gsr_capture_nvfbc_setup_session(cap_nvfbc) != 0) {
|
||||
|
||||
if(gsr_capture_nvfbc_setup_session(self) != 0) {
|
||||
fprintf(stderr, "gsr error: gsr_capture_nvfbc_capture failed to recreate nvfbc session, trying again in %f second(s)\n", nvfbc_recreate_retry_time_seconds);
|
||||
return -1;
|
||||
}
|
||||
|
||||
cap_nvfbc->nvfbc_needs_recreate = false;
|
||||
self->nvfbc_needs_recreate = false;
|
||||
} else {
|
||||
return 0;
|
||||
}
|
||||
@@ -400,24 +400,24 @@ static int gsr_capture_nvfbc_capture(gsr_capture *cap, AVFrame *frame, gsr_color
|
||||
grab_params.pFrameGrabInfo = &frame_info;
|
||||
grab_params.dwTimeoutMs = 0;
|
||||
|
||||
NVFBCSTATUS status = cap_nvfbc->nv_fbc_function_list.nvFBCToGLGrabFrame(cap_nvfbc->nv_fbc_handle, &grab_params);
|
||||
NVFBCSTATUS status = self->nv_fbc_function_list.nvFBCToGLGrabFrame(self->nv_fbc_handle, &grab_params);
|
||||
if(status != NVFBC_SUCCESS) {
|
||||
fprintf(stderr, "gsr error: gsr_capture_nvfbc_capture failed: %s (%d), recreating session after %f second(s)\n", cap_nvfbc->nv_fbc_function_list.nvFBCGetLastErrorStr(cap_nvfbc->nv_fbc_handle), status, nvfbc_recreate_retry_time_seconds);
|
||||
cap_nvfbc->nvfbc_needs_recreate = true;
|
||||
cap_nvfbc->nvfbc_dead_start = clock_get_monotonic_seconds();
|
||||
fprintf(stderr, "gsr error: gsr_capture_nvfbc_capture failed: %s (%d), recreating session after %f second(s)\n", self->nv_fbc_function_list.nvFBCGetLastErrorStr(self->nv_fbc_handle), status, nvfbc_recreate_retry_time_seconds);
|
||||
self->nvfbc_needs_recreate = true;
|
||||
self->nvfbc_dead_start = clock_get_monotonic_seconds();
|
||||
return 0;
|
||||
}
|
||||
|
||||
cap_nvfbc->params.egl->glFlush();
|
||||
cap_nvfbc->params.egl->glFinish();
|
||||
self->params.egl->glFlush();
|
||||
self->params.egl->glFinish();
|
||||
|
||||
gsr_color_conversion_draw(color_conversion, cap_nvfbc->setup_params.dwTextures[grab_params.dwTextureIndex],
|
||||
gsr_color_conversion_draw(color_conversion, self->setup_params.dwTextures[grab_params.dwTextureIndex],
|
||||
(vec2i){0, 0}, (vec2i){frame->width, frame->height},
|
||||
(vec2i){0, 0}, (vec2i){frame->width, frame->height},
|
||||
0.0f, false);
|
||||
|
||||
cap_nvfbc->params.egl->glFlush();
|
||||
cap_nvfbc->params.egl->glFinish();
|
||||
self->params.egl->glFlush();
|
||||
self->params.egl->glFinish();
|
||||
|
||||
return 0;
|
||||
}
|
||||
@@ -429,8 +429,9 @@ static gsr_source_color gsr_capture_nvfbc_get_source_color(gsr_capture *cap) {
|
||||
|
||||
static void gsr_capture_nvfbc_destroy(gsr_capture *cap, AVCodecContext *video_codec_context) {
|
||||
(void)video_codec_context;
|
||||
gsr_capture_nvfbc *cap_nvfbc = cap->priv;
|
||||
gsr_capture_nvfbc_stop(cap_nvfbc);
|
||||
gsr_capture_nvfbc *self = cap->priv;
|
||||
gsr_capture_nvfbc_stop(self);
|
||||
free(cap->priv);
|
||||
free(cap);
|
||||
}
|
||||
|
||||
@@ -465,7 +466,7 @@ gsr_capture* gsr_capture_nvfbc_create(const gsr_capture_nvfbc_params *params) {
|
||||
cap_nvfbc->params = *params;
|
||||
cap_nvfbc->params.display_to_capture = display_to_capture;
|
||||
cap_nvfbc->params.fps = max_int(cap_nvfbc->params.fps, 1);
|
||||
|
||||
|
||||
*cap = (gsr_capture) {
|
||||
.start = gsr_capture_nvfbc_start,
|
||||
.tick = NULL,
|
||||
|
||||
@@ -414,9 +414,9 @@ static void gsr_capture_portal_clear_damage(gsr_capture *cap) {
|
||||
|
||||
static void gsr_capture_portal_destroy(gsr_capture *cap, AVCodecContext *video_codec_context) {
|
||||
(void)video_codec_context;
|
||||
gsr_capture_portal *cap_portal = cap->priv;
|
||||
gsr_capture_portal *self = cap->priv;
|
||||
if(cap->priv) {
|
||||
gsr_capture_portal_stop(cap_portal);
|
||||
gsr_capture_portal_stop(self);
|
||||
free(cap->priv);
|
||||
cap->priv = NULL;
|
||||
}
|
||||
|
||||
@@ -709,7 +709,7 @@ bool gsr_egl_process_event(gsr_egl *self) {
|
||||
|
||||
void gsr_egl_swap_buffers(gsr_egl *self) {
|
||||
/* This uses less cpu than swap buffer on nvidia */
|
||||
// TODO:
|
||||
// TODO: Do these and remove swap
|
||||
//self->glFlush();
|
||||
//self->glFinish();
|
||||
if(self->egl_display) {
|
||||
|
||||
@@ -356,22 +356,22 @@ static gsr_supported_video_codecs gsr_video_encoder_cuda_get_supported_codecs(gs
|
||||
static void gsr_video_encoder_cuda_stop(gsr_video_encoder_cuda *self, AVCodecContext *video_codec_context);
|
||||
|
||||
static bool gsr_video_encoder_cuda_start(gsr_video_encoder *encoder, AVCodecContext *video_codec_context, AVFrame *frame) {
|
||||
gsr_video_encoder_cuda *encoder_cuda = encoder->priv;
|
||||
gsr_video_encoder_cuda *self = encoder->priv;
|
||||
|
||||
const bool overclock = gsr_egl_get_display_server(encoder_cuda->params.egl) == GSR_DISPLAY_SERVER_X11 ? encoder_cuda->params.overclock : false;
|
||||
if(!gsr_cuda_load(&encoder_cuda->cuda, encoder_cuda->params.egl->x11.dpy, overclock)) {
|
||||
const bool overclock = gsr_egl_get_display_server(self->params.egl) == GSR_DISPLAY_SERVER_X11 ? self->params.overclock : false;
|
||||
if(!gsr_cuda_load(&self->cuda, self->params.egl->x11.dpy, overclock)) {
|
||||
fprintf(stderr, "gsr error: gsr_video_encoder_cuda_start: failed to load cuda\n");
|
||||
gsr_video_encoder_cuda_stop(encoder_cuda, video_codec_context);
|
||||
gsr_video_encoder_cuda_stop(self, video_codec_context);
|
||||
return false;
|
||||
}
|
||||
|
||||
if(!gsr_video_encoder_cuda_setup_context(encoder_cuda, video_codec_context)) {
|
||||
gsr_video_encoder_cuda_stop(encoder_cuda, video_codec_context);
|
||||
if(!gsr_video_encoder_cuda_setup_context(self, video_codec_context)) {
|
||||
gsr_video_encoder_cuda_stop(self, video_codec_context);
|
||||
return false;
|
||||
}
|
||||
|
||||
if(!gsr_video_encoder_cuda_setup_textures(encoder_cuda, video_codec_context, frame)) {
|
||||
gsr_video_encoder_cuda_stop(encoder_cuda, video_codec_context);
|
||||
if(!gsr_video_encoder_cuda_setup_textures(self, video_codec_context, frame)) {
|
||||
gsr_video_encoder_cuda_stop(self, video_codec_context);
|
||||
return false;
|
||||
}
|
||||
|
||||
@@ -402,7 +402,7 @@ void gsr_video_encoder_cuda_stop(gsr_video_encoder_cuda *self, AVCodecContext *v
|
||||
}
|
||||
|
||||
static void gsr_video_encoder_cuda_copy_textures_to_frame(gsr_video_encoder *encoder, AVFrame *frame) {
|
||||
gsr_video_encoder_cuda *encoder_cuda = encoder->priv;
|
||||
gsr_video_encoder_cuda *self = encoder->priv;
|
||||
const int div[2] = {1, 2}; // divide UV texture size by 2 because chroma is half size
|
||||
for(int i = 0; i < 2; ++i) {
|
||||
CUDA_MEMCPY2D memcpy_struct;
|
||||
@@ -414,26 +414,26 @@ static void gsr_video_encoder_cuda_copy_textures_to_frame(gsr_video_encoder *enc
|
||||
memcpy_struct.dstY = 0;
|
||||
memcpy_struct.dstMemoryType = CU_MEMORYTYPE_DEVICE;
|
||||
|
||||
memcpy_struct.srcArray = encoder_cuda->mapped_arrays[i];
|
||||
memcpy_struct.srcArray = self->mapped_arrays[i];
|
||||
memcpy_struct.srcPitch = frame->width / div[i];
|
||||
memcpy_struct.dstDevice = (CUdeviceptr)frame->data[i];
|
||||
memcpy_struct.dstPitch = frame->linesize[i];
|
||||
memcpy_struct.WidthInBytes = frame->width * (encoder_cuda->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? 2 : 1);
|
||||
memcpy_struct.WidthInBytes = frame->width * (self->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? 2 : 1);
|
||||
memcpy_struct.Height = frame->height / div[i];
|
||||
// TODO: Remove this copy if possible
|
||||
encoder_cuda->cuda.cuMemcpy2DAsync_v2(&memcpy_struct, encoder_cuda->cuda_stream);
|
||||
self->cuda.cuMemcpy2DAsync_v2(&memcpy_struct, self->cuda_stream);
|
||||
}
|
||||
|
||||
// TODO: needed?
|
||||
encoder_cuda->cuda.cuStreamSynchronize(encoder_cuda->cuda_stream);
|
||||
self->cuda.cuStreamSynchronize(self->cuda_stream);
|
||||
}
|
||||
|
||||
static void gsr_video_encoder_cuda_get_textures(gsr_video_encoder *encoder, unsigned int *textures, int *num_textures, gsr_destination_color *destination_color) {
|
||||
gsr_video_encoder_cuda *encoder_cuda = encoder->priv;
|
||||
textures[0] = encoder_cuda->target_textures[0];
|
||||
textures[1] = encoder_cuda->target_textures[1];
|
||||
gsr_video_encoder_cuda *self = encoder->priv;
|
||||
textures[0] = self->target_textures[0];
|
||||
textures[1] = self->target_textures[1];
|
||||
*num_textures = 2;
|
||||
*destination_color = encoder_cuda->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? GSR_DESTINATION_COLOR_P010 : GSR_DESTINATION_COLOR_NV12;
|
||||
*destination_color = self->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? GSR_DESTINATION_COLOR_P010 : GSR_DESTINATION_COLOR_NV12;
|
||||
}
|
||||
|
||||
static void gsr_video_encoder_cuda_destroy(gsr_video_encoder *encoder, AVCodecContext *video_codec_context) {
|
||||
|
||||
@@ -77,7 +77,7 @@ static gsr_supported_video_codecs gsr_video_encoder_software_get_supported_codec
|
||||
static void gsr_video_encoder_software_stop(gsr_video_encoder_software *self, AVCodecContext *video_codec_context);
|
||||
|
||||
static bool gsr_video_encoder_software_start(gsr_video_encoder *encoder, AVCodecContext *video_codec_context, AVFrame *frame) {
|
||||
gsr_video_encoder_software *encoder_software = encoder->priv;
|
||||
gsr_video_encoder_software *self = encoder->priv;
|
||||
|
||||
video_codec_context->width = FFALIGN(video_codec_context->width, LINESIZE_ALIGNMENT);
|
||||
video_codec_context->height = FFALIGN(video_codec_context->height, 2);
|
||||
@@ -85,8 +85,8 @@ static bool gsr_video_encoder_software_start(gsr_video_encoder *encoder, AVCodec
|
||||
frame->width = video_codec_context->width;
|
||||
frame->height = video_codec_context->height;
|
||||
|
||||
if(!gsr_video_encoder_software_setup_textures(encoder_software, video_codec_context, frame)) {
|
||||
gsr_video_encoder_software_stop(encoder_software, video_codec_context);
|
||||
if(!gsr_video_encoder_software_setup_textures(self, video_codec_context, frame)) {
|
||||
gsr_video_encoder_software_stop(self, video_codec_context);
|
||||
return false;
|
||||
}
|
||||
|
||||
@@ -101,28 +101,28 @@ void gsr_video_encoder_software_stop(gsr_video_encoder_software *self, AVCodecCo
|
||||
}
|
||||
|
||||
static void gsr_video_encoder_software_copy_textures_to_frame(gsr_video_encoder *encoder, AVFrame *frame) {
|
||||
gsr_video_encoder_software *encoder_software = encoder->priv;
|
||||
gsr_video_encoder_software *self = encoder->priv;
|
||||
// TODO: hdr support
|
||||
const unsigned int formats[2] = { GL_RED, GL_RG };
|
||||
for(int i = 0; i < 2; ++i) {
|
||||
encoder_software->params.egl->glBindTexture(GL_TEXTURE_2D, encoder_software->target_textures[i]);
|
||||
self->params.egl->glBindTexture(GL_TEXTURE_2D, self->target_textures[i]);
|
||||
// We could use glGetTexSubImage and then we wouldn't have to use a specific linesize (LINESIZE_ALIGNMENT) that adds padding,
|
||||
// but glGetTexSubImage is only available starting from opengl 4.5.
|
||||
encoder_software->params.egl->glGetTexImage(GL_TEXTURE_2D, 0, formats[i], GL_UNSIGNED_BYTE, frame->data[i]);
|
||||
self->params.egl->glGetTexImage(GL_TEXTURE_2D, 0, formats[i], GL_UNSIGNED_BYTE, frame->data[i]);
|
||||
}
|
||||
encoder_software->params.egl->glBindTexture(GL_TEXTURE_2D, 0);
|
||||
self->params.egl->glBindTexture(GL_TEXTURE_2D, 0);
|
||||
// cap_kms->kms.base.egl->eglSwapBuffers(cap_kms->kms.base.egl->egl_display, cap_kms->kms.base.egl->egl_surface);
|
||||
|
||||
encoder_software->params.egl->glFlush();
|
||||
encoder_software->params.egl->glFinish();
|
||||
self->params.egl->glFlush();
|
||||
self->params.egl->glFinish();
|
||||
}
|
||||
|
||||
static void gsr_video_encoder_software_get_textures(gsr_video_encoder *encoder, unsigned int *textures, int *num_textures, gsr_destination_color *destination_color) {
|
||||
gsr_video_encoder_software *encoder_software = encoder->priv;
|
||||
textures[0] = encoder_software->target_textures[0];
|
||||
textures[1] = encoder_software->target_textures[1];
|
||||
gsr_video_encoder_software *self = encoder->priv;
|
||||
textures[0] = self->target_textures[0];
|
||||
textures[1] = self->target_textures[1];
|
||||
*num_textures = 2;
|
||||
*destination_color = encoder_software->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? GSR_DESTINATION_COLOR_P010 : GSR_DESTINATION_COLOR_NV12;
|
||||
*destination_color = self->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? GSR_DESTINATION_COLOR_P010 : GSR_DESTINATION_COLOR_NV12;
|
||||
}
|
||||
|
||||
static void gsr_video_encoder_software_destroy(gsr_video_encoder *encoder, AVCodecContext *video_codec_context) {
|
||||
|
||||
@@ -299,12 +299,12 @@ static bool get_supported_video_codecs(VADisplay va_dpy, gsr_supported_video_cod
|
||||
}
|
||||
|
||||
static gsr_supported_video_codecs gsr_video_encoder_vaapi_get_supported_codecs(gsr_video_encoder *encoder, bool cleanup) {
|
||||
gsr_video_encoder_vaapi *encoder_vaapi = encoder->priv;
|
||||
gsr_video_encoder_vaapi *self = encoder->priv;
|
||||
gsr_supported_video_codecs supported_video_codecs = {0};
|
||||
|
||||
char render_path[128];
|
||||
if(!gsr_card_path_get_render_path(encoder_vaapi->params.egl->card_path, render_path)) {
|
||||
fprintf(stderr, "gsr error: gsr_video_encoder_vaapi_get_supported_codecs: failed to get /dev/dri/renderDXXX file from %s\n", encoder_vaapi->params.egl->card_path);
|
||||
if(!gsr_card_path_get_render_path(self->params.egl->card_path, render_path)) {
|
||||
fprintf(stderr, "gsr error: gsr_video_encoder_vaapi_get_supported_codecs: failed to get /dev/dri/renderDXXX file from %s\n", self->params.egl->card_path);
|
||||
return supported_video_codecs;
|
||||
}
|
||||
|
||||
@@ -329,13 +329,13 @@ static gsr_supported_video_codecs gsr_video_encoder_vaapi_get_supported_codecs(g
|
||||
static void gsr_video_encoder_vaapi_stop(gsr_video_encoder_vaapi *self, AVCodecContext *video_codec_context);
|
||||
|
||||
static bool gsr_video_encoder_vaapi_start(gsr_video_encoder *encoder, AVCodecContext *video_codec_context, AVFrame *frame) {
|
||||
gsr_video_encoder_vaapi *encoder_vaapi = encoder->priv;
|
||||
gsr_video_encoder_vaapi *self = encoder->priv;
|
||||
|
||||
if(encoder_vaapi->params.egl->gpu_info.vendor == GSR_GPU_VENDOR_AMD && video_codec_context->codec_id == AV_CODEC_ID_HEVC) {
|
||||
if(self->params.egl->gpu_info.vendor == GSR_GPU_VENDOR_AMD && video_codec_context->codec_id == AV_CODEC_ID_HEVC) {
|
||||
// TODO: dont do this if using ffmpeg reports that this is not needed (AMD driver bug that was fixed recently)
|
||||
video_codec_context->width = FFALIGN(video_codec_context->width, 64);
|
||||
video_codec_context->height = FFALIGN(video_codec_context->height, 16);
|
||||
} else if(encoder_vaapi->params.egl->gpu_info.vendor == GSR_GPU_VENDOR_AMD && video_codec_context->codec_id == AV_CODEC_ID_AV1) {
|
||||
} else if(self->params.egl->gpu_info.vendor == GSR_GPU_VENDOR_AMD && video_codec_context->codec_id == AV_CODEC_ID_AV1) {
|
||||
// TODO: Dont do this for VCN 5 and forward which should fix this hardware bug
|
||||
video_codec_context->width = FFALIGN(video_codec_context->width, 64);
|
||||
// AMD driver has special case handling for 1080 height to set it to 1082 instead of 1088 (1080 aligned to 16).
|
||||
@@ -354,13 +354,13 @@ static bool gsr_video_encoder_vaapi_start(gsr_video_encoder *encoder, AVCodecCon
|
||||
frame->width = video_codec_context->width;
|
||||
frame->height = video_codec_context->height;
|
||||
|
||||
if(!gsr_video_encoder_vaapi_setup_context(encoder_vaapi, video_codec_context)) {
|
||||
gsr_video_encoder_vaapi_stop(encoder_vaapi, video_codec_context);
|
||||
if(!gsr_video_encoder_vaapi_setup_context(self, video_codec_context)) {
|
||||
gsr_video_encoder_vaapi_stop(self, video_codec_context);
|
||||
return false;
|
||||
}
|
||||
|
||||
if(!gsr_video_encoder_vaapi_setup_textures(encoder_vaapi, video_codec_context, frame)) {
|
||||
gsr_video_encoder_vaapi_stop(encoder_vaapi, video_codec_context);
|
||||
if(!gsr_video_encoder_vaapi_setup_textures(self, video_codec_context, frame)) {
|
||||
gsr_video_encoder_vaapi_stop(self, video_codec_context);
|
||||
return false;
|
||||
}
|
||||
|
||||
@@ -386,11 +386,11 @@ void gsr_video_encoder_vaapi_stop(gsr_video_encoder_vaapi *self, AVCodecContext
|
||||
}
|
||||
|
||||
static void gsr_video_encoder_vaapi_get_textures(gsr_video_encoder *encoder, unsigned int *textures, int *num_textures, gsr_destination_color *destination_color) {
|
||||
gsr_video_encoder_vaapi *encoder_vaapi = encoder->priv;
|
||||
textures[0] = encoder_vaapi->target_textures[0];
|
||||
textures[1] = encoder_vaapi->target_textures[1];
|
||||
gsr_video_encoder_vaapi *self = encoder->priv;
|
||||
textures[0] = self->target_textures[0];
|
||||
textures[1] = self->target_textures[1];
|
||||
*num_textures = 2;
|
||||
*destination_color = encoder_vaapi->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? GSR_DESTINATION_COLOR_P010 : GSR_DESTINATION_COLOR_NV12;
|
||||
*destination_color = self->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? GSR_DESTINATION_COLOR_P010 : GSR_DESTINATION_COLOR_NV12;
|
||||
}
|
||||
|
||||
static void gsr_video_encoder_vaapi_destroy(gsr_video_encoder *encoder, AVCodecContext *video_codec_context) {
|
||||
|
||||
@@ -86,11 +86,11 @@ void gsr_video_encoder_vulkan_stop(gsr_video_encoder_vulkan *self, AVCodecContex
|
||||
}
|
||||
|
||||
static void gsr_video_encoder_vulkan_get_textures(gsr_video_encoder *encoder, unsigned int *textures, int *num_textures, gsr_destination_color *destination_color) {
|
||||
gsr_video_encoder_vulkan *encoder_vaapi = encoder->priv;
|
||||
textures[0] = encoder_vaapi->target_textures[0];
|
||||
textures[1] = encoder_vaapi->target_textures[1];
|
||||
gsr_video_encoder_vulkan *self = encoder->priv;
|
||||
textures[0] = self->target_textures[0];
|
||||
textures[1] = self->target_textures[1];
|
||||
*num_textures = 2;
|
||||
*destination_color = encoder_vaapi->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? GSR_DESTINATION_COLOR_P010 : GSR_DESTINATION_COLOR_NV12;
|
||||
*destination_color = self->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? GSR_DESTINATION_COLOR_P010 : GSR_DESTINATION_COLOR_NV12;
|
||||
}
|
||||
|
||||
static void gsr_video_encoder_vulkan_destroy(gsr_video_encoder *encoder, AVCodecContext *video_codec_context) {
|
||||
@@ -104,20 +104,20 @@ gsr_video_encoder* gsr_video_encoder_vulkan_create(const gsr_video_encoder_vulka
|
||||
if(!encoder)
|
||||
return NULL;
|
||||
|
||||
gsr_video_encoder_vulkan *encoder_vaapi = calloc(1, sizeof(gsr_video_encoder_vulkan));
|
||||
if(!encoder_vaapi) {
|
||||
gsr_video_encoder_vulkan *encoder_vulkan = calloc(1, sizeof(gsr_video_encoder_vulkan));
|
||||
if(!encoder_vulkan) {
|
||||
free(encoder);
|
||||
return NULL;
|
||||
}
|
||||
|
||||
encoder_vaapi->params = *params;
|
||||
encoder_vulkan->params = *params;
|
||||
|
||||
*encoder = (gsr_video_encoder) {
|
||||
.start = gsr_video_encoder_vulkan_start,
|
||||
.copy_textures_to_frame = NULL,
|
||||
.get_textures = gsr_video_encoder_vulkan_get_textures,
|
||||
.destroy = gsr_video_encoder_vulkan_destroy,
|
||||
.priv = encoder_vaapi
|
||||
.priv = encoder_vulkan
|
||||
};
|
||||
|
||||
return encoder;
|
||||
|
||||
Reference in New Issue
Block a user