Name capture/encoder variable self

This commit is contained in:
dec05eba
2024-09-26 02:36:31 +02:00
parent aa2fa1e17e
commit 0b20a46e58
8 changed files with 161 additions and 159 deletions

View File

@@ -16,6 +16,7 @@ src = [
'src/encoder/video/video.c', 'src/encoder/video/video.c',
'src/encoder/video/cuda.c', 'src/encoder/video/cuda.c',
'src/encoder/video/vaapi.c', 'src/encoder/video/vaapi.c',
'src/encoder/video/vulkan.c',
'src/encoder/video/software.c', 'src/encoder/video/software.c',
'src/egl.c', 'src/egl.c',
'src/cuda.c', 'src/cuda.c',

View File

@@ -102,7 +102,7 @@ static void set_func_ptr(void **dst, void *src) {
} }
static bool gsr_capture_nvfbc_load_library(gsr_capture *cap) { static bool gsr_capture_nvfbc_load_library(gsr_capture *cap) {
gsr_capture_nvfbc *cap_nvfbc = cap->priv; gsr_capture_nvfbc *self = cap->priv;
dlerror(); /* clear */ dlerror(); /* clear */
void *lib = dlopen("libnvidia-fbc.so.1", RTLD_LAZY); void *lib = dlopen("libnvidia-fbc.so.1", RTLD_LAZY);
@@ -111,23 +111,23 @@ static bool gsr_capture_nvfbc_load_library(gsr_capture *cap) {
return false; return false;
} }
set_func_ptr((void**)&cap_nvfbc->nv_fbc_create_instance, dlsym(lib, "NvFBCCreateInstance")); set_func_ptr((void**)&self->nv_fbc_create_instance, dlsym(lib, "NvFBCCreateInstance"));
if(!cap_nvfbc->nv_fbc_create_instance) { if(!self->nv_fbc_create_instance) {
fprintf(stderr, "gsr error: unable to resolve symbol 'NvFBCCreateInstance'\n"); fprintf(stderr, "gsr error: unable to resolve symbol 'NvFBCCreateInstance'\n");
dlclose(lib); dlclose(lib);
return false; return false;
} }
memset(&cap_nvfbc->nv_fbc_function_list, 0, sizeof(cap_nvfbc->nv_fbc_function_list)); memset(&self->nv_fbc_function_list, 0, sizeof(self->nv_fbc_function_list));
cap_nvfbc->nv_fbc_function_list.dwVersion = NVFBC_VERSION; self->nv_fbc_function_list.dwVersion = NVFBC_VERSION;
NVFBCSTATUS status = cap_nvfbc->nv_fbc_create_instance(&cap_nvfbc->nv_fbc_function_list); NVFBCSTATUS status = self->nv_fbc_create_instance(&self->nv_fbc_function_list);
if(status != NVFBC_SUCCESS) { if(status != NVFBC_SUCCESS) {
fprintf(stderr, "gsr error: failed to create NvFBC instance (status: %d)\n", status); fprintf(stderr, "gsr error: failed to create NvFBC instance (status: %d)\n", status);
dlclose(lib); dlclose(lib);
return false; return false;
} }
cap_nvfbc->library = lib; self->library = lib;
return true; return true;
} }
@@ -153,64 +153,64 @@ static void set_vertical_sync_enabled(gsr_egl *egl, int enabled) {
fprintf(stderr, "gsr warning: setting vertical sync failed\n"); fprintf(stderr, "gsr warning: setting vertical sync failed\n");
} }
static void gsr_capture_nvfbc_destroy_session(gsr_capture_nvfbc *cap_nvfbc) { static void gsr_capture_nvfbc_destroy_session(gsr_capture_nvfbc *self) {
if(cap_nvfbc->fbc_handle_created && cap_nvfbc->capture_session_created) { if(self->fbc_handle_created && self->capture_session_created) {
NVFBC_DESTROY_CAPTURE_SESSION_PARAMS destroy_capture_params; NVFBC_DESTROY_CAPTURE_SESSION_PARAMS destroy_capture_params;
memset(&destroy_capture_params, 0, sizeof(destroy_capture_params)); memset(&destroy_capture_params, 0, sizeof(destroy_capture_params));
destroy_capture_params.dwVersion = NVFBC_DESTROY_CAPTURE_SESSION_PARAMS_VER; destroy_capture_params.dwVersion = NVFBC_DESTROY_CAPTURE_SESSION_PARAMS_VER;
cap_nvfbc->nv_fbc_function_list.nvFBCDestroyCaptureSession(cap_nvfbc->nv_fbc_handle, &destroy_capture_params); self->nv_fbc_function_list.nvFBCDestroyCaptureSession(self->nv_fbc_handle, &destroy_capture_params);
cap_nvfbc->capture_session_created = false; self->capture_session_created = false;
} }
} }
static void gsr_capture_nvfbc_destroy_handle(gsr_capture_nvfbc *cap_nvfbc) { static void gsr_capture_nvfbc_destroy_handle(gsr_capture_nvfbc *self) {
if(cap_nvfbc->fbc_handle_created) { if(self->fbc_handle_created) {
NVFBC_DESTROY_HANDLE_PARAMS destroy_params; NVFBC_DESTROY_HANDLE_PARAMS destroy_params;
memset(&destroy_params, 0, sizeof(destroy_params)); memset(&destroy_params, 0, sizeof(destroy_params));
destroy_params.dwVersion = NVFBC_DESTROY_HANDLE_PARAMS_VER; destroy_params.dwVersion = NVFBC_DESTROY_HANDLE_PARAMS_VER;
cap_nvfbc->nv_fbc_function_list.nvFBCDestroyHandle(cap_nvfbc->nv_fbc_handle, &destroy_params); self->nv_fbc_function_list.nvFBCDestroyHandle(self->nv_fbc_handle, &destroy_params);
cap_nvfbc->fbc_handle_created = false; self->fbc_handle_created = false;
cap_nvfbc->nv_fbc_handle = 0; self->nv_fbc_handle = 0;
} }
} }
static void gsr_capture_nvfbc_destroy_session_and_handle(gsr_capture_nvfbc *cap_nvfbc) { static void gsr_capture_nvfbc_destroy_session_and_handle(gsr_capture_nvfbc *self) {
gsr_capture_nvfbc_destroy_session(cap_nvfbc); gsr_capture_nvfbc_destroy_session(self);
gsr_capture_nvfbc_destroy_handle(cap_nvfbc); gsr_capture_nvfbc_destroy_handle(self);
} }
static int gsr_capture_nvfbc_setup_handle(gsr_capture_nvfbc *cap_nvfbc) { static int gsr_capture_nvfbc_setup_handle(gsr_capture_nvfbc *self) {
NVFBCSTATUS status; NVFBCSTATUS status;
NVFBC_CREATE_HANDLE_PARAMS create_params; NVFBC_CREATE_HANDLE_PARAMS create_params;
memset(&create_params, 0, sizeof(create_params)); memset(&create_params, 0, sizeof(create_params));
create_params.dwVersion = NVFBC_CREATE_HANDLE_PARAMS_VER; create_params.dwVersion = NVFBC_CREATE_HANDLE_PARAMS_VER;
create_params.bExternallyManagedContext = NVFBC_TRUE; create_params.bExternallyManagedContext = NVFBC_TRUE;
create_params.glxCtx = cap_nvfbc->params.egl->glx_context; create_params.glxCtx = self->params.egl->glx_context;
create_params.glxFBConfig = cap_nvfbc->params.egl->glx_fb_config; create_params.glxFBConfig = self->params.egl->glx_fb_config;
status = cap_nvfbc->nv_fbc_function_list.nvFBCCreateHandle(&cap_nvfbc->nv_fbc_handle, &create_params); status = self->nv_fbc_function_list.nvFBCCreateHandle(&self->nv_fbc_handle, &create_params);
if(status != NVFBC_SUCCESS) { if(status != NVFBC_SUCCESS) {
// Reverse engineering for interoperability // Reverse engineering for interoperability
const uint8_t enable_key[] = { 0xac, 0x10, 0xc9, 0x2e, 0xa5, 0xe6, 0x87, 0x4f, 0x8f, 0x4b, 0xf4, 0x61, 0xf8, 0x56, 0x27, 0xe9 }; const uint8_t enable_key[] = { 0xac, 0x10, 0xc9, 0x2e, 0xa5, 0xe6, 0x87, 0x4f, 0x8f, 0x4b, 0xf4, 0x61, 0xf8, 0x56, 0x27, 0xe9 };
create_params.privateData = enable_key; create_params.privateData = enable_key;
create_params.privateDataSize = 16; create_params.privateDataSize = 16;
status = cap_nvfbc->nv_fbc_function_list.nvFBCCreateHandle(&cap_nvfbc->nv_fbc_handle, &create_params); status = self->nv_fbc_function_list.nvFBCCreateHandle(&self->nv_fbc_handle, &create_params);
if(status != NVFBC_SUCCESS) { if(status != NVFBC_SUCCESS) {
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: %s\n", cap_nvfbc->nv_fbc_function_list.nvFBCGetLastErrorStr(cap_nvfbc->nv_fbc_handle)); fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: %s\n", self->nv_fbc_function_list.nvFBCGetLastErrorStr(self->nv_fbc_handle));
goto error_cleanup; goto error_cleanup;
} }
} }
cap_nvfbc->fbc_handle_created = true; self->fbc_handle_created = true;
NVFBC_GET_STATUS_PARAMS status_params; NVFBC_GET_STATUS_PARAMS status_params;
memset(&status_params, 0, sizeof(status_params)); memset(&status_params, 0, sizeof(status_params));
status_params.dwVersion = NVFBC_GET_STATUS_PARAMS_VER; status_params.dwVersion = NVFBC_GET_STATUS_PARAMS_VER;
status = cap_nvfbc->nv_fbc_function_list.nvFBCGetStatus(cap_nvfbc->nv_fbc_handle, &status_params); status = self->nv_fbc_function_list.nvFBCGetStatus(self->nv_fbc_handle, &status_params);
if(status != NVFBC_SUCCESS) { if(status != NVFBC_SUCCESS) {
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: %s\n", cap_nvfbc->nv_fbc_function_list.nvFBCGetLastErrorStr(cap_nvfbc->nv_fbc_handle)); fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: %s\n", self->nv_fbc_function_list.nvFBCGetLastErrorStr(self->nv_fbc_handle));
goto error_cleanup; goto error_cleanup;
} }
@@ -219,10 +219,10 @@ static int gsr_capture_nvfbc_setup_handle(gsr_capture_nvfbc *cap_nvfbc) {
goto error_cleanup; goto error_cleanup;
} }
cap_nvfbc->tracking_width = XWidthOfScreen(DefaultScreenOfDisplay(cap_nvfbc->params.egl->x11.dpy)); self->tracking_width = XWidthOfScreen(DefaultScreenOfDisplay(self->params.egl->x11.dpy));
cap_nvfbc->tracking_height = XHeightOfScreen(DefaultScreenOfDisplay(cap_nvfbc->params.egl->x11.dpy)); self->tracking_height = XHeightOfScreen(DefaultScreenOfDisplay(self->params.egl->x11.dpy));
cap_nvfbc->tracking_type = strcmp(cap_nvfbc->params.display_to_capture, "screen") == 0 ? NVFBC_TRACKING_SCREEN : NVFBC_TRACKING_OUTPUT; self->tracking_type = strcmp(self->params.display_to_capture, "screen") == 0 ? NVFBC_TRACKING_SCREEN : NVFBC_TRACKING_OUTPUT;
if(cap_nvfbc->tracking_type == NVFBC_TRACKING_OUTPUT) { if(self->tracking_type == NVFBC_TRACKING_OUTPUT) {
if(!status_params.bXRandRAvailable) { if(!status_params.bXRandRAvailable) {
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: the xrandr extension is not available\n"); fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: the xrandr extension is not available\n");
goto error_cleanup; goto error_cleanup;
@@ -233,9 +233,9 @@ static int gsr_capture_nvfbc_setup_handle(gsr_capture_nvfbc *cap_nvfbc) {
goto error_cleanup; goto error_cleanup;
} }
cap_nvfbc->output_id = get_output_id_from_display_name(status_params.outputs, status_params.dwOutputNum, cap_nvfbc->params.display_to_capture, &cap_nvfbc->tracking_width, &cap_nvfbc->tracking_height); self->output_id = get_output_id_from_display_name(status_params.outputs, status_params.dwOutputNum, self->params.display_to_capture, &self->tracking_width, &self->tracking_height);
if(cap_nvfbc->output_id == 0) { if(self->output_id == 0) {
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: display '%s' not found\n", cap_nvfbc->params.display_to_capture); fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: display '%s' not found\n", self->params.display_to_capture);
goto error_cleanup; goto error_cleanup;
} }
} }
@@ -243,83 +243,83 @@ static int gsr_capture_nvfbc_setup_handle(gsr_capture_nvfbc *cap_nvfbc) {
return 0; return 0;
error_cleanup: error_cleanup:
gsr_capture_nvfbc_destroy_session_and_handle(cap_nvfbc); gsr_capture_nvfbc_destroy_session_and_handle(self);
return -1; return -1;
} }
static int gsr_capture_nvfbc_setup_session(gsr_capture_nvfbc *cap_nvfbc) { static int gsr_capture_nvfbc_setup_session(gsr_capture_nvfbc *self) {
NVFBC_CREATE_CAPTURE_SESSION_PARAMS create_capture_params; NVFBC_CREATE_CAPTURE_SESSION_PARAMS create_capture_params;
memset(&create_capture_params, 0, sizeof(create_capture_params)); memset(&create_capture_params, 0, sizeof(create_capture_params));
create_capture_params.dwVersion = NVFBC_CREATE_CAPTURE_SESSION_PARAMS_VER; create_capture_params.dwVersion = NVFBC_CREATE_CAPTURE_SESSION_PARAMS_VER;
create_capture_params.eCaptureType = NVFBC_CAPTURE_TO_GL; create_capture_params.eCaptureType = NVFBC_CAPTURE_TO_GL;
create_capture_params.bWithCursor = (!cap_nvfbc->params.direct_capture || cap_nvfbc->supports_direct_cursor) ? NVFBC_TRUE : NVFBC_FALSE; create_capture_params.bWithCursor = (!self->params.direct_capture || self->supports_direct_cursor) ? NVFBC_TRUE : NVFBC_FALSE;
if(!cap_nvfbc->params.record_cursor) if(!self->params.record_cursor)
create_capture_params.bWithCursor = false; create_capture_params.bWithCursor = false;
if(cap_nvfbc->capture_region) if(self->capture_region)
create_capture_params.captureBox = (NVFBC_BOX){ cap_nvfbc->x, cap_nvfbc->y, cap_nvfbc->width, cap_nvfbc->height }; create_capture_params.captureBox = (NVFBC_BOX){ self->x, self->y, self->width, self->height };
create_capture_params.eTrackingType = cap_nvfbc->tracking_type; create_capture_params.eTrackingType = self->tracking_type;
create_capture_params.dwSamplingRateMs = (uint32_t)ceilf(1000.0f / (float)cap_nvfbc->params.fps); create_capture_params.dwSamplingRateMs = (uint32_t)ceilf(1000.0f / (float)self->params.fps);
create_capture_params.bAllowDirectCapture = cap_nvfbc->params.direct_capture ? NVFBC_TRUE : NVFBC_FALSE; create_capture_params.bAllowDirectCapture = self->params.direct_capture ? NVFBC_TRUE : NVFBC_FALSE;
create_capture_params.bPushModel = cap_nvfbc->params.direct_capture ? NVFBC_TRUE : NVFBC_FALSE; create_capture_params.bPushModel = self->params.direct_capture ? NVFBC_TRUE : NVFBC_FALSE;
create_capture_params.bDisableAutoModesetRecovery = true; create_capture_params.bDisableAutoModesetRecovery = true;
if(cap_nvfbc->tracking_type == NVFBC_TRACKING_OUTPUT) if(self->tracking_type == NVFBC_TRACKING_OUTPUT)
create_capture_params.dwOutputId = cap_nvfbc->output_id; create_capture_params.dwOutputId = self->output_id;
NVFBCSTATUS status = cap_nvfbc->nv_fbc_function_list.nvFBCCreateCaptureSession(cap_nvfbc->nv_fbc_handle, &create_capture_params); NVFBCSTATUS status = self->nv_fbc_function_list.nvFBCCreateCaptureSession(self->nv_fbc_handle, &create_capture_params);
if(status != NVFBC_SUCCESS) { if(status != NVFBC_SUCCESS) {
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: %s\n", cap_nvfbc->nv_fbc_function_list.nvFBCGetLastErrorStr(cap_nvfbc->nv_fbc_handle)); fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: %s\n", self->nv_fbc_function_list.nvFBCGetLastErrorStr(self->nv_fbc_handle));
return -1; return -1;
} }
cap_nvfbc->capture_session_created = true; self->capture_session_created = true;
memset(&cap_nvfbc->setup_params, 0, sizeof(cap_nvfbc->setup_params)); memset(&self->setup_params, 0, sizeof(self->setup_params));
cap_nvfbc->setup_params.dwVersion = NVFBC_TOGL_SETUP_PARAMS_VER; self->setup_params.dwVersion = NVFBC_TOGL_SETUP_PARAMS_VER;
cap_nvfbc->setup_params.eBufferFormat = NVFBC_BUFFER_FORMAT_BGRA; self->setup_params.eBufferFormat = NVFBC_BUFFER_FORMAT_BGRA;
status = cap_nvfbc->nv_fbc_function_list.nvFBCToGLSetUp(cap_nvfbc->nv_fbc_handle, &cap_nvfbc->setup_params); status = self->nv_fbc_function_list.nvFBCToGLSetUp(self->nv_fbc_handle, &self->setup_params);
if(status != NVFBC_SUCCESS) { if(status != NVFBC_SUCCESS) {
fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: %s\n", cap_nvfbc->nv_fbc_function_list.nvFBCGetLastErrorStr(cap_nvfbc->nv_fbc_handle)); fprintf(stderr, "gsr error: gsr_capture_nvfbc_start failed: %s\n", self->nv_fbc_function_list.nvFBCGetLastErrorStr(self->nv_fbc_handle));
gsr_capture_nvfbc_destroy_session(cap_nvfbc); gsr_capture_nvfbc_destroy_session(self);
return -1; return -1;
} }
return 0; return 0;
} }
static void gsr_capture_nvfbc_stop(gsr_capture_nvfbc *cap_nvfbc) { static void gsr_capture_nvfbc_stop(gsr_capture_nvfbc *self) {
gsr_capture_nvfbc_destroy_session_and_handle(cap_nvfbc); gsr_capture_nvfbc_destroy_session_and_handle(self);
if(cap_nvfbc->library) { if(self->library) {
dlclose(cap_nvfbc->library); dlclose(self->library);
cap_nvfbc->library = NULL; self->library = NULL;
} }
if(cap_nvfbc->params.display_to_capture) { if(self->params.display_to_capture) {
free((void*)cap_nvfbc->params.display_to_capture); free((void*)self->params.display_to_capture);
cap_nvfbc->params.display_to_capture = NULL; self->params.display_to_capture = NULL;
} }
} }
static int gsr_capture_nvfbc_start(gsr_capture *cap, AVCodecContext *video_codec_context, AVFrame *frame) { static int gsr_capture_nvfbc_start(gsr_capture *cap, AVCodecContext *video_codec_context, AVFrame *frame) {
gsr_capture_nvfbc *cap_nvfbc = cap->priv; gsr_capture_nvfbc *self = cap->priv;
if(!gsr_capture_nvfbc_load_library(cap)) if(!gsr_capture_nvfbc_load_library(cap))
return -1; return -1;
cap_nvfbc->x = max_int(cap_nvfbc->params.pos.x, 0); self->x = max_int(self->params.pos.x, 0);
cap_nvfbc->y = max_int(cap_nvfbc->params.pos.y, 0); self->y = max_int(self->params.pos.y, 0);
cap_nvfbc->width = max_int(cap_nvfbc->params.size.x, 0); self->width = max_int(self->params.size.x, 0);
cap_nvfbc->height = max_int(cap_nvfbc->params.size.y, 0); self->height = max_int(self->params.size.y, 0);
cap_nvfbc->capture_region = (cap_nvfbc->x > 0 || cap_nvfbc->y > 0 || cap_nvfbc->width > 0 || cap_nvfbc->height > 0); self->capture_region = (self->x > 0 || self->y > 0 || self->width > 0 || self->height > 0);
cap_nvfbc->supports_direct_cursor = false; self->supports_direct_cursor = false;
int driver_major_version = 0; int driver_major_version = 0;
int driver_minor_version = 0; int driver_minor_version = 0;
if(cap_nvfbc->params.direct_capture && get_driver_version(&driver_major_version, &driver_minor_version)) { if(self->params.direct_capture && get_driver_version(&driver_major_version, &driver_minor_version)) {
fprintf(stderr, "Info: detected nvidia version: %d.%d\n", driver_major_version, driver_minor_version); fprintf(stderr, "Info: detected nvidia version: %d.%d\n", driver_major_version, driver_minor_version);
// TODO: // TODO:
if(version_at_least(driver_major_version, driver_minor_version, 515, 57) && version_less_than(driver_major_version, driver_minor_version, 520, 56)) { if(version_at_least(driver_major_version, driver_minor_version, 515, 57) && version_less_than(driver_major_version, driver_minor_version, 520, 56)) {
cap_nvfbc->params.direct_capture = false; self->params.direct_capture = false;
fprintf(stderr, "Warning: \"screen-direct\" has temporary been disabled as it causes stuttering with driver versions >= 515.57 and < 520.56. Please update your driver if possible. Capturing \"screen\" instead.\n"); fprintf(stderr, "Warning: \"screen-direct\" has temporary been disabled as it causes stuttering with driver versions >= 515.57 and < 520.56. Please update your driver if possible. Capturing \"screen\" instead.\n");
} }
@@ -328,63 +328,63 @@ static int gsr_capture_nvfbc_start(gsr_capture *cap, AVCodecContext *video_codec
/* /*
if(direct_capture) { if(direct_capture) {
if(version_at_least(driver_major_version, driver_minor_version, 515, 57)) if(version_at_least(driver_major_version, driver_minor_version, 515, 57))
cap_nvfbc->supports_direct_cursor = true; self->supports_direct_cursor = true;
else else
fprintf(stderr, "Info: capturing \"screen-direct\" but driver version appears to be less than 515.57. Disabling capture of cursor. Please update your driver if you want to capture your cursor or record \"screen\" instead.\n"); fprintf(stderr, "Info: capturing \"screen-direct\" but driver version appears to be less than 515.57. Disabling capture of cursor. Please update your driver if you want to capture your cursor or record \"screen\" instead.\n");
} }
*/ */
} }
if(gsr_capture_nvfbc_setup_handle(cap_nvfbc) != 0) { if(gsr_capture_nvfbc_setup_handle(self) != 0) {
goto error_cleanup; goto error_cleanup;
} }
if(gsr_capture_nvfbc_setup_session(cap_nvfbc) != 0) { if(gsr_capture_nvfbc_setup_session(self) != 0) {
goto error_cleanup; goto error_cleanup;
} }
if(cap_nvfbc->capture_region) { if(self->capture_region) {
video_codec_context->width = FFALIGN(cap_nvfbc->width, 2); video_codec_context->width = FFALIGN(self->width, 2);
video_codec_context->height = FFALIGN(cap_nvfbc->height, 2); video_codec_context->height = FFALIGN(self->height, 2);
} else { } else {
video_codec_context->width = FFALIGN(cap_nvfbc->tracking_width, 2); video_codec_context->width = FFALIGN(self->tracking_width, 2);
video_codec_context->height = FFALIGN(cap_nvfbc->tracking_height, 2); video_codec_context->height = FFALIGN(self->tracking_height, 2);
} }
frame->width = video_codec_context->width; frame->width = video_codec_context->width;
frame->height = video_codec_context->height; frame->height = video_codec_context->height;
/* Disable vsync */ /* Disable vsync */
set_vertical_sync_enabled(cap_nvfbc->params.egl, 0); set_vertical_sync_enabled(self->params.egl, 0);
return 0; return 0;
error_cleanup: error_cleanup:
gsr_capture_nvfbc_stop(cap_nvfbc); gsr_capture_nvfbc_stop(self);
return -1; return -1;
} }
static int gsr_capture_nvfbc_capture(gsr_capture *cap, AVFrame *frame, gsr_color_conversion *color_conversion) { static int gsr_capture_nvfbc_capture(gsr_capture *cap, AVFrame *frame, gsr_color_conversion *color_conversion) {
gsr_capture_nvfbc *cap_nvfbc = cap->priv; gsr_capture_nvfbc *self = cap->priv;
const double nvfbc_recreate_retry_time_seconds = 1.0; const double nvfbc_recreate_retry_time_seconds = 1.0;
if(cap_nvfbc->nvfbc_needs_recreate) { if(self->nvfbc_needs_recreate) {
const double now = clock_get_monotonic_seconds(); const double now = clock_get_monotonic_seconds();
if(now - cap_nvfbc->nvfbc_dead_start >= nvfbc_recreate_retry_time_seconds) { if(now - self->nvfbc_dead_start >= nvfbc_recreate_retry_time_seconds) {
cap_nvfbc->nvfbc_dead_start = now; self->nvfbc_dead_start = now;
gsr_capture_nvfbc_destroy_session_and_handle(cap_nvfbc); gsr_capture_nvfbc_destroy_session_and_handle(self);
if(gsr_capture_nvfbc_setup_handle(cap_nvfbc) != 0) { if(gsr_capture_nvfbc_setup_handle(self) != 0) {
fprintf(stderr, "gsr error: gsr_capture_nvfbc_capture failed to recreate nvfbc handle, trying again in %f second(s)\n", nvfbc_recreate_retry_time_seconds); fprintf(stderr, "gsr error: gsr_capture_nvfbc_capture failed to recreate nvfbc handle, trying again in %f second(s)\n", nvfbc_recreate_retry_time_seconds);
return -1; return -1;
} }
if(gsr_capture_nvfbc_setup_session(cap_nvfbc) != 0) { if(gsr_capture_nvfbc_setup_session(self) != 0) {
fprintf(stderr, "gsr error: gsr_capture_nvfbc_capture failed to recreate nvfbc session, trying again in %f second(s)\n", nvfbc_recreate_retry_time_seconds); fprintf(stderr, "gsr error: gsr_capture_nvfbc_capture failed to recreate nvfbc session, trying again in %f second(s)\n", nvfbc_recreate_retry_time_seconds);
return -1; return -1;
} }
cap_nvfbc->nvfbc_needs_recreate = false; self->nvfbc_needs_recreate = false;
} else { } else {
return 0; return 0;
} }
@@ -400,24 +400,24 @@ static int gsr_capture_nvfbc_capture(gsr_capture *cap, AVFrame *frame, gsr_color
grab_params.pFrameGrabInfo = &frame_info; grab_params.pFrameGrabInfo = &frame_info;
grab_params.dwTimeoutMs = 0; grab_params.dwTimeoutMs = 0;
NVFBCSTATUS status = cap_nvfbc->nv_fbc_function_list.nvFBCToGLGrabFrame(cap_nvfbc->nv_fbc_handle, &grab_params); NVFBCSTATUS status = self->nv_fbc_function_list.nvFBCToGLGrabFrame(self->nv_fbc_handle, &grab_params);
if(status != NVFBC_SUCCESS) { if(status != NVFBC_SUCCESS) {
fprintf(stderr, "gsr error: gsr_capture_nvfbc_capture failed: %s (%d), recreating session after %f second(s)\n", cap_nvfbc->nv_fbc_function_list.nvFBCGetLastErrorStr(cap_nvfbc->nv_fbc_handle), status, nvfbc_recreate_retry_time_seconds); fprintf(stderr, "gsr error: gsr_capture_nvfbc_capture failed: %s (%d), recreating session after %f second(s)\n", self->nv_fbc_function_list.nvFBCGetLastErrorStr(self->nv_fbc_handle), status, nvfbc_recreate_retry_time_seconds);
cap_nvfbc->nvfbc_needs_recreate = true; self->nvfbc_needs_recreate = true;
cap_nvfbc->nvfbc_dead_start = clock_get_monotonic_seconds(); self->nvfbc_dead_start = clock_get_monotonic_seconds();
return 0; return 0;
} }
cap_nvfbc->params.egl->glFlush(); self->params.egl->glFlush();
cap_nvfbc->params.egl->glFinish(); self->params.egl->glFinish();
gsr_color_conversion_draw(color_conversion, cap_nvfbc->setup_params.dwTextures[grab_params.dwTextureIndex], gsr_color_conversion_draw(color_conversion, self->setup_params.dwTextures[grab_params.dwTextureIndex],
(vec2i){0, 0}, (vec2i){frame->width, frame->height}, (vec2i){0, 0}, (vec2i){frame->width, frame->height},
(vec2i){0, 0}, (vec2i){frame->width, frame->height}, (vec2i){0, 0}, (vec2i){frame->width, frame->height},
0.0f, false); 0.0f, false);
cap_nvfbc->params.egl->glFlush(); self->params.egl->glFlush();
cap_nvfbc->params.egl->glFinish(); self->params.egl->glFinish();
return 0; return 0;
} }
@@ -429,8 +429,9 @@ static gsr_source_color gsr_capture_nvfbc_get_source_color(gsr_capture *cap) {
static void gsr_capture_nvfbc_destroy(gsr_capture *cap, AVCodecContext *video_codec_context) { static void gsr_capture_nvfbc_destroy(gsr_capture *cap, AVCodecContext *video_codec_context) {
(void)video_codec_context; (void)video_codec_context;
gsr_capture_nvfbc *cap_nvfbc = cap->priv; gsr_capture_nvfbc *self = cap->priv;
gsr_capture_nvfbc_stop(cap_nvfbc); gsr_capture_nvfbc_stop(self);
free(cap->priv);
free(cap); free(cap);
} }
@@ -465,7 +466,7 @@ gsr_capture* gsr_capture_nvfbc_create(const gsr_capture_nvfbc_params *params) {
cap_nvfbc->params = *params; cap_nvfbc->params = *params;
cap_nvfbc->params.display_to_capture = display_to_capture; cap_nvfbc->params.display_to_capture = display_to_capture;
cap_nvfbc->params.fps = max_int(cap_nvfbc->params.fps, 1); cap_nvfbc->params.fps = max_int(cap_nvfbc->params.fps, 1);
*cap = (gsr_capture) { *cap = (gsr_capture) {
.start = gsr_capture_nvfbc_start, .start = gsr_capture_nvfbc_start,
.tick = NULL, .tick = NULL,

View File

@@ -414,9 +414,9 @@ static void gsr_capture_portal_clear_damage(gsr_capture *cap) {
static void gsr_capture_portal_destroy(gsr_capture *cap, AVCodecContext *video_codec_context) { static void gsr_capture_portal_destroy(gsr_capture *cap, AVCodecContext *video_codec_context) {
(void)video_codec_context; (void)video_codec_context;
gsr_capture_portal *cap_portal = cap->priv; gsr_capture_portal *self = cap->priv;
if(cap->priv) { if(cap->priv) {
gsr_capture_portal_stop(cap_portal); gsr_capture_portal_stop(self);
free(cap->priv); free(cap->priv);
cap->priv = NULL; cap->priv = NULL;
} }

View File

@@ -709,7 +709,7 @@ bool gsr_egl_process_event(gsr_egl *self) {
void gsr_egl_swap_buffers(gsr_egl *self) { void gsr_egl_swap_buffers(gsr_egl *self) {
/* This uses less cpu than swap buffer on nvidia */ /* This uses less cpu than swap buffer on nvidia */
// TODO: // TODO: Do these and remove swap
//self->glFlush(); //self->glFlush();
//self->glFinish(); //self->glFinish();
if(self->egl_display) { if(self->egl_display) {

View File

@@ -356,22 +356,22 @@ static gsr_supported_video_codecs gsr_video_encoder_cuda_get_supported_codecs(gs
static void gsr_video_encoder_cuda_stop(gsr_video_encoder_cuda *self, AVCodecContext *video_codec_context); static void gsr_video_encoder_cuda_stop(gsr_video_encoder_cuda *self, AVCodecContext *video_codec_context);
static bool gsr_video_encoder_cuda_start(gsr_video_encoder *encoder, AVCodecContext *video_codec_context, AVFrame *frame) { static bool gsr_video_encoder_cuda_start(gsr_video_encoder *encoder, AVCodecContext *video_codec_context, AVFrame *frame) {
gsr_video_encoder_cuda *encoder_cuda = encoder->priv; gsr_video_encoder_cuda *self = encoder->priv;
const bool overclock = gsr_egl_get_display_server(encoder_cuda->params.egl) == GSR_DISPLAY_SERVER_X11 ? encoder_cuda->params.overclock : false; const bool overclock = gsr_egl_get_display_server(self->params.egl) == GSR_DISPLAY_SERVER_X11 ? self->params.overclock : false;
if(!gsr_cuda_load(&encoder_cuda->cuda, encoder_cuda->params.egl->x11.dpy, overclock)) { if(!gsr_cuda_load(&self->cuda, self->params.egl->x11.dpy, overclock)) {
fprintf(stderr, "gsr error: gsr_video_encoder_cuda_start: failed to load cuda\n"); fprintf(stderr, "gsr error: gsr_video_encoder_cuda_start: failed to load cuda\n");
gsr_video_encoder_cuda_stop(encoder_cuda, video_codec_context); gsr_video_encoder_cuda_stop(self, video_codec_context);
return false; return false;
} }
if(!gsr_video_encoder_cuda_setup_context(encoder_cuda, video_codec_context)) { if(!gsr_video_encoder_cuda_setup_context(self, video_codec_context)) {
gsr_video_encoder_cuda_stop(encoder_cuda, video_codec_context); gsr_video_encoder_cuda_stop(self, video_codec_context);
return false; return false;
} }
if(!gsr_video_encoder_cuda_setup_textures(encoder_cuda, video_codec_context, frame)) { if(!gsr_video_encoder_cuda_setup_textures(self, video_codec_context, frame)) {
gsr_video_encoder_cuda_stop(encoder_cuda, video_codec_context); gsr_video_encoder_cuda_stop(self, video_codec_context);
return false; return false;
} }
@@ -402,7 +402,7 @@ void gsr_video_encoder_cuda_stop(gsr_video_encoder_cuda *self, AVCodecContext *v
} }
static void gsr_video_encoder_cuda_copy_textures_to_frame(gsr_video_encoder *encoder, AVFrame *frame) { static void gsr_video_encoder_cuda_copy_textures_to_frame(gsr_video_encoder *encoder, AVFrame *frame) {
gsr_video_encoder_cuda *encoder_cuda = encoder->priv; gsr_video_encoder_cuda *self = encoder->priv;
const int div[2] = {1, 2}; // divide UV texture size by 2 because chroma is half size const int div[2] = {1, 2}; // divide UV texture size by 2 because chroma is half size
for(int i = 0; i < 2; ++i) { for(int i = 0; i < 2; ++i) {
CUDA_MEMCPY2D memcpy_struct; CUDA_MEMCPY2D memcpy_struct;
@@ -414,26 +414,26 @@ static void gsr_video_encoder_cuda_copy_textures_to_frame(gsr_video_encoder *enc
memcpy_struct.dstY = 0; memcpy_struct.dstY = 0;
memcpy_struct.dstMemoryType = CU_MEMORYTYPE_DEVICE; memcpy_struct.dstMemoryType = CU_MEMORYTYPE_DEVICE;
memcpy_struct.srcArray = encoder_cuda->mapped_arrays[i]; memcpy_struct.srcArray = self->mapped_arrays[i];
memcpy_struct.srcPitch = frame->width / div[i]; memcpy_struct.srcPitch = frame->width / div[i];
memcpy_struct.dstDevice = (CUdeviceptr)frame->data[i]; memcpy_struct.dstDevice = (CUdeviceptr)frame->data[i];
memcpy_struct.dstPitch = frame->linesize[i]; memcpy_struct.dstPitch = frame->linesize[i];
memcpy_struct.WidthInBytes = frame->width * (encoder_cuda->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? 2 : 1); memcpy_struct.WidthInBytes = frame->width * (self->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? 2 : 1);
memcpy_struct.Height = frame->height / div[i]; memcpy_struct.Height = frame->height / div[i];
// TODO: Remove this copy if possible // TODO: Remove this copy if possible
encoder_cuda->cuda.cuMemcpy2DAsync_v2(&memcpy_struct, encoder_cuda->cuda_stream); self->cuda.cuMemcpy2DAsync_v2(&memcpy_struct, self->cuda_stream);
} }
// TODO: needed? // TODO: needed?
encoder_cuda->cuda.cuStreamSynchronize(encoder_cuda->cuda_stream); self->cuda.cuStreamSynchronize(self->cuda_stream);
} }
static void gsr_video_encoder_cuda_get_textures(gsr_video_encoder *encoder, unsigned int *textures, int *num_textures, gsr_destination_color *destination_color) { static void gsr_video_encoder_cuda_get_textures(gsr_video_encoder *encoder, unsigned int *textures, int *num_textures, gsr_destination_color *destination_color) {
gsr_video_encoder_cuda *encoder_cuda = encoder->priv; gsr_video_encoder_cuda *self = encoder->priv;
textures[0] = encoder_cuda->target_textures[0]; textures[0] = self->target_textures[0];
textures[1] = encoder_cuda->target_textures[1]; textures[1] = self->target_textures[1];
*num_textures = 2; *num_textures = 2;
*destination_color = encoder_cuda->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? GSR_DESTINATION_COLOR_P010 : GSR_DESTINATION_COLOR_NV12; *destination_color = self->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? GSR_DESTINATION_COLOR_P010 : GSR_DESTINATION_COLOR_NV12;
} }
static void gsr_video_encoder_cuda_destroy(gsr_video_encoder *encoder, AVCodecContext *video_codec_context) { static void gsr_video_encoder_cuda_destroy(gsr_video_encoder *encoder, AVCodecContext *video_codec_context) {

View File

@@ -77,7 +77,7 @@ static gsr_supported_video_codecs gsr_video_encoder_software_get_supported_codec
static void gsr_video_encoder_software_stop(gsr_video_encoder_software *self, AVCodecContext *video_codec_context); static void gsr_video_encoder_software_stop(gsr_video_encoder_software *self, AVCodecContext *video_codec_context);
static bool gsr_video_encoder_software_start(gsr_video_encoder *encoder, AVCodecContext *video_codec_context, AVFrame *frame) { static bool gsr_video_encoder_software_start(gsr_video_encoder *encoder, AVCodecContext *video_codec_context, AVFrame *frame) {
gsr_video_encoder_software *encoder_software = encoder->priv; gsr_video_encoder_software *self = encoder->priv;
video_codec_context->width = FFALIGN(video_codec_context->width, LINESIZE_ALIGNMENT); video_codec_context->width = FFALIGN(video_codec_context->width, LINESIZE_ALIGNMENT);
video_codec_context->height = FFALIGN(video_codec_context->height, 2); video_codec_context->height = FFALIGN(video_codec_context->height, 2);
@@ -85,8 +85,8 @@ static bool gsr_video_encoder_software_start(gsr_video_encoder *encoder, AVCodec
frame->width = video_codec_context->width; frame->width = video_codec_context->width;
frame->height = video_codec_context->height; frame->height = video_codec_context->height;
if(!gsr_video_encoder_software_setup_textures(encoder_software, video_codec_context, frame)) { if(!gsr_video_encoder_software_setup_textures(self, video_codec_context, frame)) {
gsr_video_encoder_software_stop(encoder_software, video_codec_context); gsr_video_encoder_software_stop(self, video_codec_context);
return false; return false;
} }
@@ -101,28 +101,28 @@ void gsr_video_encoder_software_stop(gsr_video_encoder_software *self, AVCodecCo
} }
static void gsr_video_encoder_software_copy_textures_to_frame(gsr_video_encoder *encoder, AVFrame *frame) { static void gsr_video_encoder_software_copy_textures_to_frame(gsr_video_encoder *encoder, AVFrame *frame) {
gsr_video_encoder_software *encoder_software = encoder->priv; gsr_video_encoder_software *self = encoder->priv;
// TODO: hdr support // TODO: hdr support
const unsigned int formats[2] = { GL_RED, GL_RG }; const unsigned int formats[2] = { GL_RED, GL_RG };
for(int i = 0; i < 2; ++i) { for(int i = 0; i < 2; ++i) {
encoder_software->params.egl->glBindTexture(GL_TEXTURE_2D, encoder_software->target_textures[i]); self->params.egl->glBindTexture(GL_TEXTURE_2D, self->target_textures[i]);
// We could use glGetTexSubImage and then we wouldn't have to use a specific linesize (LINESIZE_ALIGNMENT) that adds padding, // We could use glGetTexSubImage and then we wouldn't have to use a specific linesize (LINESIZE_ALIGNMENT) that adds padding,
// but glGetTexSubImage is only available starting from opengl 4.5. // but glGetTexSubImage is only available starting from opengl 4.5.
encoder_software->params.egl->glGetTexImage(GL_TEXTURE_2D, 0, formats[i], GL_UNSIGNED_BYTE, frame->data[i]); self->params.egl->glGetTexImage(GL_TEXTURE_2D, 0, formats[i], GL_UNSIGNED_BYTE, frame->data[i]);
} }
encoder_software->params.egl->glBindTexture(GL_TEXTURE_2D, 0); self->params.egl->glBindTexture(GL_TEXTURE_2D, 0);
// cap_kms->kms.base.egl->eglSwapBuffers(cap_kms->kms.base.egl->egl_display, cap_kms->kms.base.egl->egl_surface); // cap_kms->kms.base.egl->eglSwapBuffers(cap_kms->kms.base.egl->egl_display, cap_kms->kms.base.egl->egl_surface);
encoder_software->params.egl->glFlush(); self->params.egl->glFlush();
encoder_software->params.egl->glFinish(); self->params.egl->glFinish();
} }
static void gsr_video_encoder_software_get_textures(gsr_video_encoder *encoder, unsigned int *textures, int *num_textures, gsr_destination_color *destination_color) { static void gsr_video_encoder_software_get_textures(gsr_video_encoder *encoder, unsigned int *textures, int *num_textures, gsr_destination_color *destination_color) {
gsr_video_encoder_software *encoder_software = encoder->priv; gsr_video_encoder_software *self = encoder->priv;
textures[0] = encoder_software->target_textures[0]; textures[0] = self->target_textures[0];
textures[1] = encoder_software->target_textures[1]; textures[1] = self->target_textures[1];
*num_textures = 2; *num_textures = 2;
*destination_color = encoder_software->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? GSR_DESTINATION_COLOR_P010 : GSR_DESTINATION_COLOR_NV12; *destination_color = self->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? GSR_DESTINATION_COLOR_P010 : GSR_DESTINATION_COLOR_NV12;
} }
static void gsr_video_encoder_software_destroy(gsr_video_encoder *encoder, AVCodecContext *video_codec_context) { static void gsr_video_encoder_software_destroy(gsr_video_encoder *encoder, AVCodecContext *video_codec_context) {

View File

@@ -299,12 +299,12 @@ static bool get_supported_video_codecs(VADisplay va_dpy, gsr_supported_video_cod
} }
static gsr_supported_video_codecs gsr_video_encoder_vaapi_get_supported_codecs(gsr_video_encoder *encoder, bool cleanup) { static gsr_supported_video_codecs gsr_video_encoder_vaapi_get_supported_codecs(gsr_video_encoder *encoder, bool cleanup) {
gsr_video_encoder_vaapi *encoder_vaapi = encoder->priv; gsr_video_encoder_vaapi *self = encoder->priv;
gsr_supported_video_codecs supported_video_codecs = {0}; gsr_supported_video_codecs supported_video_codecs = {0};
char render_path[128]; char render_path[128];
if(!gsr_card_path_get_render_path(encoder_vaapi->params.egl->card_path, render_path)) { if(!gsr_card_path_get_render_path(self->params.egl->card_path, render_path)) {
fprintf(stderr, "gsr error: gsr_video_encoder_vaapi_get_supported_codecs: failed to get /dev/dri/renderDXXX file from %s\n", encoder_vaapi->params.egl->card_path); fprintf(stderr, "gsr error: gsr_video_encoder_vaapi_get_supported_codecs: failed to get /dev/dri/renderDXXX file from %s\n", self->params.egl->card_path);
return supported_video_codecs; return supported_video_codecs;
} }
@@ -329,13 +329,13 @@ static gsr_supported_video_codecs gsr_video_encoder_vaapi_get_supported_codecs(g
static void gsr_video_encoder_vaapi_stop(gsr_video_encoder_vaapi *self, AVCodecContext *video_codec_context); static void gsr_video_encoder_vaapi_stop(gsr_video_encoder_vaapi *self, AVCodecContext *video_codec_context);
static bool gsr_video_encoder_vaapi_start(gsr_video_encoder *encoder, AVCodecContext *video_codec_context, AVFrame *frame) { static bool gsr_video_encoder_vaapi_start(gsr_video_encoder *encoder, AVCodecContext *video_codec_context, AVFrame *frame) {
gsr_video_encoder_vaapi *encoder_vaapi = encoder->priv; gsr_video_encoder_vaapi *self = encoder->priv;
if(encoder_vaapi->params.egl->gpu_info.vendor == GSR_GPU_VENDOR_AMD && video_codec_context->codec_id == AV_CODEC_ID_HEVC) { if(self->params.egl->gpu_info.vendor == GSR_GPU_VENDOR_AMD && video_codec_context->codec_id == AV_CODEC_ID_HEVC) {
// TODO: dont do this if using ffmpeg reports that this is not needed (AMD driver bug that was fixed recently) // TODO: dont do this if using ffmpeg reports that this is not needed (AMD driver bug that was fixed recently)
video_codec_context->width = FFALIGN(video_codec_context->width, 64); video_codec_context->width = FFALIGN(video_codec_context->width, 64);
video_codec_context->height = FFALIGN(video_codec_context->height, 16); video_codec_context->height = FFALIGN(video_codec_context->height, 16);
} else if(encoder_vaapi->params.egl->gpu_info.vendor == GSR_GPU_VENDOR_AMD && video_codec_context->codec_id == AV_CODEC_ID_AV1) { } else if(self->params.egl->gpu_info.vendor == GSR_GPU_VENDOR_AMD && video_codec_context->codec_id == AV_CODEC_ID_AV1) {
// TODO: Dont do this for VCN 5 and forward which should fix this hardware bug // TODO: Dont do this for VCN 5 and forward which should fix this hardware bug
video_codec_context->width = FFALIGN(video_codec_context->width, 64); video_codec_context->width = FFALIGN(video_codec_context->width, 64);
// AMD driver has special case handling for 1080 height to set it to 1082 instead of 1088 (1080 aligned to 16). // AMD driver has special case handling for 1080 height to set it to 1082 instead of 1088 (1080 aligned to 16).
@@ -354,13 +354,13 @@ static bool gsr_video_encoder_vaapi_start(gsr_video_encoder *encoder, AVCodecCon
frame->width = video_codec_context->width; frame->width = video_codec_context->width;
frame->height = video_codec_context->height; frame->height = video_codec_context->height;
if(!gsr_video_encoder_vaapi_setup_context(encoder_vaapi, video_codec_context)) { if(!gsr_video_encoder_vaapi_setup_context(self, video_codec_context)) {
gsr_video_encoder_vaapi_stop(encoder_vaapi, video_codec_context); gsr_video_encoder_vaapi_stop(self, video_codec_context);
return false; return false;
} }
if(!gsr_video_encoder_vaapi_setup_textures(encoder_vaapi, video_codec_context, frame)) { if(!gsr_video_encoder_vaapi_setup_textures(self, video_codec_context, frame)) {
gsr_video_encoder_vaapi_stop(encoder_vaapi, video_codec_context); gsr_video_encoder_vaapi_stop(self, video_codec_context);
return false; return false;
} }
@@ -386,11 +386,11 @@ void gsr_video_encoder_vaapi_stop(gsr_video_encoder_vaapi *self, AVCodecContext
} }
static void gsr_video_encoder_vaapi_get_textures(gsr_video_encoder *encoder, unsigned int *textures, int *num_textures, gsr_destination_color *destination_color) { static void gsr_video_encoder_vaapi_get_textures(gsr_video_encoder *encoder, unsigned int *textures, int *num_textures, gsr_destination_color *destination_color) {
gsr_video_encoder_vaapi *encoder_vaapi = encoder->priv; gsr_video_encoder_vaapi *self = encoder->priv;
textures[0] = encoder_vaapi->target_textures[0]; textures[0] = self->target_textures[0];
textures[1] = encoder_vaapi->target_textures[1]; textures[1] = self->target_textures[1];
*num_textures = 2; *num_textures = 2;
*destination_color = encoder_vaapi->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? GSR_DESTINATION_COLOR_P010 : GSR_DESTINATION_COLOR_NV12; *destination_color = self->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? GSR_DESTINATION_COLOR_P010 : GSR_DESTINATION_COLOR_NV12;
} }
static void gsr_video_encoder_vaapi_destroy(gsr_video_encoder *encoder, AVCodecContext *video_codec_context) { static void gsr_video_encoder_vaapi_destroy(gsr_video_encoder *encoder, AVCodecContext *video_codec_context) {

View File

@@ -86,11 +86,11 @@ void gsr_video_encoder_vulkan_stop(gsr_video_encoder_vulkan *self, AVCodecContex
} }
static void gsr_video_encoder_vulkan_get_textures(gsr_video_encoder *encoder, unsigned int *textures, int *num_textures, gsr_destination_color *destination_color) { static void gsr_video_encoder_vulkan_get_textures(gsr_video_encoder *encoder, unsigned int *textures, int *num_textures, gsr_destination_color *destination_color) {
gsr_video_encoder_vulkan *encoder_vaapi = encoder->priv; gsr_video_encoder_vulkan *self = encoder->priv;
textures[0] = encoder_vaapi->target_textures[0]; textures[0] = self->target_textures[0];
textures[1] = encoder_vaapi->target_textures[1]; textures[1] = self->target_textures[1];
*num_textures = 2; *num_textures = 2;
*destination_color = encoder_vaapi->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? GSR_DESTINATION_COLOR_P010 : GSR_DESTINATION_COLOR_NV12; *destination_color = self->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? GSR_DESTINATION_COLOR_P010 : GSR_DESTINATION_COLOR_NV12;
} }
static void gsr_video_encoder_vulkan_destroy(gsr_video_encoder *encoder, AVCodecContext *video_codec_context) { static void gsr_video_encoder_vulkan_destroy(gsr_video_encoder *encoder, AVCodecContext *video_codec_context) {
@@ -104,20 +104,20 @@ gsr_video_encoder* gsr_video_encoder_vulkan_create(const gsr_video_encoder_vulka
if(!encoder) if(!encoder)
return NULL; return NULL;
gsr_video_encoder_vulkan *encoder_vaapi = calloc(1, sizeof(gsr_video_encoder_vulkan)); gsr_video_encoder_vulkan *encoder_vulkan = calloc(1, sizeof(gsr_video_encoder_vulkan));
if(!encoder_vaapi) { if(!encoder_vulkan) {
free(encoder); free(encoder);
return NULL; return NULL;
} }
encoder_vaapi->params = *params; encoder_vulkan->params = *params;
*encoder = (gsr_video_encoder) { *encoder = (gsr_video_encoder) {
.start = gsr_video_encoder_vulkan_start, .start = gsr_video_encoder_vulkan_start,
.copy_textures_to_frame = NULL, .copy_textures_to_frame = NULL,
.get_textures = gsr_video_encoder_vulkan_get_textures, .get_textures = gsr_video_encoder_vulkan_get_textures,
.destroy = gsr_video_encoder_vulkan_destroy, .destroy = gsr_video_encoder_vulkan_destroy,
.priv = encoder_vaapi .priv = encoder_vulkan
}; };
return encoder; return encoder;