mirror of
https://repo.dec05eba.com/gpu-screen-recorder
synced 2026-03-31 09:07:13 +09:00
Refactor xcomposite
Move common code between cuda/vaapi implementations into xcomposite file. This also fixes limited/full range colors on nvidia.
This commit is contained in:
3
build.sh
3
build.sh
@@ -25,6 +25,7 @@ build_gsr() {
|
|||||||
libs="$(pkg-config --libs $dependencies) -ldl -pthread -lm"
|
libs="$(pkg-config --libs $dependencies) -ldl -pthread -lm"
|
||||||
$CC -c src/capture/capture.c $opts $includes
|
$CC -c src/capture/capture.c $opts $includes
|
||||||
$CC -c src/capture/nvfbc.c $opts $includes
|
$CC -c src/capture/nvfbc.c $opts $includes
|
||||||
|
$CC -c src/capture/xcomposite.c $opts $includes
|
||||||
$CC -c src/capture/xcomposite_cuda.c $opts $includes
|
$CC -c src/capture/xcomposite_cuda.c $opts $includes
|
||||||
$CC -c src/capture/xcomposite_vaapi.c $opts $includes
|
$CC -c src/capture/xcomposite_vaapi.c $opts $includes
|
||||||
$CC -c src/capture/kms_vaapi.c $opts $includes
|
$CC -c src/capture/kms_vaapi.c $opts $includes
|
||||||
@@ -43,7 +44,7 @@ build_gsr() {
|
|||||||
$CXX -c src/sound.cpp $opts $includes
|
$CXX -c src/sound.cpp $opts $includes
|
||||||
$CXX -c src/main.cpp $opts $includes
|
$CXX -c src/main.cpp $opts $includes
|
||||||
$CXX -o gpu-screen-recorder capture.o nvfbc.o kms_client.o egl.o cuda.o xnvctrl.o overclock.o window_texture.o shader.o \
|
$CXX -o gpu-screen-recorder capture.o nvfbc.o kms_client.o egl.o cuda.o xnvctrl.o overclock.o window_texture.o shader.o \
|
||||||
color_conversion.o utils.o library_loader.o xcomposite_cuda.o xcomposite_vaapi.o kms_vaapi.o kms_cuda.o kms.o sound.o main.o $libs $opts
|
color_conversion.o utils.o library_loader.o xcomposite.o xcomposite_cuda.o xcomposite_vaapi.o kms_vaapi.o kms_cuda.o kms.o sound.o main.o $libs $opts
|
||||||
}
|
}
|
||||||
|
|
||||||
build_gsr_kms_server
|
build_gsr_kms_server
|
||||||
|
|||||||
@@ -12,6 +12,8 @@ typedef struct gsr_cuda gsr_cuda;
|
|||||||
typedef struct AVFrame AVFrame;
|
typedef struct AVFrame AVFrame;
|
||||||
typedef struct CUgraphicsResource_st *CUgraphicsResource;
|
typedef struct CUgraphicsResource_st *CUgraphicsResource;
|
||||||
typedef struct CUarray_st *CUarray;
|
typedef struct CUarray_st *CUarray;
|
||||||
|
typedef struct CUctx_st *CUcontext;
|
||||||
|
typedef struct CUstream_st *CUstream;
|
||||||
|
|
||||||
typedef struct gsr_capture gsr_capture;
|
typedef struct gsr_capture gsr_capture;
|
||||||
|
|
||||||
@@ -58,4 +60,7 @@ bool gsr_capture_base_setup_vaapi_textures(gsr_capture_base *self, AVFrame *fram
|
|||||||
bool gsr_capture_base_setup_cuda_textures(gsr_capture_base *base, AVFrame *frame, gsr_cuda_context *cuda_context, gsr_egl *egl, gsr_color_range color_range, gsr_source_color source_color, bool hdr);
|
bool gsr_capture_base_setup_cuda_textures(gsr_capture_base *base, AVFrame *frame, gsr_cuda_context *cuda_context, gsr_egl *egl, gsr_color_range color_range, gsr_source_color source_color, bool hdr);
|
||||||
void gsr_capture_base_stop(gsr_capture_base *self, gsr_egl *egl);
|
void gsr_capture_base_stop(gsr_capture_base *self, gsr_egl *egl);
|
||||||
|
|
||||||
|
bool drm_create_codec_context(const char *card_path, AVCodecContext *video_codec_context, bool hdr, VADisplay *va_dpy);
|
||||||
|
bool cuda_create_codec_context(CUcontext cu_ctx, AVCodecContext *video_codec_context, CUstream *cuda_stream);
|
||||||
|
|
||||||
#endif /* GSR_CAPTURE_CAPTURE_H */
|
#endif /* GSR_CAPTURE_CAPTURE_H */
|
||||||
|
|||||||
46
include/capture/xcomposite.h
Normal file
46
include/capture/xcomposite.h
Normal file
@@ -0,0 +1,46 @@
|
|||||||
|
#ifndef GSR_CAPTURE_XCOMPOSITE_H
|
||||||
|
#define GSR_CAPTURE_XCOMPOSITE_H
|
||||||
|
|
||||||
|
#include "capture.h"
|
||||||
|
#include "../egl.h"
|
||||||
|
#include "../vec2.h"
|
||||||
|
#include "../color_conversion.h"
|
||||||
|
#include "../window_texture.h"
|
||||||
|
|
||||||
|
typedef struct {
|
||||||
|
gsr_egl *egl;
|
||||||
|
Window window;
|
||||||
|
bool follow_focused; /* If this is set then |window| is ignored */
|
||||||
|
vec2i region_size; /* This is currently only used with |follow_focused| */
|
||||||
|
gsr_color_range color_range;
|
||||||
|
} gsr_capture_xcomposite_params;
|
||||||
|
|
||||||
|
typedef struct {
|
||||||
|
gsr_capture_base base;
|
||||||
|
gsr_capture_xcomposite_params params;
|
||||||
|
XEvent xev;
|
||||||
|
|
||||||
|
bool should_stop;
|
||||||
|
bool stop_is_error;
|
||||||
|
bool window_resized;
|
||||||
|
bool follow_focused_initialized;
|
||||||
|
|
||||||
|
Window window;
|
||||||
|
vec2i window_size;
|
||||||
|
vec2i texture_size;
|
||||||
|
double window_resize_timer;
|
||||||
|
|
||||||
|
WindowTexture window_texture;
|
||||||
|
|
||||||
|
Atom net_active_window_atom;
|
||||||
|
} gsr_capture_xcomposite;
|
||||||
|
|
||||||
|
void gsr_capture_xcomposite_init(gsr_capture_xcomposite *self, const gsr_capture_xcomposite_params *params);
|
||||||
|
|
||||||
|
int gsr_capture_xcomposite_start(gsr_capture_xcomposite *self, AVCodecContext *video_codec_context, AVFrame *frame);
|
||||||
|
void gsr_capture_xcomposite_stop(gsr_capture_xcomposite *self, AVCodecContext *video_codec_context);
|
||||||
|
void gsr_capture_xcomposite_tick(gsr_capture_xcomposite *self, AVCodecContext *video_codec_context);
|
||||||
|
bool gsr_capture_xcomposite_should_stop(gsr_capture_xcomposite *self, bool *err);
|
||||||
|
int gsr_capture_xcomposite_capture(gsr_capture_xcomposite *self, AVFrame *frame);
|
||||||
|
|
||||||
|
#endif /* GSR_CAPTURE_XCOMPOSITE_H */
|
||||||
@@ -2,14 +2,10 @@
|
|||||||
#define GSR_CAPTURE_XCOMPOSITE_CUDA_H
|
#define GSR_CAPTURE_XCOMPOSITE_CUDA_H
|
||||||
|
|
||||||
#include "capture.h"
|
#include "capture.h"
|
||||||
#include "../egl.h"
|
#include "xcomposite.h"
|
||||||
#include "../vec2.h"
|
|
||||||
|
|
||||||
typedef struct {
|
typedef struct {
|
||||||
gsr_egl *egl;
|
gsr_capture_xcomposite_params base;
|
||||||
Window window;
|
|
||||||
bool follow_focused; /* If this is set then |window| is ignored */
|
|
||||||
vec2i region_size; /* This is currently only used with |follow_focused| */
|
|
||||||
bool overclock;
|
bool overclock;
|
||||||
} gsr_capture_xcomposite_cuda_params;
|
} gsr_capture_xcomposite_cuda_params;
|
||||||
|
|
||||||
|
|||||||
@@ -2,16 +2,10 @@
|
|||||||
#define GSR_CAPTURE_XCOMPOSITE_VAAPI_H
|
#define GSR_CAPTURE_XCOMPOSITE_VAAPI_H
|
||||||
|
|
||||||
#include "capture.h"
|
#include "capture.h"
|
||||||
#include "../egl.h"
|
#include "xcomposite.h"
|
||||||
#include "../vec2.h"
|
|
||||||
#include "../color_conversion.h"
|
|
||||||
|
|
||||||
typedef struct {
|
typedef struct {
|
||||||
gsr_egl *egl;
|
gsr_capture_xcomposite_params base;
|
||||||
Window window;
|
|
||||||
bool follow_focused; /* If this is set then |window| is ignored */
|
|
||||||
vec2i region_size; /* This is currently only used with |follow_focused| */
|
|
||||||
gsr_color_range color_range;
|
|
||||||
} gsr_capture_xcomposite_vaapi_params;
|
} gsr_capture_xcomposite_vaapi_params;
|
||||||
|
|
||||||
gsr_capture* gsr_capture_xcomposite_vaapi_create(const gsr_capture_xcomposite_vaapi_params *params);
|
gsr_capture* gsr_capture_xcomposite_vaapi_create(const gsr_capture_xcomposite_vaapi_params *params);
|
||||||
|
|||||||
@@ -212,7 +212,7 @@ int gsr_kms_client_init(gsr_kms_client *self, const char *card_path) {
|
|||||||
//fprintf(stderr, "No permission:(\n");
|
//fprintf(stderr, "No permission:(\n");
|
||||||
}
|
}
|
||||||
cap_free(kms_server_cap);
|
cap_free(kms_server_cap);
|
||||||
} else {
|
} else if(!inside_flatpak) {
|
||||||
if(errno == ENODATA)
|
if(errno == ENODATA)
|
||||||
fprintf(stderr, "gsr info: gsr_kms_client_init: gsr-kms-server is missing sys_admin cap and will require root authentication. To bypass this automatically, run: sudo setcap cap_sys_admin+ep '%s'\n", server_filepath);
|
fprintf(stderr, "gsr info: gsr_kms_client_init: gsr-kms-server is missing sys_admin cap and will require root authentication. To bypass this automatically, run: sudo setcap cap_sys_admin+ep '%s'\n", server_filepath);
|
||||||
else
|
else
|
||||||
|
|||||||
@@ -1,11 +1,14 @@
|
|||||||
#include "../../include/capture/capture.h"
|
#include "../../include/capture/capture.h"
|
||||||
#include "../../include/egl.h"
|
#include "../../include/egl.h"
|
||||||
#include "../../include/cuda.h"
|
#include "../../include/cuda.h"
|
||||||
|
#include "../../include/utils.h"
|
||||||
#include <stdio.h>
|
#include <stdio.h>
|
||||||
#include <stdint.h>
|
#include <stdint.h>
|
||||||
#include <va/va.h>
|
#include <va/va.h>
|
||||||
#include <va/va_drmcommon.h>
|
#include <va/va_drmcommon.h>
|
||||||
#include <libavutil/frame.h>
|
#include <libavutil/frame.h>
|
||||||
|
#include <libavutil/hwcontext_vaapi.h>
|
||||||
|
#include <libavutil/hwcontext_cuda.h>
|
||||||
#include <libavcodec/avcodec.h>
|
#include <libavcodec/avcodec.h>
|
||||||
|
|
||||||
#define FOURCC_NV12 842094158
|
#define FOURCC_NV12 842094158
|
||||||
@@ -301,3 +304,95 @@ void gsr_capture_base_stop(gsr_capture_base *self, gsr_egl *egl) {
|
|||||||
self->target_textures[1] = 0;
|
self->target_textures[1] = 0;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
bool drm_create_codec_context(const char *card_path, AVCodecContext *video_codec_context, bool hdr, VADisplay *va_dpy) {
|
||||||
|
char render_path[128];
|
||||||
|
if(!gsr_card_path_get_render_path(card_path, render_path)) {
|
||||||
|
fprintf(stderr, "gsr error: failed to get /dev/dri/renderDXXX file from %s\n", card_path);
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
AVBufferRef *device_ctx;
|
||||||
|
if(av_hwdevice_ctx_create(&device_ctx, AV_HWDEVICE_TYPE_VAAPI, render_path, NULL, 0) < 0) {
|
||||||
|
fprintf(stderr, "Error: Failed to create hardware device context\n");
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
AVBufferRef *frame_context = av_hwframe_ctx_alloc(device_ctx);
|
||||||
|
if(!frame_context) {
|
||||||
|
fprintf(stderr, "Error: Failed to create hwframe context\n");
|
||||||
|
av_buffer_unref(&device_ctx);
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
AVHWFramesContext *hw_frame_context =
|
||||||
|
(AVHWFramesContext *)frame_context->data;
|
||||||
|
hw_frame_context->width = video_codec_context->width;
|
||||||
|
hw_frame_context->height = video_codec_context->height;
|
||||||
|
hw_frame_context->sw_format = hdr ? AV_PIX_FMT_P010LE : AV_PIX_FMT_NV12;
|
||||||
|
hw_frame_context->format = video_codec_context->pix_fmt;
|
||||||
|
hw_frame_context->device_ref = device_ctx;
|
||||||
|
hw_frame_context->device_ctx = (AVHWDeviceContext*)device_ctx->data;
|
||||||
|
|
||||||
|
//hw_frame_context->initial_pool_size = 20;
|
||||||
|
|
||||||
|
AVVAAPIDeviceContext *vactx =((AVHWDeviceContext*)device_ctx->data)->hwctx;
|
||||||
|
*va_dpy = vactx->display;
|
||||||
|
|
||||||
|
if (av_hwframe_ctx_init(frame_context) < 0) {
|
||||||
|
fprintf(stderr, "Error: Failed to initialize hardware frame context "
|
||||||
|
"(note: ffmpeg version needs to be > 4.0)\n");
|
||||||
|
av_buffer_unref(&device_ctx);
|
||||||
|
//av_buffer_unref(&frame_context);
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
video_codec_context->hw_device_ctx = av_buffer_ref(device_ctx);
|
||||||
|
video_codec_context->hw_frames_ctx = av_buffer_ref(frame_context);
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
bool cuda_create_codec_context(CUcontext cu_ctx, AVCodecContext *video_codec_context, CUstream *cuda_stream) {
|
||||||
|
AVBufferRef *device_ctx = av_hwdevice_ctx_alloc(AV_HWDEVICE_TYPE_CUDA);
|
||||||
|
if(!device_ctx) {
|
||||||
|
fprintf(stderr, "gsr error: cuda_create_codec_context failed: failed to create hardware device context\n");
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
AVHWDeviceContext *hw_device_context = (AVHWDeviceContext*)device_ctx->data;
|
||||||
|
AVCUDADeviceContext *cuda_device_context = (AVCUDADeviceContext*)hw_device_context->hwctx;
|
||||||
|
cuda_device_context->cuda_ctx = cu_ctx;
|
||||||
|
if(av_hwdevice_ctx_init(device_ctx) < 0) {
|
||||||
|
fprintf(stderr, "gsr error: cuda_create_codec_context failed: failed to create hardware device context\n");
|
||||||
|
av_buffer_unref(&device_ctx);
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
AVBufferRef *frame_context = av_hwframe_ctx_alloc(device_ctx);
|
||||||
|
if(!frame_context) {
|
||||||
|
fprintf(stderr, "gsr error: cuda_create_codec_context failed: failed to create hwframe context\n");
|
||||||
|
av_buffer_unref(&device_ctx);
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
AVHWFramesContext *hw_frame_context = (AVHWFramesContext*)frame_context->data;
|
||||||
|
hw_frame_context->width = video_codec_context->width;
|
||||||
|
hw_frame_context->height = video_codec_context->height;
|
||||||
|
hw_frame_context->sw_format = AV_PIX_FMT_NV12;
|
||||||
|
hw_frame_context->format = video_codec_context->pix_fmt;
|
||||||
|
hw_frame_context->device_ref = device_ctx;
|
||||||
|
hw_frame_context->device_ctx = (AVHWDeviceContext*)device_ctx->data;
|
||||||
|
|
||||||
|
if (av_hwframe_ctx_init(frame_context) < 0) {
|
||||||
|
fprintf(stderr, "gsr error: cuda_create_codec_context failed: failed to initialize hardware frame context "
|
||||||
|
"(note: ffmpeg version needs to be > 4.0)\n");
|
||||||
|
av_buffer_unref(&device_ctx);
|
||||||
|
//av_buffer_unref(&frame_context);
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
*cuda_stream = cuda_device_context->stream;
|
||||||
|
video_codec_context->hw_device_ctx = av_buffer_ref(device_ctx);
|
||||||
|
video_codec_context->hw_frames_ctx = av_buffer_ref(frame_context);
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|||||||
@@ -23,59 +23,6 @@ typedef struct {
|
|||||||
|
|
||||||
static void gsr_capture_kms_cuda_stop(gsr_capture *cap, AVCodecContext *video_codec_context);
|
static void gsr_capture_kms_cuda_stop(gsr_capture *cap, AVCodecContext *video_codec_context);
|
||||||
|
|
||||||
static bool cuda_create_codec_context(gsr_capture_kms_cuda *cap_kms, AVCodecContext *video_codec_context) {
|
|
||||||
CUcontext old_ctx;
|
|
||||||
cap_kms->cuda.cuCtxPushCurrent_v2(cap_kms->cuda.cu_ctx);
|
|
||||||
|
|
||||||
AVBufferRef *device_ctx = av_hwdevice_ctx_alloc(AV_HWDEVICE_TYPE_CUDA);
|
|
||||||
if(!device_ctx) {
|
|
||||||
fprintf(stderr, "Error: Failed to create hardware device context\n");
|
|
||||||
cap_kms->cuda.cuCtxPopCurrent_v2(&old_ctx);
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
AVHWDeviceContext *hw_device_context = (AVHWDeviceContext*)device_ctx->data;
|
|
||||||
AVCUDADeviceContext *cuda_device_context = (AVCUDADeviceContext*)hw_device_context->hwctx;
|
|
||||||
cuda_device_context->cuda_ctx = cap_kms->cuda.cu_ctx;
|
|
||||||
if(av_hwdevice_ctx_init(device_ctx) < 0) {
|
|
||||||
fprintf(stderr, "Error: Failed to create hardware device context\n");
|
|
||||||
av_buffer_unref(&device_ctx);
|
|
||||||
cap_kms->cuda.cuCtxPopCurrent_v2(&old_ctx);
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
AVBufferRef *frame_context = av_hwframe_ctx_alloc(device_ctx);
|
|
||||||
if(!frame_context) {
|
|
||||||
fprintf(stderr, "Error: Failed to create hwframe context\n");
|
|
||||||
av_buffer_unref(&device_ctx);
|
|
||||||
cap_kms->cuda.cuCtxPopCurrent_v2(&old_ctx);
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
AVHWFramesContext *hw_frame_context =
|
|
||||||
(AVHWFramesContext *)frame_context->data;
|
|
||||||
hw_frame_context->width = video_codec_context->width;
|
|
||||||
hw_frame_context->height = video_codec_context->height;
|
|
||||||
hw_frame_context->sw_format = cap_kms->params.hdr ? AV_PIX_FMT_P010LE : AV_PIX_FMT_NV12;
|
|
||||||
hw_frame_context->format = video_codec_context->pix_fmt;
|
|
||||||
hw_frame_context->device_ref = device_ctx;
|
|
||||||
hw_frame_context->device_ctx = (AVHWDeviceContext*)device_ctx->data;
|
|
||||||
|
|
||||||
if (av_hwframe_ctx_init(frame_context) < 0) {
|
|
||||||
fprintf(stderr, "Error: Failed to initialize hardware frame context "
|
|
||||||
"(note: ffmpeg version needs to be > 4.0)\n");
|
|
||||||
av_buffer_unref(&device_ctx);
|
|
||||||
//av_buffer_unref(&frame_context);
|
|
||||||
cap_kms->cuda.cuCtxPopCurrent_v2(&old_ctx);
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
cap_kms->cuda_stream = cuda_device_context->stream;
|
|
||||||
video_codec_context->hw_device_ctx = av_buffer_ref(device_ctx);
|
|
||||||
video_codec_context->hw_frames_ctx = av_buffer_ref(frame_context);
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
static int gsr_capture_kms_cuda_start(gsr_capture *cap, AVCodecContext *video_codec_context, AVFrame *frame) {
|
static int gsr_capture_kms_cuda_start(gsr_capture *cap, AVCodecContext *video_codec_context, AVFrame *frame) {
|
||||||
gsr_capture_kms_cuda *cap_kms = cap->priv;
|
gsr_capture_kms_cuda *cap_kms = cap->priv;
|
||||||
|
|
||||||
@@ -92,7 +39,7 @@ static int gsr_capture_kms_cuda_start(gsr_capture *cap, AVCodecContext *video_co
|
|||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
if(!cuda_create_codec_context(cap_kms, video_codec_context)) {
|
if(!cuda_create_codec_context(cap_kms->cuda.cu_ctx, video_codec_context, &cap_kms->cuda_stream)) {
|
||||||
gsr_capture_kms_cuda_stop(cap, video_codec_context);
|
gsr_capture_kms_cuda_stop(cap, video_codec_context);
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
@@ -102,6 +49,7 @@ static int gsr_capture_kms_cuda_start(gsr_capture *cap, AVCodecContext *video_co
|
|||||||
.cuda_graphics_resources = cap_kms->cuda_graphics_resources,
|
.cuda_graphics_resources = cap_kms->cuda_graphics_resources,
|
||||||
.mapped_arrays = cap_kms->mapped_arrays
|
.mapped_arrays = cap_kms->mapped_arrays
|
||||||
};
|
};
|
||||||
|
|
||||||
if(!gsr_capture_base_setup_cuda_textures(&cap_kms->base, frame, &cuda_context, cap_kms->params.egl, cap_kms->params.color_range, GSR_SOURCE_COLOR_RGB, cap_kms->params.hdr)) {
|
if(!gsr_capture_base_setup_cuda_textures(&cap_kms->base, frame, &cuda_context, cap_kms->params.egl, cap_kms->params.color_range, GSR_SOURCE_COLOR_RGB, cap_kms->params.hdr)) {
|
||||||
gsr_capture_kms_cuda_stop(cap, video_codec_context);
|
gsr_capture_kms_cuda_stop(cap, video_codec_context);
|
||||||
return -1;
|
return -1;
|
||||||
|
|||||||
@@ -21,53 +21,6 @@ typedef struct {
|
|||||||
|
|
||||||
static void gsr_capture_kms_vaapi_stop(gsr_capture *cap, AVCodecContext *video_codec_context);
|
static void gsr_capture_kms_vaapi_stop(gsr_capture *cap, AVCodecContext *video_codec_context);
|
||||||
|
|
||||||
static bool drm_create_codec_context(gsr_capture_kms_vaapi *cap_kms, AVCodecContext *video_codec_context) {
|
|
||||||
char render_path[128];
|
|
||||||
if(!gsr_card_path_get_render_path(cap_kms->params.egl->card_path, render_path)) {
|
|
||||||
fprintf(stderr, "gsr error: failed to get /dev/dri/renderDXXX file from %s\n", cap_kms->params.egl->card_path);
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
AVBufferRef *device_ctx;
|
|
||||||
if(av_hwdevice_ctx_create(&device_ctx, AV_HWDEVICE_TYPE_VAAPI, render_path, NULL, 0) < 0) {
|
|
||||||
fprintf(stderr, "Error: Failed to create hardware device context\n");
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
AVBufferRef *frame_context = av_hwframe_ctx_alloc(device_ctx);
|
|
||||||
if(!frame_context) {
|
|
||||||
fprintf(stderr, "Error: Failed to create hwframe context\n");
|
|
||||||
av_buffer_unref(&device_ctx);
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
AVHWFramesContext *hw_frame_context =
|
|
||||||
(AVHWFramesContext *)frame_context->data;
|
|
||||||
hw_frame_context->width = video_codec_context->width;
|
|
||||||
hw_frame_context->height = video_codec_context->height;
|
|
||||||
hw_frame_context->sw_format = cap_kms->params.hdr ? AV_PIX_FMT_P010LE : AV_PIX_FMT_NV12;
|
|
||||||
hw_frame_context->format = video_codec_context->pix_fmt;
|
|
||||||
hw_frame_context->device_ref = device_ctx;
|
|
||||||
hw_frame_context->device_ctx = (AVHWDeviceContext*)device_ctx->data;
|
|
||||||
|
|
||||||
//hw_frame_context->initial_pool_size = 20;
|
|
||||||
|
|
||||||
AVVAAPIDeviceContext *vactx =((AVHWDeviceContext*)device_ctx->data)->hwctx;
|
|
||||||
cap_kms->va_dpy = vactx->display;
|
|
||||||
|
|
||||||
if (av_hwframe_ctx_init(frame_context) < 0) {
|
|
||||||
fprintf(stderr, "Error: Failed to initialize hardware frame context "
|
|
||||||
"(note: ffmpeg version needs to be > 4.0)\n");
|
|
||||||
av_buffer_unref(&device_ctx);
|
|
||||||
//av_buffer_unref(&frame_context);
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
video_codec_context->hw_device_ctx = av_buffer_ref(device_ctx);
|
|
||||||
video_codec_context->hw_frames_ctx = av_buffer_ref(frame_context);
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
static int gsr_capture_kms_vaapi_start(gsr_capture *cap, AVCodecContext *video_codec_context, AVFrame *frame) {
|
static int gsr_capture_kms_vaapi_start(gsr_capture *cap, AVCodecContext *video_codec_context, AVFrame *frame) {
|
||||||
gsr_capture_kms_vaapi *cap_kms = cap->priv;
|
gsr_capture_kms_vaapi *cap_kms = cap->priv;
|
||||||
|
|
||||||
@@ -77,7 +30,7 @@ static int gsr_capture_kms_vaapi_start(gsr_capture *cap, AVCodecContext *video_c
|
|||||||
return res;
|
return res;
|
||||||
}
|
}
|
||||||
|
|
||||||
if(!drm_create_codec_context(cap_kms, video_codec_context)) {
|
if(!drm_create_codec_context(cap_kms->params.egl->card_path, video_codec_context, cap_kms->params.hdr, &cap_kms->va_dpy)) {
|
||||||
gsr_capture_kms_vaapi_stop(cap, video_codec_context);
|
gsr_capture_kms_vaapi_stop(cap, video_codec_context);
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -135,51 +135,6 @@ static bool gsr_capture_nvfbc_load_library(gsr_capture *cap) {
|
|||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
static bool ffmpeg_create_cuda_contexts(gsr_capture_nvfbc *cap_nvfbc, AVCodecContext *video_codec_context) {
|
|
||||||
AVBufferRef *device_ctx = av_hwdevice_ctx_alloc(AV_HWDEVICE_TYPE_CUDA);
|
|
||||||
if(!device_ctx) {
|
|
||||||
fprintf(stderr, "gsr error: cuda_create_codec_context failed: failed to create hardware device context\n");
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
AVHWDeviceContext *hw_device_context = (AVHWDeviceContext*)device_ctx->data;
|
|
||||||
AVCUDADeviceContext *cuda_device_context = (AVCUDADeviceContext*)hw_device_context->hwctx;
|
|
||||||
cuda_device_context->cuda_ctx = cap_nvfbc->cuda.cu_ctx;
|
|
||||||
if(av_hwdevice_ctx_init(device_ctx) < 0) {
|
|
||||||
fprintf(stderr, "gsr error: cuda_create_codec_context failed: failed to create hardware device context\n");
|
|
||||||
av_buffer_unref(&device_ctx);
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
AVBufferRef *frame_context = av_hwframe_ctx_alloc(device_ctx);
|
|
||||||
if(!frame_context) {
|
|
||||||
fprintf(stderr, "gsr error: cuda_create_codec_context failed: failed to create hwframe context\n");
|
|
||||||
av_buffer_unref(&device_ctx);
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
AVHWFramesContext *hw_frame_context = (AVHWFramesContext*)frame_context->data;
|
|
||||||
hw_frame_context->width = video_codec_context->width;
|
|
||||||
hw_frame_context->height = video_codec_context->height;
|
|
||||||
hw_frame_context->sw_format = AV_PIX_FMT_NV12;
|
|
||||||
hw_frame_context->format = video_codec_context->pix_fmt;
|
|
||||||
hw_frame_context->device_ref = device_ctx;
|
|
||||||
hw_frame_context->device_ctx = (AVHWDeviceContext*)device_ctx->data;
|
|
||||||
|
|
||||||
if (av_hwframe_ctx_init(frame_context) < 0) {
|
|
||||||
fprintf(stderr, "gsr error: cuda_create_codec_context failed: failed to initialize hardware frame context "
|
|
||||||
"(note: ffmpeg version needs to be > 4.0)\n");
|
|
||||||
av_buffer_unref(&device_ctx);
|
|
||||||
//av_buffer_unref(&frame_context);
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
cap_nvfbc->cuda_stream = cuda_device_context->stream;
|
|
||||||
video_codec_context->hw_device_ctx = av_buffer_ref(device_ctx);
|
|
||||||
video_codec_context->hw_frames_ctx = av_buffer_ref(frame_context);
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
/* TODO: check for glx swap control extension string (GLX_EXT_swap_control, etc) */
|
/* TODO: check for glx swap control extension string (GLX_EXT_swap_control, etc) */
|
||||||
static void set_vertical_sync_enabled(gsr_egl *egl, int enabled) {
|
static void set_vertical_sync_enabled(gsr_egl *egl, int enabled) {
|
||||||
int result = 0;
|
int result = 0;
|
||||||
@@ -352,7 +307,7 @@ static int gsr_capture_nvfbc_start(gsr_capture *cap, AVCodecContext *video_codec
|
|||||||
frame->width = video_codec_context->width;
|
frame->width = video_codec_context->width;
|
||||||
frame->height = video_codec_context->height;
|
frame->height = video_codec_context->height;
|
||||||
|
|
||||||
if(!ffmpeg_create_cuda_contexts(cap_nvfbc, video_codec_context))
|
if(!cuda_create_codec_context(cap_nvfbc->cuda.cu_ctx, video_codec_context, &cap_nvfbc->cuda_stream))
|
||||||
goto error_cleanup;
|
goto error_cleanup;
|
||||||
|
|
||||||
gsr_cuda_context cuda_context = {
|
gsr_cuda_context cuda_context = {
|
||||||
@@ -360,6 +315,7 @@ static int gsr_capture_nvfbc_start(gsr_capture *cap, AVCodecContext *video_codec
|
|||||||
.cuda_graphics_resources = cap_nvfbc->cuda_graphics_resources,
|
.cuda_graphics_resources = cap_nvfbc->cuda_graphics_resources,
|
||||||
.mapped_arrays = cap_nvfbc->mapped_arrays
|
.mapped_arrays = cap_nvfbc->mapped_arrays
|
||||||
};
|
};
|
||||||
|
|
||||||
// TODO: Remove this, it creates shit we dont need
|
// TODO: Remove this, it creates shit we dont need
|
||||||
if(!gsr_capture_base_setup_cuda_textures(&cap_nvfbc->base, frame, &cuda_context, cap_nvfbc->params.egl, cap_nvfbc->params.color_range, GSR_SOURCE_COLOR_BGR, cap_nvfbc->params.hdr)) {
|
if(!gsr_capture_base_setup_cuda_textures(&cap_nvfbc->base, frame, &cuda_context, cap_nvfbc->params.egl, cap_nvfbc->params.color_range, GSR_SOURCE_COLOR_BGR, cap_nvfbc->params.hdr)) {
|
||||||
goto error_cleanup;
|
goto error_cleanup;
|
||||||
|
|||||||
265
src/capture/xcomposite.c
Normal file
265
src/capture/xcomposite.c
Normal file
@@ -0,0 +1,265 @@
|
|||||||
|
#include "../../include/capture/xcomposite.h"
|
||||||
|
#include "../../include/window_texture.h"
|
||||||
|
#include "../../include/utils.h"
|
||||||
|
#include <stdlib.h>
|
||||||
|
#include <stdio.h>
|
||||||
|
#include <unistd.h>
|
||||||
|
#include <assert.h>
|
||||||
|
#include <X11/Xlib.h>
|
||||||
|
#include <libavutil/hwcontext.h>
|
||||||
|
#include <libavutil/hwcontext.h>
|
||||||
|
#include <libavutil/frame.h>
|
||||||
|
#include <libavcodec/avcodec.h>
|
||||||
|
#include <va/va.h>
|
||||||
|
#include <va/va_drmcommon.h>
|
||||||
|
|
||||||
|
static int max_int(int a, int b) {
|
||||||
|
return a > b ? a : b;
|
||||||
|
}
|
||||||
|
|
||||||
|
static int min_int(int a, int b) {
|
||||||
|
return a < b ? a : b;
|
||||||
|
}
|
||||||
|
|
||||||
|
void gsr_capture_xcomposite_init(gsr_capture_xcomposite *self, const gsr_capture_xcomposite_params *params) {
|
||||||
|
memset(self, 0, sizeof(*self));
|
||||||
|
self->params = *params;
|
||||||
|
}
|
||||||
|
|
||||||
|
static Window get_focused_window(Display *display, Atom net_active_window_atom) {
|
||||||
|
Atom type;
|
||||||
|
int format = 0;
|
||||||
|
unsigned long num_items = 0;
|
||||||
|
unsigned long bytes_after = 0;
|
||||||
|
unsigned char *properties = NULL;
|
||||||
|
if(XGetWindowProperty(display, DefaultRootWindow(display), net_active_window_atom, 0, 1024, False, AnyPropertyType, &type, &format, &num_items, &bytes_after, &properties) == Success && properties) {
|
||||||
|
Window focused_window = *(unsigned long*)properties;
|
||||||
|
XFree(properties);
|
||||||
|
return focused_window;
|
||||||
|
}
|
||||||
|
return None;
|
||||||
|
}
|
||||||
|
|
||||||
|
int gsr_capture_xcomposite_start(gsr_capture_xcomposite *self, AVCodecContext *video_codec_context, AVFrame *frame) {
|
||||||
|
self->base.video_codec_context = video_codec_context;
|
||||||
|
|
||||||
|
if(self->params.follow_focused) {
|
||||||
|
self->net_active_window_atom = XInternAtom(self->params.egl->x11.dpy, "_NET_ACTIVE_WINDOW", False);
|
||||||
|
if(!self->net_active_window_atom) {
|
||||||
|
fprintf(stderr, "gsr error: gsr_capture_xcomposite_start failed: failed to get _NET_ACTIVE_WINDOW atom\n");
|
||||||
|
return -1;
|
||||||
|
}
|
||||||
|
self->window = get_focused_window(self->params.egl->x11.dpy, self->net_active_window_atom);
|
||||||
|
} else {
|
||||||
|
self->window = self->params.window;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* TODO: Do these in tick, and allow error if follow_focused */
|
||||||
|
|
||||||
|
XWindowAttributes attr;
|
||||||
|
if(!XGetWindowAttributes(self->params.egl->x11.dpy, self->params.window, &attr) && !self->params.follow_focused) {
|
||||||
|
fprintf(stderr, "gsr error: gsr_capture_xcomposite_start failed: invalid window id: %lu\n", self->params.window);
|
||||||
|
return -1;
|
||||||
|
}
|
||||||
|
|
||||||
|
self->window_size.x = max_int(attr.width, 0);
|
||||||
|
self->window_size.y = max_int(attr.height, 0);
|
||||||
|
|
||||||
|
if(self->params.follow_focused)
|
||||||
|
XSelectInput(self->params.egl->x11.dpy, DefaultRootWindow(self->params.egl->x11.dpy), PropertyChangeMask);
|
||||||
|
|
||||||
|
// TODO: Get select and add these on top of it and then restore at the end. Also do the same in other xcomposite
|
||||||
|
XSelectInput(self->params.egl->x11.dpy, self->params.window, StructureNotifyMask | ExposureMask);
|
||||||
|
|
||||||
|
if(!self->params.egl->eglExportDMABUFImageQueryMESA) {
|
||||||
|
fprintf(stderr, "gsr error: gsr_capture_xcomposite_start: could not find eglExportDMABUFImageQueryMESA\n");
|
||||||
|
return -1;
|
||||||
|
}
|
||||||
|
|
||||||
|
if(!self->params.egl->eglExportDMABUFImageMESA) {
|
||||||
|
fprintf(stderr, "gsr error: gsr_capture_xcomposite_start: could not find eglExportDMABUFImageMESA\n");
|
||||||
|
return -1;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Disable vsync */
|
||||||
|
self->params.egl->eglSwapInterval(self->params.egl->egl_display, 0);
|
||||||
|
if(window_texture_init(&self->window_texture, self->params.egl->x11.dpy, self->params.window, self->params.egl) != 0 && !self->params.follow_focused) {
|
||||||
|
fprintf(stderr, "gsr error: gsr_capture_xcomposite_start: failed to get window texture for window %ld\n", self->params.window);
|
||||||
|
return -1;
|
||||||
|
}
|
||||||
|
|
||||||
|
self->texture_size.x = 0;
|
||||||
|
self->texture_size.y = 0;
|
||||||
|
|
||||||
|
self->params.egl->glBindTexture(GL_TEXTURE_2D, window_texture_get_opengl_texture_id(&self->window_texture));
|
||||||
|
self->params.egl->glGetTexLevelParameteriv(GL_TEXTURE_2D, 0, GL_TEXTURE_WIDTH, &self->texture_size.x);
|
||||||
|
self->params.egl->glGetTexLevelParameteriv(GL_TEXTURE_2D, 0, GL_TEXTURE_HEIGHT, &self->texture_size.y);
|
||||||
|
self->params.egl->glBindTexture(GL_TEXTURE_2D, 0);
|
||||||
|
|
||||||
|
self->texture_size.x = max_int(2, even_number_ceil(self->texture_size.x));
|
||||||
|
self->texture_size.y = max_int(2, even_number_ceil(self->texture_size.y));
|
||||||
|
|
||||||
|
video_codec_context->width = self->texture_size.x;
|
||||||
|
video_codec_context->height = self->texture_size.y;
|
||||||
|
|
||||||
|
if(self->params.region_size.x > 0 && self->params.region_size.y > 0) {
|
||||||
|
video_codec_context->width = max_int(2, even_number_ceil(self->params.region_size.x));
|
||||||
|
video_codec_context->height = max_int(2, even_number_ceil(self->params.region_size.y));
|
||||||
|
}
|
||||||
|
|
||||||
|
frame->width = video_codec_context->width;
|
||||||
|
frame->height = video_codec_context->height;
|
||||||
|
|
||||||
|
self->window_resize_timer = clock_get_monotonic_seconds();
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
void gsr_capture_xcomposite_stop(gsr_capture_xcomposite *self, AVCodecContext *video_codec_context) {
|
||||||
|
window_texture_deinit(&self->window_texture);
|
||||||
|
|
||||||
|
if(video_codec_context->hw_device_ctx)
|
||||||
|
av_buffer_unref(&video_codec_context->hw_device_ctx);
|
||||||
|
if(video_codec_context->hw_frames_ctx)
|
||||||
|
av_buffer_unref(&video_codec_context->hw_frames_ctx);
|
||||||
|
|
||||||
|
gsr_capture_base_stop(&self->base, self->params.egl);
|
||||||
|
}
|
||||||
|
|
||||||
|
void gsr_capture_xcomposite_tick(gsr_capture_xcomposite *self, AVCodecContext *video_codec_context) {
|
||||||
|
//self->params.egl->glClearColor(0.0f, 0.0f, 0.0f, 1.0f);
|
||||||
|
self->params.egl->glClear(0);
|
||||||
|
|
||||||
|
bool init_new_window = false;
|
||||||
|
while(XPending(self->params.egl->x11.dpy)) {
|
||||||
|
XNextEvent(self->params.egl->x11.dpy, &self->xev);
|
||||||
|
|
||||||
|
switch(self->xev.type) {
|
||||||
|
case DestroyNotify: {
|
||||||
|
/* Window died (when not following focused window), so we stop recording */
|
||||||
|
if(!self->params.follow_focused && self->xev.xdestroywindow.window == self->window) {
|
||||||
|
self->should_stop = true;
|
||||||
|
self->stop_is_error = false;
|
||||||
|
}
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
case Expose: {
|
||||||
|
/* Requires window texture recreate */
|
||||||
|
if(self->xev.xexpose.count == 0 && self->xev.xexpose.window == self->window) {
|
||||||
|
self->window_resize_timer = clock_get_monotonic_seconds();
|
||||||
|
self->window_resized = true;
|
||||||
|
}
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
case ConfigureNotify: {
|
||||||
|
/* Window resized */
|
||||||
|
if(self->xev.xconfigure.window == self->window && (self->xev.xconfigure.width != self->window_size.x || self->xev.xconfigure.height != self->window_size.y)) {
|
||||||
|
self->window_size.x = max_int(self->xev.xconfigure.width, 0);
|
||||||
|
self->window_size.y = max_int(self->xev.xconfigure.height, 0);
|
||||||
|
self->window_resize_timer = clock_get_monotonic_seconds();
|
||||||
|
self->window_resized = true;
|
||||||
|
}
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
case PropertyNotify: {
|
||||||
|
/* Focused window changed */
|
||||||
|
if(self->params.follow_focused && self->xev.xproperty.atom == self->net_active_window_atom) {
|
||||||
|
init_new_window = true;
|
||||||
|
}
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if(self->params.follow_focused && !self->follow_focused_initialized) {
|
||||||
|
init_new_window = true;
|
||||||
|
}
|
||||||
|
|
||||||
|
if(init_new_window) {
|
||||||
|
Window focused_window = get_focused_window(self->params.egl->x11.dpy, self->net_active_window_atom);
|
||||||
|
if(focused_window != self->window || !self->follow_focused_initialized) {
|
||||||
|
self->follow_focused_initialized = true;
|
||||||
|
XSelectInput(self->params.egl->x11.dpy, self->window, 0);
|
||||||
|
self->window = focused_window;
|
||||||
|
XSelectInput(self->params.egl->x11.dpy, self->window, StructureNotifyMask | ExposureMask);
|
||||||
|
|
||||||
|
XWindowAttributes attr;
|
||||||
|
attr.width = 0;
|
||||||
|
attr.height = 0;
|
||||||
|
if(!XGetWindowAttributes(self->params.egl->x11.dpy, self->window, &attr))
|
||||||
|
fprintf(stderr, "gsr error: gsr_capture_xcomposite_tick failed: invalid window id: %lu\n", self->window);
|
||||||
|
|
||||||
|
self->window_size.x = max_int(attr.width, 0);
|
||||||
|
self->window_size.y = max_int(attr.height, 0);
|
||||||
|
self->window_resized = true;
|
||||||
|
|
||||||
|
window_texture_deinit(&self->window_texture);
|
||||||
|
window_texture_init(&self->window_texture, self->params.egl->x11.dpy, self->window, self->params.egl); // TODO: Do not do the below window_texture_on_resize after this
|
||||||
|
|
||||||
|
self->texture_size.x = 0;
|
||||||
|
self->texture_size.y = 0;
|
||||||
|
|
||||||
|
self->params.egl->glBindTexture(GL_TEXTURE_2D, window_texture_get_opengl_texture_id(&self->window_texture));
|
||||||
|
self->params.egl->glGetTexLevelParameteriv(GL_TEXTURE_2D, 0, GL_TEXTURE_WIDTH, &self->texture_size.x);
|
||||||
|
self->params.egl->glGetTexLevelParameteriv(GL_TEXTURE_2D, 0, GL_TEXTURE_HEIGHT, &self->texture_size.y);
|
||||||
|
self->params.egl->glBindTexture(GL_TEXTURE_2D, 0);
|
||||||
|
|
||||||
|
self->texture_size.x = min_int(video_codec_context->width, max_int(2, even_number_ceil(self->texture_size.x)));
|
||||||
|
self->texture_size.y = min_int(video_codec_context->height, max_int(2, even_number_ceil(self->texture_size.y)));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
const double window_resize_timeout = 1.0; // 1 second
|
||||||
|
if(self->window_resized && clock_get_monotonic_seconds() - self->window_resize_timer >= window_resize_timeout) {
|
||||||
|
self->window_resized = false;
|
||||||
|
|
||||||
|
if(window_texture_on_resize(&self->window_texture) != 0) {
|
||||||
|
fprintf(stderr, "gsr error: gsr_capture_xcomposite_tick: window_texture_on_resize failed\n");
|
||||||
|
//self->should_stop = true;
|
||||||
|
//self->stop_is_error = true;
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
self->texture_size.x = 0;
|
||||||
|
self->texture_size.y = 0;
|
||||||
|
|
||||||
|
self->params.egl->glBindTexture(GL_TEXTURE_2D, window_texture_get_opengl_texture_id(&self->window_texture));
|
||||||
|
self->params.egl->glGetTexLevelParameteriv(GL_TEXTURE_2D, 0, GL_TEXTURE_WIDTH, &self->texture_size.x);
|
||||||
|
self->params.egl->glGetTexLevelParameteriv(GL_TEXTURE_2D, 0, GL_TEXTURE_HEIGHT, &self->texture_size.y);
|
||||||
|
self->params.egl->glBindTexture(GL_TEXTURE_2D, 0);
|
||||||
|
|
||||||
|
self->texture_size.x = min_int(video_codec_context->width, max_int(2, even_number_ceil(self->texture_size.x)));
|
||||||
|
self->texture_size.y = min_int(video_codec_context->height, max_int(2, even_number_ceil(self->texture_size.y)));
|
||||||
|
|
||||||
|
gsr_color_conversion_clear(&self->base.color_conversion);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
bool gsr_capture_xcomposite_should_stop(gsr_capture_xcomposite *self, bool *err) {
|
||||||
|
if(self->should_stop) {
|
||||||
|
if(err)
|
||||||
|
*err = self->stop_is_error;
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
if(err)
|
||||||
|
*err = false;
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
int gsr_capture_xcomposite_capture(gsr_capture_xcomposite *self, AVFrame *frame) {
|
||||||
|
(void)frame;
|
||||||
|
|
||||||
|
const int target_x = max_int(0, frame->width / 2 - self->texture_size.x / 2);
|
||||||
|
const int target_y = max_int(0, frame->height / 2 - self->texture_size.y / 2);
|
||||||
|
|
||||||
|
gsr_color_conversion_draw(&self->base.color_conversion, window_texture_get_opengl_texture_id(&self->window_texture),
|
||||||
|
(vec2i){target_x, target_y}, self->texture_size,
|
||||||
|
(vec2i){0, 0}, self->texture_size,
|
||||||
|
0.0f, false);
|
||||||
|
|
||||||
|
self->params.egl->eglSwapBuffers(self->params.egl->egl_display, self->params.egl->egl_surface);
|
||||||
|
//self->params.egl->glFlush();
|
||||||
|
//self->params.egl->glFinish();
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
@@ -1,438 +1,97 @@
|
|||||||
#include "../../include/capture/xcomposite_cuda.h"
|
#include "../../include/capture/xcomposite_cuda.h"
|
||||||
#include "../../include/cuda.h"
|
#include "../../include/cuda.h"
|
||||||
#include "../../include/window_texture.h"
|
#include <stdio.h>
|
||||||
#include "../../include/utils.h"
|
#include <stdlib.h>
|
||||||
#include <libavutil/hwcontext.h>
|
|
||||||
#include <libavutil/hwcontext_cuda.h>
|
|
||||||
#include <libavutil/frame.h>
|
#include <libavutil/frame.h>
|
||||||
#include <libavcodec/avcodec.h>
|
|
||||||
|
|
||||||
typedef struct {
|
typedef struct {
|
||||||
gsr_capture_xcomposite_cuda_params params;
|
gsr_capture_xcomposite xcomposite;
|
||||||
XEvent xev;
|
bool overclock;
|
||||||
|
|
||||||
bool should_stop;
|
|
||||||
bool stop_is_error;
|
|
||||||
bool window_resized;
|
|
||||||
bool follow_focused_initialized;
|
|
||||||
double window_resize_timer;
|
|
||||||
|
|
||||||
vec2i window_size;
|
|
||||||
|
|
||||||
unsigned int target_texture_id;
|
|
||||||
vec2i texture_size;
|
|
||||||
Window window;
|
|
||||||
WindowTexture window_texture;
|
|
||||||
Atom net_active_window_atom;
|
|
||||||
|
|
||||||
CUgraphicsResource cuda_graphics_resource;
|
|
||||||
CUarray mapped_array;
|
|
||||||
|
|
||||||
gsr_cuda cuda;
|
gsr_cuda cuda;
|
||||||
|
CUgraphicsResource cuda_graphics_resources[2];
|
||||||
|
CUarray mapped_arrays[2];
|
||||||
|
CUstream cuda_stream;
|
||||||
} gsr_capture_xcomposite_cuda;
|
} gsr_capture_xcomposite_cuda;
|
||||||
|
|
||||||
static int max_int(int a, int b) {
|
|
||||||
return a > b ? a : b;
|
|
||||||
}
|
|
||||||
|
|
||||||
static int min_int(int a, int b) {
|
|
||||||
return a < b ? a : b;
|
|
||||||
}
|
|
||||||
|
|
||||||
static Window get_focused_window(Display *display, Atom net_active_window_atom) {
|
|
||||||
Atom type;
|
|
||||||
int format = 0;
|
|
||||||
unsigned long num_items = 0;
|
|
||||||
unsigned long bytes_after = 0;
|
|
||||||
unsigned char *properties = NULL;
|
|
||||||
if(XGetWindowProperty(display, DefaultRootWindow(display), net_active_window_atom, 0, 1024, False, AnyPropertyType, &type, &format, &num_items, &bytes_after, &properties) == Success && properties) {
|
|
||||||
Window focused_window = *(unsigned long*)properties;
|
|
||||||
XFree(properties);
|
|
||||||
return focused_window;
|
|
||||||
}
|
|
||||||
return None;
|
|
||||||
}
|
|
||||||
|
|
||||||
static void gsr_capture_xcomposite_cuda_stop(gsr_capture *cap, AVCodecContext *video_codec_context);
|
static void gsr_capture_xcomposite_cuda_stop(gsr_capture *cap, AVCodecContext *video_codec_context);
|
||||||
|
|
||||||
static bool cuda_register_opengl_texture(gsr_capture_xcomposite_cuda *cap_xcomp) {
|
|
||||||
CUresult res;
|
|
||||||
CUcontext old_ctx;
|
|
||||||
res = cap_xcomp->cuda.cuCtxPushCurrent_v2(cap_xcomp->cuda.cu_ctx);
|
|
||||||
// TODO: Use cuGraphicsEGLRegisterImage instead with the window egl image (dont use window_texture).
|
|
||||||
// That removes the need for an extra texture and texture copy
|
|
||||||
res = cap_xcomp->cuda.cuGraphicsGLRegisterImage(
|
|
||||||
&cap_xcomp->cuda_graphics_resource, cap_xcomp->target_texture_id, GL_TEXTURE_2D,
|
|
||||||
CU_GRAPHICS_REGISTER_FLAGS_READ_ONLY);
|
|
||||||
if (res != CUDA_SUCCESS) {
|
|
||||||
const char *err_str = "unknown";
|
|
||||||
cap_xcomp->cuda.cuGetErrorString(res, &err_str);
|
|
||||||
fprintf(stderr, "gsr error: cuda_register_opengl_texture: cuGraphicsGLRegisterImage failed, error: %s, texture " "id: %u\n", err_str, cap_xcomp->target_texture_id);
|
|
||||||
res = cap_xcomp->cuda.cuCtxPopCurrent_v2(&old_ctx);
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
res = cap_xcomp->cuda.cuGraphicsResourceSetMapFlags(cap_xcomp->cuda_graphics_resource, CU_GRAPHICS_MAP_RESOURCE_FLAGS_READ_ONLY);
|
|
||||||
res = cap_xcomp->cuda.cuGraphicsMapResources(1, &cap_xcomp->cuda_graphics_resource, 0);
|
|
||||||
|
|
||||||
res = cap_xcomp->cuda.cuGraphicsSubResourceGetMappedArray(&cap_xcomp->mapped_array, cap_xcomp->cuda_graphics_resource, 0, 0);
|
|
||||||
res = cap_xcomp->cuda.cuCtxPopCurrent_v2(&old_ctx);
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
static bool cuda_create_codec_context(gsr_capture_xcomposite_cuda *cap_xcomp, AVCodecContext *video_codec_context) {
|
|
||||||
CUcontext old_ctx;
|
|
||||||
cap_xcomp->cuda.cuCtxPushCurrent_v2(cap_xcomp->cuda.cu_ctx);
|
|
||||||
|
|
||||||
AVBufferRef *device_ctx = av_hwdevice_ctx_alloc(AV_HWDEVICE_TYPE_CUDA);
|
|
||||||
if(!device_ctx) {
|
|
||||||
fprintf(stderr, "Error: Failed to create hardware device context\n");
|
|
||||||
cap_xcomp->cuda.cuCtxPopCurrent_v2(&old_ctx);
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
AVHWDeviceContext *hw_device_context = (AVHWDeviceContext*)device_ctx->data;
|
|
||||||
AVCUDADeviceContext *cuda_device_context = (AVCUDADeviceContext*)hw_device_context->hwctx;
|
|
||||||
cuda_device_context->cuda_ctx = cap_xcomp->cuda.cu_ctx;
|
|
||||||
if(av_hwdevice_ctx_init(device_ctx) < 0) {
|
|
||||||
fprintf(stderr, "Error: Failed to create hardware device context\n");
|
|
||||||
av_buffer_unref(&device_ctx);
|
|
||||||
cap_xcomp->cuda.cuCtxPopCurrent_v2(&old_ctx);
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
AVBufferRef *frame_context = av_hwframe_ctx_alloc(device_ctx);
|
|
||||||
if(!frame_context) {
|
|
||||||
fprintf(stderr, "Error: Failed to create hwframe context\n");
|
|
||||||
av_buffer_unref(&device_ctx);
|
|
||||||
cap_xcomp->cuda.cuCtxPopCurrent_v2(&old_ctx);
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
AVHWFramesContext *hw_frame_context =
|
|
||||||
(AVHWFramesContext *)frame_context->data;
|
|
||||||
hw_frame_context->width = video_codec_context->width;
|
|
||||||
hw_frame_context->height = video_codec_context->height;
|
|
||||||
hw_frame_context->sw_format = AV_PIX_FMT_BGR0;
|
|
||||||
hw_frame_context->format = video_codec_context->pix_fmt;
|
|
||||||
hw_frame_context->device_ref = device_ctx;
|
|
||||||
hw_frame_context->device_ctx = (AVHWDeviceContext*)device_ctx->data;
|
|
||||||
|
|
||||||
if (av_hwframe_ctx_init(frame_context) < 0) {
|
|
||||||
fprintf(stderr, "Error: Failed to initialize hardware frame context "
|
|
||||||
"(note: ffmpeg version needs to be > 4.0)\n");
|
|
||||||
av_buffer_unref(&device_ctx);
|
|
||||||
//av_buffer_unref(&frame_context);
|
|
||||||
cap_xcomp->cuda.cuCtxPopCurrent_v2(&old_ctx);
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
video_codec_context->hw_device_ctx = av_buffer_ref(device_ctx);
|
|
||||||
video_codec_context->hw_frames_ctx = av_buffer_ref(frame_context);
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
static unsigned int gl_create_texture(gsr_capture_xcomposite_cuda *cap_xcomp, int width, int height) {
|
|
||||||
unsigned int texture_id = 0;
|
|
||||||
cap_xcomp->params.egl->glGenTextures(1, &texture_id);
|
|
||||||
cap_xcomp->params.egl->glBindTexture(GL_TEXTURE_2D, texture_id);
|
|
||||||
cap_xcomp->params.egl->glTexImage2D(GL_TEXTURE_2D, 0, GL_RGB, width, height, 0, GL_RGB, GL_UNSIGNED_BYTE, NULL);
|
|
||||||
|
|
||||||
cap_xcomp->params.egl->glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_S, GL_CLAMP_TO_EDGE);
|
|
||||||
cap_xcomp->params.egl->glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_T, GL_CLAMP_TO_EDGE);
|
|
||||||
cap_xcomp->params.egl->glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MAG_FILTER, GL_NEAREST);
|
|
||||||
cap_xcomp->params.egl->glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_NEAREST);
|
|
||||||
|
|
||||||
cap_xcomp->params.egl->glBindTexture(GL_TEXTURE_2D, 0);
|
|
||||||
return texture_id;
|
|
||||||
}
|
|
||||||
|
|
||||||
static int gsr_capture_xcomposite_cuda_start(gsr_capture *cap, AVCodecContext *video_codec_context, AVFrame *frame) {
|
static int gsr_capture_xcomposite_cuda_start(gsr_capture *cap, AVCodecContext *video_codec_context, AVFrame *frame) {
|
||||||
gsr_capture_xcomposite_cuda *cap_xcomp = cap->priv;
|
gsr_capture_xcomposite_cuda *cap_xcomp = cap->priv;
|
||||||
|
|
||||||
if(cap_xcomp->params.follow_focused) {
|
const int res = gsr_capture_xcomposite_start(&cap_xcomp->xcomposite, video_codec_context, frame);
|
||||||
cap_xcomp->net_active_window_atom = XInternAtom(cap_xcomp->params.egl->x11.dpy, "_NET_ACTIVE_WINDOW", False);
|
if(res != 0) {
|
||||||
if(!cap_xcomp->net_active_window_atom) {
|
gsr_capture_xcomposite_cuda_stop(cap, video_codec_context);
|
||||||
fprintf(stderr, "gsr error: gsr_capture_xcomposite_cuda_start failed: failed to get _NET_ACTIVE_WINDOW atom\n");
|
return res;
|
||||||
return -1;
|
|
||||||
}
|
|
||||||
cap_xcomp->window = get_focused_window(cap_xcomp->params.egl->x11.dpy, cap_xcomp->net_active_window_atom);
|
|
||||||
} else {
|
|
||||||
cap_xcomp->window = cap_xcomp->params.window;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/* TODO: Do these in tick, and allow error if follow_focused */
|
// TODO: overclocking is not supported on wayland...
|
||||||
|
if(!gsr_cuda_load(&cap_xcomp->cuda, NULL, false)) {
|
||||||
XWindowAttributes attr;
|
fprintf(stderr, "gsr error: gsr_capture_kms_cuda_start: failed to load cuda\n");
|
||||||
attr.width = 0;
|
|
||||||
attr.height = 0;
|
|
||||||
if(!XGetWindowAttributes(cap_xcomp->params.egl->x11.dpy, cap_xcomp->window, &attr) && !cap_xcomp->params.follow_focused) {
|
|
||||||
fprintf(stderr, "gsr error: gsr_capture_xcomposite_cuda_start failed: invalid window id: %lu\n", cap_xcomp->window);
|
|
||||||
return -1;
|
|
||||||
}
|
|
||||||
|
|
||||||
cap_xcomp->window_size.x = max_int(attr.width, 0);
|
|
||||||
cap_xcomp->window_size.y = max_int(attr.height, 0);
|
|
||||||
|
|
||||||
if(cap_xcomp->params.follow_focused)
|
|
||||||
XSelectInput(cap_xcomp->params.egl->x11.dpy, DefaultRootWindow(cap_xcomp->params.egl->x11.dpy), PropertyChangeMask);
|
|
||||||
|
|
||||||
XSelectInput(cap_xcomp->params.egl->x11.dpy, cap_xcomp->window, StructureNotifyMask | ExposureMask);
|
|
||||||
|
|
||||||
cap_xcomp->params.egl->eglSwapInterval(cap_xcomp->params.egl->egl_display, 0);
|
|
||||||
if(window_texture_init(&cap_xcomp->window_texture, cap_xcomp->params.egl->x11.dpy, cap_xcomp->window, cap_xcomp->params.egl) != 0 && !cap_xcomp->params.follow_focused) {
|
|
||||||
fprintf(stderr, "gsr error: gsr_capture_xcomposite_cuda_start: failed to get window texture for window %ld\n", cap_xcomp->window);
|
|
||||||
return -1;
|
|
||||||
}
|
|
||||||
|
|
||||||
cap_xcomp->texture_size.x = 0;
|
|
||||||
cap_xcomp->texture_size.y = 0;
|
|
||||||
|
|
||||||
cap_xcomp->params.egl->glBindTexture(GL_TEXTURE_2D, window_texture_get_opengl_texture_id(&cap_xcomp->window_texture));
|
|
||||||
cap_xcomp->params.egl->glGetTexLevelParameteriv(GL_TEXTURE_2D, 0, GL_TEXTURE_WIDTH, &cap_xcomp->texture_size.x);
|
|
||||||
cap_xcomp->params.egl->glGetTexLevelParameteriv(GL_TEXTURE_2D, 0, GL_TEXTURE_HEIGHT, &cap_xcomp->texture_size.y);
|
|
||||||
cap_xcomp->params.egl->glBindTexture(GL_TEXTURE_2D, 0);
|
|
||||||
|
|
||||||
cap_xcomp->texture_size.x = max_int(2, cap_xcomp->texture_size.x & ~1);
|
|
||||||
cap_xcomp->texture_size.y = max_int(2, cap_xcomp->texture_size.y & ~1);
|
|
||||||
|
|
||||||
video_codec_context->width = cap_xcomp->texture_size.x;
|
|
||||||
video_codec_context->height = cap_xcomp->texture_size.y;
|
|
||||||
|
|
||||||
if(cap_xcomp->params.region_size.x > 0 && cap_xcomp->params.region_size.y > 0) {
|
|
||||||
video_codec_context->width = max_int(2, cap_xcomp->params.region_size.x & ~1);
|
|
||||||
video_codec_context->height = max_int(2, cap_xcomp->params.region_size.y & ~1);
|
|
||||||
}
|
|
||||||
|
|
||||||
frame->width = video_codec_context->width;
|
|
||||||
frame->height = video_codec_context->height;
|
|
||||||
|
|
||||||
cap_xcomp->target_texture_id = gl_create_texture(cap_xcomp, video_codec_context->width, video_codec_context->height);
|
|
||||||
if(cap_xcomp->target_texture_id == 0) {
|
|
||||||
fprintf(stderr, "gsr error: gsr_capture_xcomposite_cuda_start: failed to create opengl texture\n");
|
|
||||||
gsr_capture_xcomposite_cuda_stop(cap, video_codec_context);
|
gsr_capture_xcomposite_cuda_stop(cap, video_codec_context);
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
if(!gsr_cuda_load(&cap_xcomp->cuda, cap_xcomp->params.egl->x11.dpy, cap_xcomp->params.overclock)) {
|
if(!cuda_create_codec_context(cap_xcomp->cuda.cu_ctx, video_codec_context, &cap_xcomp->cuda_stream)) {
|
||||||
gsr_capture_xcomposite_cuda_stop(cap, video_codec_context);
|
gsr_capture_xcomposite_cuda_stop(cap, video_codec_context);
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
if(!cuda_create_codec_context(cap_xcomp, video_codec_context)) {
|
gsr_cuda_context cuda_context = {
|
||||||
|
.cuda = &cap_xcomp->cuda,
|
||||||
|
.cuda_graphics_resources = cap_xcomp->cuda_graphics_resources,
|
||||||
|
.mapped_arrays = cap_xcomp->mapped_arrays
|
||||||
|
};
|
||||||
|
|
||||||
|
if(!gsr_capture_base_setup_cuda_textures(&cap_xcomp->xcomposite.base, frame, &cuda_context, cap_xcomp->xcomposite.params.egl, cap_xcomp->xcomposite.params.color_range, GSR_SOURCE_COLOR_RGB, false)) {
|
||||||
gsr_capture_xcomposite_cuda_stop(cap, video_codec_context);
|
gsr_capture_xcomposite_cuda_stop(cap, video_codec_context);
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
if(!cuda_register_opengl_texture(cap_xcomp)) {
|
|
||||||
gsr_capture_xcomposite_cuda_stop(cap, video_codec_context);
|
|
||||||
return -1;
|
|
||||||
}
|
|
||||||
|
|
||||||
if(av_hwframe_get_buffer(video_codec_context->hw_frames_ctx, frame, 0) < 0) {
|
|
||||||
fprintf(stderr, "gsr error: gsr_capture_xcomposite_cuda_start: av_hwframe_get_buffer failed\n");
|
|
||||||
gsr_capture_xcomposite_cuda_stop(cap, video_codec_context);
|
|
||||||
return -1;
|
|
||||||
}
|
|
||||||
|
|
||||||
cap_xcomp->window_resize_timer = clock_get_monotonic_seconds();
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void gsr_capture_xcomposite_cuda_stop(gsr_capture *cap, AVCodecContext *video_codec_context) {
|
static void gsr_capture_xcomposite_unload_cuda_graphics(gsr_capture_xcomposite_cuda *cap_xcomp) {
|
||||||
gsr_capture_xcomposite_cuda *cap_xcomp = cap->priv;
|
|
||||||
|
|
||||||
if(cap_xcomp->cuda.cu_ctx) {
|
if(cap_xcomp->cuda.cu_ctx) {
|
||||||
CUcontext old_ctx;
|
CUcontext old_ctx;
|
||||||
cap_xcomp->cuda.cuCtxPushCurrent_v2(cap_xcomp->cuda.cu_ctx);
|
cap_xcomp->cuda.cuCtxPushCurrent_v2(cap_xcomp->cuda.cu_ctx);
|
||||||
|
|
||||||
if(cap_xcomp->cuda_graphics_resource) {
|
for(int i = 0; i < 2; ++i) {
|
||||||
cap_xcomp->cuda.cuGraphicsUnmapResources(1, &cap_xcomp->cuda_graphics_resource, 0);
|
if(cap_xcomp->cuda_graphics_resources[i]) {
|
||||||
cap_xcomp->cuda.cuGraphicsUnregisterResource(cap_xcomp->cuda_graphics_resource);
|
cap_xcomp->cuda.cuGraphicsUnmapResources(1, &cap_xcomp->cuda_graphics_resources[i], 0);
|
||||||
|
cap_xcomp->cuda.cuGraphicsUnregisterResource(cap_xcomp->cuda_graphics_resources[i]);
|
||||||
|
cap_xcomp->cuda_graphics_resources[i] = 0;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
cap_xcomp->cuda.cuCtxPopCurrent_v2(&old_ctx);
|
cap_xcomp->cuda.cuCtxPopCurrent_v2(&old_ctx);
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
window_texture_deinit(&cap_xcomp->window_texture);
|
static void gsr_capture_xcomposite_cuda_stop(gsr_capture *cap, AVCodecContext *video_codec_context) {
|
||||||
|
gsr_capture_xcomposite_cuda *cap_xcomp = cap->priv;
|
||||||
if(cap_xcomp->target_texture_id) {
|
gsr_capture_xcomposite_stop(&cap_xcomp->xcomposite, video_codec_context);
|
||||||
cap_xcomp->params.egl->glDeleteTextures(1, &cap_xcomp->target_texture_id);
|
gsr_capture_xcomposite_unload_cuda_graphics(cap_xcomp);
|
||||||
cap_xcomp->target_texture_id = 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
if(video_codec_context->hw_device_ctx)
|
|
||||||
av_buffer_unref(&video_codec_context->hw_device_ctx);
|
|
||||||
if(video_codec_context->hw_frames_ctx)
|
|
||||||
av_buffer_unref(&video_codec_context->hw_frames_ctx);
|
|
||||||
|
|
||||||
gsr_cuda_unload(&cap_xcomp->cuda);
|
gsr_cuda_unload(&cap_xcomp->cuda);
|
||||||
|
|
||||||
if(cap_xcomp->params.egl->x11.dpy) {
|
|
||||||
// TODO: This causes a crash, why? maybe some other library dlclose xlib and that also happened to unload this???
|
|
||||||
//XCloseDisplay(cap_xcomp->dpy);
|
|
||||||
cap_xcomp->params.egl->x11.dpy = NULL;
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static void gsr_capture_xcomposite_cuda_tick(gsr_capture *cap, AVCodecContext *video_codec_context) {
|
static void gsr_capture_xcomposite_cuda_tick(gsr_capture *cap, AVCodecContext *video_codec_context) {
|
||||||
gsr_capture_xcomposite_cuda *cap_xcomp = cap->priv;
|
gsr_capture_xcomposite_cuda *cap_xcomp = cap->priv;
|
||||||
|
gsr_capture_xcomposite_tick(&cap_xcomp->xcomposite, video_codec_context);
|
||||||
bool init_new_window = false;
|
|
||||||
while(XPending(cap_xcomp->params.egl->x11.dpy)) {
|
|
||||||
XNextEvent(cap_xcomp->params.egl->x11.dpy, &cap_xcomp->xev);
|
|
||||||
|
|
||||||
switch(cap_xcomp->xev.type) {
|
|
||||||
case DestroyNotify: {
|
|
||||||
/* Window died (when not following focused window), so we stop recording */
|
|
||||||
if(!cap_xcomp->params.follow_focused && cap_xcomp->xev.xdestroywindow.window == cap_xcomp->window) {
|
|
||||||
cap_xcomp->should_stop = true;
|
|
||||||
cap_xcomp->stop_is_error = false;
|
|
||||||
}
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
case Expose: {
|
|
||||||
/* Requires window texture recreate */
|
|
||||||
if(cap_xcomp->xev.xexpose.count == 0 && cap_xcomp->xev.xexpose.window == cap_xcomp->window) {
|
|
||||||
cap_xcomp->window_resize_timer = clock_get_monotonic_seconds();
|
|
||||||
cap_xcomp->window_resized = true;
|
|
||||||
}
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
case ConfigureNotify: {
|
|
||||||
/* Window resized */
|
|
||||||
if(cap_xcomp->xev.xconfigure.window == cap_xcomp->window && (cap_xcomp->xev.xconfigure.width != cap_xcomp->window_size.x || cap_xcomp->xev.xconfigure.height != cap_xcomp->window_size.y)) {
|
|
||||||
cap_xcomp->window_size.x = max_int(cap_xcomp->xev.xconfigure.width, 0);
|
|
||||||
cap_xcomp->window_size.y = max_int(cap_xcomp->xev.xconfigure.height, 0);
|
|
||||||
cap_xcomp->window_resize_timer = clock_get_monotonic_seconds();
|
|
||||||
cap_xcomp->window_resized = true;
|
|
||||||
}
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
case PropertyNotify: {
|
|
||||||
/* Focused window changed */
|
|
||||||
if(cap_xcomp->params.follow_focused && cap_xcomp->xev.xproperty.atom == cap_xcomp->net_active_window_atom) {
|
|
||||||
init_new_window = true;
|
|
||||||
}
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if(cap_xcomp->params.follow_focused && !cap_xcomp->follow_focused_initialized) {
|
|
||||||
init_new_window = true;
|
|
||||||
}
|
|
||||||
|
|
||||||
if(init_new_window) {
|
|
||||||
Window focused_window = get_focused_window(cap_xcomp->params.egl->x11.dpy, cap_xcomp->net_active_window_atom);
|
|
||||||
if(focused_window != cap_xcomp->window || !cap_xcomp->follow_focused_initialized) {
|
|
||||||
cap_xcomp->follow_focused_initialized = true;
|
|
||||||
XSelectInput(cap_xcomp->params.egl->x11.dpy, cap_xcomp->window, 0);
|
|
||||||
cap_xcomp->window = focused_window;
|
|
||||||
XSelectInput(cap_xcomp->params.egl->x11.dpy, cap_xcomp->window, StructureNotifyMask | ExposureMask);
|
|
||||||
|
|
||||||
XWindowAttributes attr;
|
|
||||||
attr.width = 0;
|
|
||||||
attr.height = 0;
|
|
||||||
if(!XGetWindowAttributes(cap_xcomp->params.egl->x11.dpy, cap_xcomp->window, &attr))
|
|
||||||
fprintf(stderr, "gsr error: gsr_capture_xcomposite_cuda_tick failed: invalid window id: %lu\n", cap_xcomp->window);
|
|
||||||
|
|
||||||
cap_xcomp->window_size.x = max_int(attr.width, 0);
|
|
||||||
cap_xcomp->window_size.y = max_int(attr.height, 0);
|
|
||||||
cap_xcomp->window_resized = true;
|
|
||||||
|
|
||||||
window_texture_deinit(&cap_xcomp->window_texture);
|
|
||||||
window_texture_init(&cap_xcomp->window_texture, cap_xcomp->params.egl->x11.dpy, cap_xcomp->window, cap_xcomp->params.egl); // TODO: Do not do the below window_texture_on_resize after this
|
|
||||||
|
|
||||||
cap_xcomp->texture_size.x = 0;
|
|
||||||
cap_xcomp->texture_size.y = 0;
|
|
||||||
|
|
||||||
cap_xcomp->params.egl->glBindTexture(GL_TEXTURE_2D, window_texture_get_opengl_texture_id(&cap_xcomp->window_texture));
|
|
||||||
cap_xcomp->params.egl->glGetTexLevelParameteriv(GL_TEXTURE_2D, 0, GL_TEXTURE_WIDTH, &cap_xcomp->texture_size.x);
|
|
||||||
cap_xcomp->params.egl->glGetTexLevelParameteriv(GL_TEXTURE_2D, 0, GL_TEXTURE_HEIGHT, &cap_xcomp->texture_size.y);
|
|
||||||
cap_xcomp->params.egl->glBindTexture(GL_TEXTURE_2D, 0);
|
|
||||||
|
|
||||||
cap_xcomp->texture_size.x = min_int(video_codec_context->width, max_int(2, cap_xcomp->texture_size.x & ~1));
|
|
||||||
cap_xcomp->texture_size.y = min_int(video_codec_context->height, max_int(2, cap_xcomp->texture_size.y & ~1));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
const double window_resize_timeout = 1.0; // 1 second
|
|
||||||
if(cap_xcomp->window_resized && clock_get_monotonic_seconds() - cap_xcomp->window_resize_timer >= window_resize_timeout) {
|
|
||||||
cap_xcomp->window_resized = false;
|
|
||||||
if(window_texture_on_resize(&cap_xcomp->window_texture) != 0) {
|
|
||||||
fprintf(stderr, "gsr error: gsr_capture_xcomposite_cuda_tick: window_texture_on_resize failed\n");
|
|
||||||
//cap_xcomp->should_stop = true;
|
|
||||||
//cap_xcomp->stop_is_error = true;
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
cap_xcomp->texture_size.x = 0;
|
|
||||||
cap_xcomp->texture_size.y = 0;
|
|
||||||
|
|
||||||
cap_xcomp->params.egl->glBindTexture(GL_TEXTURE_2D, window_texture_get_opengl_texture_id(&cap_xcomp->window_texture));
|
|
||||||
cap_xcomp->params.egl->glGetTexLevelParameteriv(GL_TEXTURE_2D, 0, GL_TEXTURE_WIDTH, &cap_xcomp->texture_size.x);
|
|
||||||
cap_xcomp->params.egl->glGetTexLevelParameteriv(GL_TEXTURE_2D, 0, GL_TEXTURE_HEIGHT, &cap_xcomp->texture_size.y);
|
|
||||||
cap_xcomp->params.egl->glBindTexture(GL_TEXTURE_2D, 0);
|
|
||||||
|
|
||||||
cap_xcomp->texture_size.x = min_int(video_codec_context->width, max_int(2, cap_xcomp->texture_size.x & ~1));
|
|
||||||
cap_xcomp->texture_size.y = min_int(video_codec_context->height, max_int(2, cap_xcomp->texture_size.y & ~1));
|
|
||||||
|
|
||||||
// Clear texture with black background because the source texture (window_texture_get_opengl_texture_id(&cap_xcomp->window_texture))
|
|
||||||
// might be smaller than cap_xcomp->target_texture_id
|
|
||||||
cap_xcomp->params.egl->glClearTexImage(cap_xcomp->target_texture_id, 0, GL_RGB, GL_UNSIGNED_BYTE, NULL);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static bool gsr_capture_xcomposite_cuda_should_stop(gsr_capture *cap, bool *err) {
|
static bool gsr_capture_xcomposite_cuda_should_stop(gsr_capture *cap, bool *err) {
|
||||||
gsr_capture_xcomposite_cuda *cap_xcomp = cap->priv;
|
gsr_capture_xcomposite_cuda *cap_xcomp = cap->priv;
|
||||||
if(cap_xcomp->should_stop) {
|
return gsr_capture_xcomposite_should_stop(&cap_xcomp->xcomposite, err);
|
||||||
if(err)
|
|
||||||
*err = cap_xcomp->stop_is_error;
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
if(err)
|
|
||||||
*err = false;
|
|
||||||
return false;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static int gsr_capture_xcomposite_cuda_capture(gsr_capture *cap, AVFrame *frame) {
|
static int gsr_capture_xcomposite_cuda_capture(gsr_capture *cap, AVFrame *frame) {
|
||||||
gsr_capture_xcomposite_cuda *cap_xcomp = cap->priv;
|
gsr_capture_xcomposite_cuda *cap_xcomp = cap->priv;
|
||||||
|
|
||||||
//cap_xcomp->params.egl->glClearColor(0.0f, 0.0f, 0.0f, 1.0f);
|
gsr_capture_xcomposite_capture(&cap_xcomp->xcomposite, frame);
|
||||||
cap_xcomp->params.egl->glClear(0);
|
|
||||||
|
|
||||||
vec2i source_pos = { 0, 0 };
|
|
||||||
vec2i source_size = cap_xcomp->texture_size;
|
|
||||||
|
|
||||||
if(cap_xcomp->window_texture.texture_id != 0) {
|
|
||||||
while(cap_xcomp->params.egl->glGetError()) {}
|
|
||||||
|
|
||||||
const int target_x = max_int(0, frame->width / 2 - cap_xcomp->texture_size.x / 2);
|
|
||||||
const int target_y = max_int(0, frame->height / 2 - cap_xcomp->texture_size.y / 2);
|
|
||||||
|
|
||||||
/* TODO: Remove this copy, which is only possible by using nvenc directly and encoding window_pixmap.target_texture_id */
|
|
||||||
cap_xcomp->params.egl->glCopyImageSubData(
|
|
||||||
window_texture_get_opengl_texture_id(&cap_xcomp->window_texture), GL_TEXTURE_2D, 0, source_pos.x, source_pos.y, 0,
|
|
||||||
cap_xcomp->target_texture_id, GL_TEXTURE_2D, 0, target_x, target_y, 0,
|
|
||||||
source_size.x, source_size.y, 1);
|
|
||||||
unsigned int err = cap_xcomp->params.egl->glGetError();
|
|
||||||
if(err != 0) {
|
|
||||||
static bool error_shown = false;
|
|
||||||
if(!error_shown) {
|
|
||||||
error_shown = true;
|
|
||||||
fprintf(stderr, "gsr error: gsr_capture_xcomposite_cuda_capture: glCopyImageSubData failed, gl error: %d\n", err);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
cap_xcomp->params.egl->eglSwapBuffers(cap_xcomp->params.egl->egl_display, cap_xcomp->params.egl->egl_surface);
|
|
||||||
|
|
||||||
frame->linesize[0] = frame->width * 4;
|
|
||||||
//frame->linesize[0] = frame->width * 1;
|
|
||||||
//frame->linesize[1] = frame->width * 1;
|
|
||||||
//frame->linesize[2] = frame->width * 1;
|
|
||||||
|
|
||||||
|
const int div[2] = {1, 2}; // divide UV texture size by 2 because chroma is half size
|
||||||
|
for(int i = 0; i < 2; ++i) {
|
||||||
CUDA_MEMCPY2D memcpy_struct;
|
CUDA_MEMCPY2D memcpy_struct;
|
||||||
memcpy_struct.srcXInBytes = 0;
|
memcpy_struct.srcXInBytes = 0;
|
||||||
memcpy_struct.srcY = 0;
|
memcpy_struct.srcY = 0;
|
||||||
@@ -442,15 +101,18 @@ static int gsr_capture_xcomposite_cuda_capture(gsr_capture *cap, AVFrame *frame)
|
|||||||
memcpy_struct.dstY = 0;
|
memcpy_struct.dstY = 0;
|
||||||
memcpy_struct.dstMemoryType = CU_MEMORYTYPE_DEVICE;
|
memcpy_struct.dstMemoryType = CU_MEMORYTYPE_DEVICE;
|
||||||
|
|
||||||
memcpy_struct.srcArray = cap_xcomp->mapped_array;
|
memcpy_struct.srcArray = cap_xcomp->mapped_arrays[i];
|
||||||
memcpy_struct.dstDevice = (CUdeviceptr)frame->data[0];
|
memcpy_struct.srcPitch = frame->width / div[i];
|
||||||
memcpy_struct.dstPitch = frame->linesize[0];
|
memcpy_struct.dstDevice = (CUdeviceptr)frame->data[i];
|
||||||
memcpy_struct.WidthInBytes = frame->width * 4;//frame->width * 1;
|
memcpy_struct.dstPitch = frame->linesize[i];
|
||||||
memcpy_struct.Height = frame->height;
|
memcpy_struct.WidthInBytes = frame->width;
|
||||||
cap_xcomp->cuda.cuMemcpy2D_v2(&memcpy_struct);
|
memcpy_struct.Height = frame->height / div[i];
|
||||||
|
// TODO: Remove this copy if possible
|
||||||
|
cap_xcomp->cuda.cuMemcpy2DAsync_v2(&memcpy_struct, cap_xcomp->cuda_stream);
|
||||||
|
}
|
||||||
|
|
||||||
//frame->data[1] = frame->data[0];
|
// TODO: needed?
|
||||||
//frame->data[2] = frame->data[0];
|
cap_xcomp->cuda.cuStreamSynchronize(cap_xcomp->cuda_stream);
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
@@ -480,7 +142,8 @@ gsr_capture* gsr_capture_xcomposite_cuda_create(const gsr_capture_xcomposite_cud
|
|||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
cap_xcomp->params = *params;
|
gsr_capture_xcomposite_init(&cap_xcomp->xcomposite, ¶ms->base);
|
||||||
|
cap_xcomp->overclock = params->overclock;
|
||||||
|
|
||||||
*cap = (gsr_capture) {
|
*cap = (gsr_capture) {
|
||||||
.start = gsr_capture_xcomposite_cuda_start,
|
.start = gsr_capture_xcomposite_cuda_start,
|
||||||
|
|||||||
@@ -1,341 +1,55 @@
|
|||||||
#include "../../include/capture/xcomposite_vaapi.h"
|
#include "../../include/capture/xcomposite_vaapi.h"
|
||||||
#include "../../include/window_texture.h"
|
#include "../../include/capture/xcomposite.h"
|
||||||
#include "../../include/utils.h"
|
#include <unistd.h>
|
||||||
#include <stdlib.h>
|
#include <stdlib.h>
|
||||||
#include <stdio.h>
|
#include <stdio.h>
|
||||||
#include <unistd.h>
|
|
||||||
#include <assert.h>
|
|
||||||
#include <X11/Xlib.h>
|
|
||||||
#include <libavutil/hwcontext.h>
|
|
||||||
#include <libavutil/hwcontext_vaapi.h>
|
|
||||||
#include <libavutil/frame.h>
|
|
||||||
#include <libavcodec/avcodec.h>
|
|
||||||
#include <va/va.h>
|
#include <va/va.h>
|
||||||
#include <va/va_drmcommon.h>
|
#include <va/va_drmcommon.h>
|
||||||
|
|
||||||
typedef struct {
|
typedef struct {
|
||||||
gsr_capture_base base;
|
gsr_capture_xcomposite xcomposite;
|
||||||
gsr_capture_xcomposite_vaapi_params params;
|
|
||||||
XEvent xev;
|
|
||||||
|
|
||||||
bool should_stop;
|
|
||||||
bool stop_is_error;
|
|
||||||
bool window_resized;
|
|
||||||
bool follow_focused_initialized;
|
|
||||||
|
|
||||||
Window window;
|
|
||||||
vec2i window_size;
|
|
||||||
vec2i texture_size;
|
|
||||||
double window_resize_timer;
|
|
||||||
|
|
||||||
WindowTexture window_texture;
|
|
||||||
|
|
||||||
VADisplay va_dpy;
|
VADisplay va_dpy;
|
||||||
VADRMPRIMESurfaceDescriptor prime;
|
VADRMPRIMESurfaceDescriptor prime;
|
||||||
|
|
||||||
Atom net_active_window_atom;
|
|
||||||
} gsr_capture_xcomposite_vaapi;
|
} gsr_capture_xcomposite_vaapi;
|
||||||
|
|
||||||
static int max_int(int a, int b) {
|
|
||||||
return a > b ? a : b;
|
|
||||||
}
|
|
||||||
|
|
||||||
static int min_int(int a, int b) {
|
|
||||||
return a < b ? a : b;
|
|
||||||
}
|
|
||||||
|
|
||||||
static void gsr_capture_xcomposite_vaapi_stop(gsr_capture *cap, AVCodecContext *video_codec_context);
|
static void gsr_capture_xcomposite_vaapi_stop(gsr_capture *cap, AVCodecContext *video_codec_context);
|
||||||
|
|
||||||
static Window get_focused_window(Display *display, Atom net_active_window_atom) {
|
|
||||||
Atom type;
|
|
||||||
int format = 0;
|
|
||||||
unsigned long num_items = 0;
|
|
||||||
unsigned long bytes_after = 0;
|
|
||||||
unsigned char *properties = NULL;
|
|
||||||
if(XGetWindowProperty(display, DefaultRootWindow(display), net_active_window_atom, 0, 1024, False, AnyPropertyType, &type, &format, &num_items, &bytes_after, &properties) == Success && properties) {
|
|
||||||
Window focused_window = *(unsigned long*)properties;
|
|
||||||
XFree(properties);
|
|
||||||
return focused_window;
|
|
||||||
}
|
|
||||||
return None;
|
|
||||||
}
|
|
||||||
|
|
||||||
static bool drm_create_codec_context(gsr_capture_xcomposite_vaapi *cap_xcomp, AVCodecContext *video_codec_context) {
|
|
||||||
char render_path[128];
|
|
||||||
if(!gsr_card_path_get_render_path(cap_xcomp->params.egl->card_path, render_path)) {
|
|
||||||
fprintf(stderr, "gsr error: failed to get /dev/dri/renderDXXX file from %s\n", cap_xcomp->params.egl->card_path);
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
AVBufferRef *device_ctx;
|
|
||||||
if(av_hwdevice_ctx_create(&device_ctx, AV_HWDEVICE_TYPE_VAAPI, render_path, NULL, 0) < 0) {
|
|
||||||
fprintf(stderr, "Error: Failed to create hardware device context\n");
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
AVBufferRef *frame_context = av_hwframe_ctx_alloc(device_ctx);
|
|
||||||
if(!frame_context) {
|
|
||||||
fprintf(stderr, "Error: Failed to create hwframe context\n");
|
|
||||||
av_buffer_unref(&device_ctx);
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
AVHWFramesContext *hw_frame_context =
|
|
||||||
(AVHWFramesContext *)frame_context->data;
|
|
||||||
hw_frame_context->width = video_codec_context->width;
|
|
||||||
hw_frame_context->height = video_codec_context->height;
|
|
||||||
hw_frame_context->sw_format = AV_PIX_FMT_NV12;//AV_PIX_FMT_0RGB32;//AV_PIX_FMT_YUV420P;//AV_PIX_FMT_0RGB32;//AV_PIX_FMT_NV12;
|
|
||||||
hw_frame_context->format = video_codec_context->pix_fmt;
|
|
||||||
hw_frame_context->device_ref = device_ctx;
|
|
||||||
hw_frame_context->device_ctx = (AVHWDeviceContext*)device_ctx->data;
|
|
||||||
|
|
||||||
//hw_frame_context->initial_pool_size = 20;
|
|
||||||
|
|
||||||
AVVAAPIDeviceContext *vactx =((AVHWDeviceContext*)device_ctx->data)->hwctx;
|
|
||||||
cap_xcomp->va_dpy = vactx->display;
|
|
||||||
|
|
||||||
if (av_hwframe_ctx_init(frame_context) < 0) {
|
|
||||||
fprintf(stderr, "Error: Failed to initialize hardware frame context "
|
|
||||||
"(note: ffmpeg version needs to be > 4.0)\n");
|
|
||||||
av_buffer_unref(&device_ctx);
|
|
||||||
//av_buffer_unref(&frame_context);
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
video_codec_context->hw_device_ctx = av_buffer_ref(device_ctx);
|
|
||||||
video_codec_context->hw_frames_ctx = av_buffer_ref(frame_context);
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
#define DRM_FORMAT_MOD_INVALID 0xffffffffffffffULL
|
|
||||||
|
|
||||||
static int gsr_capture_xcomposite_vaapi_start(gsr_capture *cap, AVCodecContext *video_codec_context, AVFrame *frame) {
|
static int gsr_capture_xcomposite_vaapi_start(gsr_capture *cap, AVCodecContext *video_codec_context, AVFrame *frame) {
|
||||||
gsr_capture_xcomposite_vaapi *cap_xcomp = cap->priv;
|
gsr_capture_xcomposite_vaapi *cap_xcomp = cap->priv;
|
||||||
|
|
||||||
cap_xcomp->base.video_codec_context = video_codec_context;
|
const int res = gsr_capture_xcomposite_start(&cap_xcomp->xcomposite, video_codec_context, frame);
|
||||||
|
if(res != 0) {
|
||||||
if(cap_xcomp->params.follow_focused) {
|
gsr_capture_xcomposite_vaapi_stop(cap, video_codec_context);
|
||||||
cap_xcomp->net_active_window_atom = XInternAtom(cap_xcomp->params.egl->x11.dpy, "_NET_ACTIVE_WINDOW", False);
|
return res;
|
||||||
if(!cap_xcomp->net_active_window_atom) {
|
|
||||||
fprintf(stderr, "gsr error: gsr_capture_xcomposite_vaapi_start failed: failed to get _NET_ACTIVE_WINDOW atom\n");
|
|
||||||
return -1;
|
|
||||||
}
|
|
||||||
cap_xcomp->window = get_focused_window(cap_xcomp->params.egl->x11.dpy, cap_xcomp->net_active_window_atom);
|
|
||||||
} else {
|
|
||||||
cap_xcomp->window = cap_xcomp->params.window;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/* TODO: Do these in tick, and allow error if follow_focused */
|
if(!drm_create_codec_context(cap_xcomp->xcomposite.params.egl->card_path, video_codec_context, false, &cap_xcomp->va_dpy)) {
|
||||||
|
|
||||||
XWindowAttributes attr;
|
|
||||||
if(!XGetWindowAttributes(cap_xcomp->params.egl->x11.dpy, cap_xcomp->params.window, &attr) && !cap_xcomp->params.follow_focused) {
|
|
||||||
fprintf(stderr, "gsr error: gsr_capture_xcomposite_vaapi_start failed: invalid window id: %lu\n", cap_xcomp->params.window);
|
|
||||||
return -1;
|
|
||||||
}
|
|
||||||
|
|
||||||
cap_xcomp->window_size.x = max_int(attr.width, 0);
|
|
||||||
cap_xcomp->window_size.y = max_int(attr.height, 0);
|
|
||||||
|
|
||||||
if(cap_xcomp->params.follow_focused)
|
|
||||||
XSelectInput(cap_xcomp->params.egl->x11.dpy, DefaultRootWindow(cap_xcomp->params.egl->x11.dpy), PropertyChangeMask);
|
|
||||||
|
|
||||||
// TODO: Get select and add these on top of it and then restore at the end. Also do the same in other xcomposite
|
|
||||||
XSelectInput(cap_xcomp->params.egl->x11.dpy, cap_xcomp->params.window, StructureNotifyMask | ExposureMask);
|
|
||||||
|
|
||||||
if(!cap_xcomp->params.egl->eglExportDMABUFImageQueryMESA) {
|
|
||||||
fprintf(stderr, "gsr error: gsr_capture_xcomposite_vaapi_start: could not find eglExportDMABUFImageQueryMESA\n");
|
|
||||||
return -1;
|
|
||||||
}
|
|
||||||
|
|
||||||
if(!cap_xcomp->params.egl->eglExportDMABUFImageMESA) {
|
|
||||||
fprintf(stderr, "gsr error: gsr_capture_xcomposite_vaapi_start: could not find eglExportDMABUFImageMESA\n");
|
|
||||||
return -1;
|
|
||||||
}
|
|
||||||
|
|
||||||
/* Disable vsync */
|
|
||||||
cap_xcomp->params.egl->eglSwapInterval(cap_xcomp->params.egl->egl_display, 0);
|
|
||||||
if(window_texture_init(&cap_xcomp->window_texture, cap_xcomp->params.egl->x11.dpy, cap_xcomp->params.window, cap_xcomp->params.egl) != 0 && !cap_xcomp->params.follow_focused) {
|
|
||||||
fprintf(stderr, "gsr error: gsr_capture_xcomposite_vaapi_start: failed to get window texture for window %ld\n", cap_xcomp->params.window);
|
|
||||||
return -1;
|
|
||||||
}
|
|
||||||
|
|
||||||
cap_xcomp->texture_size.x = 0;
|
|
||||||
cap_xcomp->texture_size.y = 0;
|
|
||||||
|
|
||||||
cap_xcomp->params.egl->glBindTexture(GL_TEXTURE_2D, window_texture_get_opengl_texture_id(&cap_xcomp->window_texture));
|
|
||||||
cap_xcomp->params.egl->glGetTexLevelParameteriv(GL_TEXTURE_2D, 0, GL_TEXTURE_WIDTH, &cap_xcomp->texture_size.x);
|
|
||||||
cap_xcomp->params.egl->glGetTexLevelParameteriv(GL_TEXTURE_2D, 0, GL_TEXTURE_HEIGHT, &cap_xcomp->texture_size.y);
|
|
||||||
cap_xcomp->params.egl->glBindTexture(GL_TEXTURE_2D, 0);
|
|
||||||
|
|
||||||
cap_xcomp->texture_size.x = max_int(2, even_number_ceil(cap_xcomp->texture_size.x));
|
|
||||||
cap_xcomp->texture_size.y = max_int(2, even_number_ceil(cap_xcomp->texture_size.y));
|
|
||||||
|
|
||||||
video_codec_context->width = cap_xcomp->texture_size.x;
|
|
||||||
video_codec_context->height = cap_xcomp->texture_size.y;
|
|
||||||
|
|
||||||
if(cap_xcomp->params.region_size.x > 0 && cap_xcomp->params.region_size.y > 0) {
|
|
||||||
video_codec_context->width = max_int(2, even_number_ceil(cap_xcomp->params.region_size.x));
|
|
||||||
video_codec_context->height = max_int(2, even_number_ceil(cap_xcomp->params.region_size.y));
|
|
||||||
}
|
|
||||||
|
|
||||||
frame->width = video_codec_context->width;
|
|
||||||
frame->height = video_codec_context->height;
|
|
||||||
|
|
||||||
if(!drm_create_codec_context(cap_xcomp, video_codec_context)) {
|
|
||||||
gsr_capture_xcomposite_vaapi_stop(cap, video_codec_context);
|
gsr_capture_xcomposite_vaapi_stop(cap, video_codec_context);
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
if(!gsr_capture_base_setup_vaapi_textures(&cap_xcomp->base, frame, cap_xcomp->params.egl, cap_xcomp->va_dpy, &cap_xcomp->prime, cap_xcomp->params.color_range)) {
|
if(!gsr_capture_base_setup_vaapi_textures(&cap_xcomp->xcomposite.base, frame, cap_xcomp->xcomposite.params.egl, cap_xcomp->va_dpy, &cap_xcomp->prime, cap_xcomp->xcomposite.params.color_range)) {
|
||||||
gsr_capture_xcomposite_vaapi_stop(cap, video_codec_context);
|
gsr_capture_xcomposite_vaapi_stop(cap, video_codec_context);
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
cap_xcomp->window_resize_timer = clock_get_monotonic_seconds();
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void gsr_capture_xcomposite_vaapi_tick(gsr_capture *cap, AVCodecContext *video_codec_context) {
|
static void gsr_capture_xcomposite_vaapi_tick(gsr_capture *cap, AVCodecContext *video_codec_context) {
|
||||||
gsr_capture_xcomposite_vaapi *cap_xcomp = cap->priv;
|
gsr_capture_xcomposite_vaapi *cap_xcomp = cap->priv;
|
||||||
|
gsr_capture_xcomposite_tick(&cap_xcomp->xcomposite, video_codec_context);
|
||||||
//cap_xcomp->params.egl->glClearColor(0.0f, 0.0f, 0.0f, 1.0f);
|
|
||||||
cap_xcomp->params.egl->glClear(0);
|
|
||||||
|
|
||||||
bool init_new_window = false;
|
|
||||||
while(XPending(cap_xcomp->params.egl->x11.dpy)) {
|
|
||||||
XNextEvent(cap_xcomp->params.egl->x11.dpy, &cap_xcomp->xev);
|
|
||||||
|
|
||||||
switch(cap_xcomp->xev.type) {
|
|
||||||
case DestroyNotify: {
|
|
||||||
/* Window died (when not following focused window), so we stop recording */
|
|
||||||
if(!cap_xcomp->params.follow_focused && cap_xcomp->xev.xdestroywindow.window == cap_xcomp->window) {
|
|
||||||
cap_xcomp->should_stop = true;
|
|
||||||
cap_xcomp->stop_is_error = false;
|
|
||||||
}
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
case Expose: {
|
|
||||||
/* Requires window texture recreate */
|
|
||||||
if(cap_xcomp->xev.xexpose.count == 0 && cap_xcomp->xev.xexpose.window == cap_xcomp->window) {
|
|
||||||
cap_xcomp->window_resize_timer = clock_get_monotonic_seconds();
|
|
||||||
cap_xcomp->window_resized = true;
|
|
||||||
}
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
case ConfigureNotify: {
|
|
||||||
/* Window resized */
|
|
||||||
if(cap_xcomp->xev.xconfigure.window == cap_xcomp->window && (cap_xcomp->xev.xconfigure.width != cap_xcomp->window_size.x || cap_xcomp->xev.xconfigure.height != cap_xcomp->window_size.y)) {
|
|
||||||
cap_xcomp->window_size.x = max_int(cap_xcomp->xev.xconfigure.width, 0);
|
|
||||||
cap_xcomp->window_size.y = max_int(cap_xcomp->xev.xconfigure.height, 0);
|
|
||||||
cap_xcomp->window_resize_timer = clock_get_monotonic_seconds();
|
|
||||||
cap_xcomp->window_resized = true;
|
|
||||||
}
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
case PropertyNotify: {
|
|
||||||
/* Focused window changed */
|
|
||||||
if(cap_xcomp->params.follow_focused && cap_xcomp->xev.xproperty.atom == cap_xcomp->net_active_window_atom) {
|
|
||||||
init_new_window = true;
|
|
||||||
}
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if(cap_xcomp->params.follow_focused && !cap_xcomp->follow_focused_initialized) {
|
|
||||||
init_new_window = true;
|
|
||||||
}
|
|
||||||
|
|
||||||
if(init_new_window) {
|
|
||||||
Window focused_window = get_focused_window(cap_xcomp->params.egl->x11.dpy, cap_xcomp->net_active_window_atom);
|
|
||||||
if(focused_window != cap_xcomp->window || !cap_xcomp->follow_focused_initialized) {
|
|
||||||
cap_xcomp->follow_focused_initialized = true;
|
|
||||||
XSelectInput(cap_xcomp->params.egl->x11.dpy, cap_xcomp->window, 0);
|
|
||||||
cap_xcomp->window = focused_window;
|
|
||||||
XSelectInput(cap_xcomp->params.egl->x11.dpy, cap_xcomp->window, StructureNotifyMask | ExposureMask);
|
|
||||||
|
|
||||||
XWindowAttributes attr;
|
|
||||||
attr.width = 0;
|
|
||||||
attr.height = 0;
|
|
||||||
if(!XGetWindowAttributes(cap_xcomp->params.egl->x11.dpy, cap_xcomp->window, &attr))
|
|
||||||
fprintf(stderr, "gsr error: gsr_capture_xcomposite_vaapi_tick failed: invalid window id: %lu\n", cap_xcomp->window);
|
|
||||||
|
|
||||||
cap_xcomp->window_size.x = max_int(attr.width, 0);
|
|
||||||
cap_xcomp->window_size.y = max_int(attr.height, 0);
|
|
||||||
cap_xcomp->window_resized = true;
|
|
||||||
|
|
||||||
window_texture_deinit(&cap_xcomp->window_texture);
|
|
||||||
window_texture_init(&cap_xcomp->window_texture, cap_xcomp->params.egl->x11.dpy, cap_xcomp->window, cap_xcomp->params.egl); // TODO: Do not do the below window_texture_on_resize after this
|
|
||||||
|
|
||||||
cap_xcomp->texture_size.x = 0;
|
|
||||||
cap_xcomp->texture_size.y = 0;
|
|
||||||
|
|
||||||
cap_xcomp->params.egl->glBindTexture(GL_TEXTURE_2D, window_texture_get_opengl_texture_id(&cap_xcomp->window_texture));
|
|
||||||
cap_xcomp->params.egl->glGetTexLevelParameteriv(GL_TEXTURE_2D, 0, GL_TEXTURE_WIDTH, &cap_xcomp->texture_size.x);
|
|
||||||
cap_xcomp->params.egl->glGetTexLevelParameteriv(GL_TEXTURE_2D, 0, GL_TEXTURE_HEIGHT, &cap_xcomp->texture_size.y);
|
|
||||||
cap_xcomp->params.egl->glBindTexture(GL_TEXTURE_2D, 0);
|
|
||||||
|
|
||||||
cap_xcomp->texture_size.x = min_int(video_codec_context->width, max_int(2, even_number_ceil(cap_xcomp->texture_size.x)));
|
|
||||||
cap_xcomp->texture_size.y = min_int(video_codec_context->height, max_int(2, even_number_ceil(cap_xcomp->texture_size.y)));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
const double window_resize_timeout = 1.0; // 1 second
|
|
||||||
if(cap_xcomp->window_resized && clock_get_monotonic_seconds() - cap_xcomp->window_resize_timer >= window_resize_timeout) {
|
|
||||||
cap_xcomp->window_resized = false;
|
|
||||||
|
|
||||||
if(window_texture_on_resize(&cap_xcomp->window_texture) != 0) {
|
|
||||||
fprintf(stderr, "gsr error: gsr_capture_xcomposite_vaapi_tick: window_texture_on_resize failed\n");
|
|
||||||
//cap_xcomp->should_stop = true;
|
|
||||||
//cap_xcomp->stop_is_error = true;
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
cap_xcomp->texture_size.x = 0;
|
|
||||||
cap_xcomp->texture_size.y = 0;
|
|
||||||
|
|
||||||
cap_xcomp->params.egl->glBindTexture(GL_TEXTURE_2D, window_texture_get_opengl_texture_id(&cap_xcomp->window_texture));
|
|
||||||
cap_xcomp->params.egl->glGetTexLevelParameteriv(GL_TEXTURE_2D, 0, GL_TEXTURE_WIDTH, &cap_xcomp->texture_size.x);
|
|
||||||
cap_xcomp->params.egl->glGetTexLevelParameteriv(GL_TEXTURE_2D, 0, GL_TEXTURE_HEIGHT, &cap_xcomp->texture_size.y);
|
|
||||||
cap_xcomp->params.egl->glBindTexture(GL_TEXTURE_2D, 0);
|
|
||||||
|
|
||||||
cap_xcomp->texture_size.x = min_int(video_codec_context->width, max_int(2, even_number_ceil(cap_xcomp->texture_size.x)));
|
|
||||||
cap_xcomp->texture_size.y = min_int(video_codec_context->height, max_int(2, even_number_ceil(cap_xcomp->texture_size.y)));
|
|
||||||
|
|
||||||
gsr_color_conversion_clear(&cap_xcomp->base.color_conversion);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static bool gsr_capture_xcomposite_vaapi_should_stop(gsr_capture *cap, bool *err) {
|
static bool gsr_capture_xcomposite_vaapi_should_stop(gsr_capture *cap, bool *err) {
|
||||||
gsr_capture_xcomposite_vaapi *cap_xcomp = cap->priv;
|
gsr_capture_xcomposite_vaapi *cap_xcomp = cap->priv;
|
||||||
if(cap_xcomp->should_stop) {
|
return gsr_capture_xcomposite_should_stop(&cap_xcomp->xcomposite, err);
|
||||||
if(err)
|
|
||||||
*err = cap_xcomp->stop_is_error;
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
if(err)
|
|
||||||
*err = false;
|
|
||||||
return false;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static int gsr_capture_xcomposite_vaapi_capture(gsr_capture *cap, AVFrame *frame) {
|
static int gsr_capture_xcomposite_vaapi_capture(gsr_capture *cap, AVFrame *frame) {
|
||||||
(void)frame;
|
|
||||||
gsr_capture_xcomposite_vaapi *cap_xcomp = cap->priv;
|
gsr_capture_xcomposite_vaapi *cap_xcomp = cap->priv;
|
||||||
|
return gsr_capture_xcomposite_capture(&cap_xcomp->xcomposite, frame);
|
||||||
const int target_x = max_int(0, frame->width / 2 - cap_xcomp->texture_size.x / 2);
|
|
||||||
const int target_y = max_int(0, frame->height / 2 - cap_xcomp->texture_size.y / 2);
|
|
||||||
|
|
||||||
gsr_color_conversion_draw(&cap_xcomp->base.color_conversion, window_texture_get_opengl_texture_id(&cap_xcomp->window_texture),
|
|
||||||
(vec2i){target_x, target_y}, cap_xcomp->texture_size,
|
|
||||||
(vec2i){0, 0}, cap_xcomp->texture_size,
|
|
||||||
0.0f, false);
|
|
||||||
|
|
||||||
cap_xcomp->params.egl->eglSwapBuffers(cap_xcomp->params.egl->egl_display, cap_xcomp->params.egl->egl_surface);
|
|
||||||
//cap_xcomp->params.egl->glFlush();
|
|
||||||
//cap_xcomp->params.egl->glFinish();
|
|
||||||
|
|
||||||
return 0;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static void gsr_capture_xcomposite_vaapi_stop(gsr_capture *cap, AVCodecContext *video_codec_context) {
|
static void gsr_capture_xcomposite_vaapi_stop(gsr_capture *cap, AVCodecContext *video_codec_context) {
|
||||||
@@ -348,14 +62,7 @@ static void gsr_capture_xcomposite_vaapi_stop(gsr_capture *cap, AVCodecContext *
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
window_texture_deinit(&cap_xcomp->window_texture);
|
gsr_capture_xcomposite_stop(&cap_xcomp->xcomposite, video_codec_context);
|
||||||
|
|
||||||
if(video_codec_context->hw_device_ctx)
|
|
||||||
av_buffer_unref(&video_codec_context->hw_device_ctx);
|
|
||||||
if(video_codec_context->hw_frames_ctx)
|
|
||||||
av_buffer_unref(&video_codec_context->hw_frames_ctx);
|
|
||||||
|
|
||||||
gsr_capture_base_stop(&cap_xcomp->base, cap_xcomp->params.egl);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static void gsr_capture_xcomposite_vaapi_destroy(gsr_capture *cap, AVCodecContext *video_codec_context) {
|
static void gsr_capture_xcomposite_vaapi_destroy(gsr_capture *cap, AVCodecContext *video_codec_context) {
|
||||||
@@ -384,7 +91,7 @@ gsr_capture* gsr_capture_xcomposite_vaapi_create(const gsr_capture_xcomposite_va
|
|||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
cap_xcomp->params = *params;
|
gsr_capture_xcomposite_init(&cap_xcomp->xcomposite, ¶ms->base);
|
||||||
|
|
||||||
*cap = (gsr_capture) {
|
*cap = (gsr_capture) {
|
||||||
.start = gsr_capture_xcomposite_vaapi_start,
|
.start = gsr_capture_xcomposite_vaapi_start,
|
||||||
|
|||||||
19
src/main.cpp
19
src/main.cpp
@@ -838,6 +838,7 @@ static void usage_full() {
|
|||||||
fprintf(stderr, " Forcefully set to 'h264' if the file container type is 'flv'.\n");
|
fprintf(stderr, " Forcefully set to 'h264' if the file container type is 'flv'.\n");
|
||||||
fprintf(stderr, " Forcefully set to 'hevc' on AMD/intel if video codec is 'h264' and if the file container type is 'mkv'.\n");
|
fprintf(stderr, " Forcefully set to 'hevc' on AMD/intel if video codec is 'h264' and if the file container type is 'mkv'.\n");
|
||||||
fprintf(stderr, " 'hevc_hdr' and 'av1_hdr' option is not available on X11.\n");
|
fprintf(stderr, " 'hevc_hdr' and 'av1_hdr' option is not available on X11.\n");
|
||||||
|
fprintf(stderr, " Note: hdr metadata is not included in the video when recording with 'hevc_hdr'/'av1_hdr' because of bugs in AMD, Intel and NVIDIA drivers (amazin', they are bugged).\n");
|
||||||
fprintf(stderr, "\n");
|
fprintf(stderr, "\n");
|
||||||
fprintf(stderr, " -ac Audio codec to use. Should be either 'aac', 'opus' or 'flac'. Defaults to 'opus' for .mp4/.mkv files, otherwise defaults to 'aac'.\n");
|
fprintf(stderr, " -ac Audio codec to use. Should be either 'aac', 'opus' or 'flac'. Defaults to 'opus' for .mp4/.mkv files, otherwise defaults to 'aac'.\n");
|
||||||
fprintf(stderr, " 'opus' and 'flac' is only supported by .mp4/.mkv files. 'opus' is recommended for best performance and smallest audio size.\n");
|
fprintf(stderr, " 'opus' and 'flac' is only supported by .mp4/.mkv files. 'opus' is recommended for best performance and smallest audio size.\n");
|
||||||
@@ -1537,11 +1538,11 @@ static gsr_capture* create_capture_impl(const char *window_str, const char *scre
|
|||||||
case GSR_GPU_VENDOR_AMD:
|
case GSR_GPU_VENDOR_AMD:
|
||||||
case GSR_GPU_VENDOR_INTEL: {
|
case GSR_GPU_VENDOR_INTEL: {
|
||||||
gsr_capture_xcomposite_vaapi_params xcomposite_params;
|
gsr_capture_xcomposite_vaapi_params xcomposite_params;
|
||||||
xcomposite_params.egl = &egl;
|
xcomposite_params.base.egl = &egl;
|
||||||
xcomposite_params.window = src_window_id;
|
xcomposite_params.base.window = src_window_id;
|
||||||
xcomposite_params.follow_focused = follow_focused;
|
xcomposite_params.base.follow_focused = follow_focused;
|
||||||
xcomposite_params.region_size = region_size;
|
xcomposite_params.base.region_size = region_size;
|
||||||
xcomposite_params.color_range = color_range;
|
xcomposite_params.base.color_range = color_range;
|
||||||
capture = gsr_capture_xcomposite_vaapi_create(&xcomposite_params);
|
capture = gsr_capture_xcomposite_vaapi_create(&xcomposite_params);
|
||||||
if(!capture)
|
if(!capture)
|
||||||
_exit(1);
|
_exit(1);
|
||||||
@@ -1549,10 +1550,10 @@ static gsr_capture* create_capture_impl(const char *window_str, const char *scre
|
|||||||
}
|
}
|
||||||
case GSR_GPU_VENDOR_NVIDIA: {
|
case GSR_GPU_VENDOR_NVIDIA: {
|
||||||
gsr_capture_xcomposite_cuda_params xcomposite_params;
|
gsr_capture_xcomposite_cuda_params xcomposite_params;
|
||||||
xcomposite_params.egl = &egl;
|
xcomposite_params.base.egl = &egl;
|
||||||
xcomposite_params.window = src_window_id;
|
xcomposite_params.base.window = src_window_id;
|
||||||
xcomposite_params.follow_focused = follow_focused;
|
xcomposite_params.base.follow_focused = follow_focused;
|
||||||
xcomposite_params.region_size = region_size;
|
xcomposite_params.base.region_size = region_size;
|
||||||
xcomposite_params.overclock = overclock;
|
xcomposite_params.overclock = overclock;
|
||||||
capture = gsr_capture_xcomposite_cuda_create(&xcomposite_params);
|
capture = gsr_capture_xcomposite_cuda_create(&xcomposite_params);
|
||||||
if(!capture)
|
if(!capture)
|
||||||
|
|||||||
17
src/utils.c
17
src/utils.c
@@ -9,13 +9,6 @@
|
|||||||
#include <stdlib.h>
|
#include <stdlib.h>
|
||||||
#include <X11/Xatom.h>
|
#include <X11/Xatom.h>
|
||||||
|
|
||||||
typedef enum {
|
|
||||||
X11_ROT_0 = 1 << 0,
|
|
||||||
X11_ROT_90 = 1 << 1,
|
|
||||||
X11_ROT_180 = 1 << 2,
|
|
||||||
X11_ROT_270 = 1 << 3
|
|
||||||
} X11Rotation;
|
|
||||||
|
|
||||||
double clock_get_monotonic_seconds(void) {
|
double clock_get_monotonic_seconds(void) {
|
||||||
struct timespec ts;
|
struct timespec ts;
|
||||||
ts.tv_sec = 0;
|
ts.tv_sec = 0;
|
||||||
@@ -32,12 +25,12 @@ static const XRRModeInfo* get_mode_info(const XRRScreenResources *sr, RRMode id)
|
|||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
static gsr_monitor_rotation x11_rotation_to_gsr_rotation(X11Rotation rot) {
|
static gsr_monitor_rotation x11_rotation_to_gsr_rotation(int rot) {
|
||||||
switch(rot) {
|
switch(rot) {
|
||||||
case X11_ROT_0: return GSR_MONITOR_ROT_0;
|
case RR_Rotate_0: return GSR_MONITOR_ROT_0;
|
||||||
case X11_ROT_90: return GSR_MONITOR_ROT_90;
|
case RR_Rotate_90: return GSR_MONITOR_ROT_90;
|
||||||
case X11_ROT_180: return GSR_MONITOR_ROT_180;
|
case RR_Rotate_180: return GSR_MONITOR_ROT_180;
|
||||||
case X11_ROT_270: return GSR_MONITOR_ROT_270;
|
case RR_Rotate_270: return GSR_MONITOR_ROT_270;
|
||||||
}
|
}
|
||||||
return GSR_MONITOR_ROT_0;
|
return GSR_MONITOR_ROT_0;
|
||||||
}
|
}
|
||||||
|
|||||||
Reference in New Issue
Block a user