mirror of
https://repo.dec05eba.com/gpu-screen-recorder
synced 2026-03-31 09:07:13 +09:00
Support taking a screenshot (-o image.jpg/image.png)
This commit is contained in:
@@ -7,6 +7,8 @@ similar to shadowplay on windows. This is the fastest screen recording tool for
|
||||
This screen recorder can be used for recording your desktop offline, for live streaming and for nvidia shadowplay-like instant replay,
|
||||
where only the last few minutes are saved.
|
||||
|
||||
This software can also take screenshots.
|
||||
|
||||
This is a cli-only tool, if you want an UI for this check out [GPU Screen Recorder GTK](https://git.dec05eba.com/gpu-screen-recorder-gtk/) or if you prefer a ShadowPlay-like UI then check out [GPU Screen Recorder UI](https://git.dec05eba.com/gpu-screen-recorder-ui/).
|
||||
|
||||
Supported video codecs:
|
||||
@@ -20,7 +22,10 @@ Supported audio codecs:
|
||||
* Opus (default)
|
||||
* AAC
|
||||
|
||||
## Note
|
||||
Supported image formats:
|
||||
* JPEG
|
||||
* PNG
|
||||
|
||||
This software works on X11 and Wayland on AMD, Intel and NVIDIA.
|
||||
### TEMPORARY ISSUES
|
||||
1) Videos are in variable framerate format. Use MPV to play such videos, otherwise you might experience stuttering in the video if you are using a buggy video player. You can try saving the video into a .mkv file instead as some software may have better support for .mkv files (such as kdenlive). You can use the "-fm cfr" option to to use constant framerate mode.
|
||||
|
||||
2
TODO
2
TODO
@@ -228,3 +228,5 @@ Make it possible to save a shorter replay clip remotely. Maybe implement ipc fir
|
||||
Add an option to pass http headers when streaming. Some streaming services require streaming keys to be passed in a http header instead of in the url as a parameter.
|
||||
|
||||
When adding vulkan video support add VK_VIDEO_ENCODE_TUNING_MODE_LOW_LATENCY_KHR.
|
||||
|
||||
Implement screenshot without invoking opengl (which is slow to start on some systems).
|
||||
|
||||
@@ -14,7 +14,6 @@ typedef struct {
|
||||
gsr_color_depth color_depth;
|
||||
gsr_color_range color_range;
|
||||
bool record_cursor;
|
||||
bool use_software_video_encoder;
|
||||
vec2i output_resolution;
|
||||
} gsr_capture_nvfbc_params;
|
||||
|
||||
|
||||
@@ -22,7 +22,8 @@ typedef enum {
|
||||
|
||||
typedef enum {
|
||||
GSR_DESTINATION_COLOR_NV12, /* YUV420, BT709, 8-bit */
|
||||
GSR_DESTINATION_COLOR_P010 /* YUV420, BT2020, 10-bit */
|
||||
GSR_DESTINATION_COLOR_P010, /* YUV420, BT2020, 10-bit */
|
||||
GSR_DESTINATION_COLOR_RGB8
|
||||
} gsr_destination_color;
|
||||
|
||||
typedef struct {
|
||||
|
||||
@@ -104,11 +104,13 @@ typedef void(*__GLXextFuncPtr)(void);
|
||||
#define GL_RG 0x8227
|
||||
#define GL_RGB 0x1907
|
||||
#define GL_RGBA 0x1908
|
||||
#define GL_RGB8 0x8051
|
||||
#define GL_RGBA8 0x8058
|
||||
#define GL_R8 0x8229
|
||||
#define GL_RG8 0x822B
|
||||
#define GL_R16 0x822A
|
||||
#define GL_RG16 0x822C
|
||||
#define GL_RGB16 0x8054
|
||||
#define GL_UNSIGNED_BYTE 0x1401
|
||||
#define GL_COLOR_BUFFER_BIT 0x00004000
|
||||
#define GL_TEXTURE_WRAP_S 0x2802
|
||||
|
||||
15
include/encoder/video/image.h
Normal file
15
include/encoder/video/image.h
Normal file
@@ -0,0 +1,15 @@
|
||||
#ifndef GSR_ENCODER_VIDEO_IMAGE_H
|
||||
#define GSR_ENCODER_VIDEO_IMAGE_H
|
||||
|
||||
#include "video.h"
|
||||
|
||||
typedef struct gsr_egl gsr_egl;
|
||||
|
||||
typedef struct {
|
||||
gsr_egl *egl;
|
||||
gsr_color_depth color_depth;
|
||||
} gsr_video_encoder_image_params;
|
||||
|
||||
gsr_video_encoder* gsr_video_encoder_image_create(const gsr_video_encoder_image_params *params);
|
||||
|
||||
#endif /* GSR_ENCODER_VIDEO_IMAGE_H */
|
||||
@@ -18,6 +18,7 @@ src = [
|
||||
'src/encoder/video/vaapi.c',
|
||||
'src/encoder/video/vulkan.c',
|
||||
'src/encoder/video/software.c',
|
||||
'src/encoder/video/image.c',
|
||||
'src/codec_query/nvenc.c',
|
||||
'src/codec_query/vaapi.c',
|
||||
'src/codec_query/vulkan.c',
|
||||
|
||||
@@ -9,6 +9,7 @@
|
||||
|
||||
#define MAX_SHADERS 4
|
||||
#define MAX_FRAMEBUFFERS 2
|
||||
#define EXTERNAL_TEXTURE_SHADER_OFFSET 2
|
||||
|
||||
static float abs_f(float v) {
|
||||
return v >= 0.0f ? v : -v;
|
||||
@@ -69,6 +70,8 @@ static const char* color_format_range_get_transform_matrix(gsr_destination_color
|
||||
}
|
||||
break;
|
||||
}
|
||||
case GSR_DESTINATION_COLOR_RGB8:
|
||||
return "";
|
||||
default:
|
||||
return NULL;
|
||||
}
|
||||
@@ -201,6 +204,65 @@ static unsigned int load_shader_uv(gsr_shader *shader, gsr_egl *egl, gsr_color_u
|
||||
return 0;
|
||||
}
|
||||
|
||||
static unsigned int load_shader_rgb(gsr_shader *shader, gsr_egl *egl, gsr_color_uniforms *uniforms, gsr_destination_color color_format, gsr_color_range color_range, bool external_texture) {
|
||||
// TODO: Support hdr
|
||||
char vertex_shader[2048];
|
||||
snprintf(vertex_shader, sizeof(vertex_shader),
|
||||
"#version 300 es \n"
|
||||
"in vec2 pos; \n"
|
||||
"in vec2 texcoords; \n"
|
||||
"out vec2 texcoords_out; \n"
|
||||
"uniform vec2 offset; \n"
|
||||
"uniform float rotation; \n"
|
||||
ROTATE_Z
|
||||
"void main() \n"
|
||||
"{ \n"
|
||||
" texcoords_out = (vec4(texcoords.x - 0.5, texcoords.y - 0.5, 0.0, 0.0) * rotate_z(rotation)).xy + vec2(0.5, 0.5); \n"
|
||||
" gl_Position = vec4(offset.x, offset.y, 0.0, 0.0) + vec4(pos.x, pos.y, 0.0, 1.0); \n"
|
||||
"} \n");
|
||||
|
||||
const char *main_code =
|
||||
main_code =
|
||||
" vec4 pixel = texture(tex1, texcoords_out); \n"
|
||||
" FragColor = pixel; \n";
|
||||
|
||||
char fragment_shader[2048];
|
||||
if(external_texture) {
|
||||
snprintf(fragment_shader, sizeof(fragment_shader),
|
||||
"#version 300 es \n"
|
||||
"#extension GL_OES_EGL_image_external : enable \n"
|
||||
"#extension GL_OES_EGL_image_external_essl3 : require \n"
|
||||
"precision mediump float; \n"
|
||||
"in vec2 texcoords_out; \n"
|
||||
"uniform samplerExternalOES tex1; \n"
|
||||
"out vec4 FragColor; \n"
|
||||
"void main() \n"
|
||||
"{ \n"
|
||||
"%s"
|
||||
"} \n", main_code);
|
||||
} else {
|
||||
snprintf(fragment_shader, sizeof(fragment_shader),
|
||||
"#version 300 es \n"
|
||||
"precision mediump float; \n"
|
||||
"in vec2 texcoords_out; \n"
|
||||
"uniform sampler2D tex1; \n"
|
||||
"out vec4 FragColor; \n"
|
||||
"void main() \n"
|
||||
"{ \n"
|
||||
"%s"
|
||||
"} \n", main_code);
|
||||
}
|
||||
|
||||
if(gsr_shader_init(shader, egl, vertex_shader, fragment_shader) != 0)
|
||||
return -1;
|
||||
|
||||
gsr_shader_bind_attribute_location(shader, "pos", 0);
|
||||
gsr_shader_bind_attribute_location(shader, "texcoords", 1);
|
||||
uniforms->offset = egl->glGetUniformLocation(shader->program_id, "offset");
|
||||
uniforms->rotation = egl->glGetUniformLocation(shader->program_id, "rotation");
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int load_framebuffers(gsr_color_conversion *self) {
|
||||
/* TODO: Only generate the necessary amount of framebuffers (self->params.num_destination_textures) */
|
||||
const unsigned int draw_buffer = GL_COLOR_ATTACHMENT0;
|
||||
@@ -276,18 +338,37 @@ int gsr_color_conversion_init(gsr_color_conversion *self, const gsr_color_conver
|
||||
}
|
||||
|
||||
if(self->params.load_external_image_shader) {
|
||||
if(load_shader_y(&self->shaders[2], self->params.egl, &self->uniforms[2], params->destination_color, params->color_range, true) != 0) {
|
||||
if(load_shader_y(&self->shaders[EXTERNAL_TEXTURE_SHADER_OFFSET], self->params.egl, &self->uniforms[EXTERNAL_TEXTURE_SHADER_OFFSET], params->destination_color, params->color_range, true) != 0) {
|
||||
fprintf(stderr, "gsr error: gsr_color_conversion_init: failed to load Y shader\n");
|
||||
goto err;
|
||||
}
|
||||
|
||||
if(load_shader_uv(&self->shaders[3], self->params.egl, &self->uniforms[3], params->destination_color, params->color_range, true) != 0) {
|
||||
if(load_shader_uv(&self->shaders[EXTERNAL_TEXTURE_SHADER_OFFSET + 1], self->params.egl, &self->uniforms[EXTERNAL_TEXTURE_SHADER_OFFSET + 1], params->destination_color, params->color_range, true) != 0) {
|
||||
fprintf(stderr, "gsr error: gsr_color_conversion_init: failed to load UV shader\n");
|
||||
goto err;
|
||||
}
|
||||
}
|
||||
break;
|
||||
}
|
||||
case GSR_DESTINATION_COLOR_RGB8: {
|
||||
if(self->params.num_destination_textures != 1) {
|
||||
fprintf(stderr, "gsr error: gsr_color_conversion_init: expected 1 destination textures for destination color RGB8, got %d destination texture(s)\n", self->params.num_destination_textures);
|
||||
return -1;
|
||||
}
|
||||
|
||||
if(load_shader_rgb(&self->shaders[0], self->params.egl, &self->uniforms[0], params->destination_color, params->color_range, false) != 0) {
|
||||
fprintf(stderr, "gsr error: gsr_color_conversion_init: failed to load Y shader\n");
|
||||
goto err;
|
||||
}
|
||||
|
||||
if(self->params.load_external_image_shader) {
|
||||
if(load_shader_rgb(&self->shaders[EXTERNAL_TEXTURE_SHADER_OFFSET], self->params.egl, &self->uniforms[EXTERNAL_TEXTURE_SHADER_OFFSET], params->destination_color, params->color_range, true) != 0) {
|
||||
fprintf(stderr, "gsr error: gsr_color_conversion_init: failed to load Y shader\n");
|
||||
goto err;
|
||||
}
|
||||
}
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
if(load_framebuffers(self) != 0)
|
||||
@@ -419,7 +500,7 @@ void gsr_color_conversion_draw(gsr_color_conversion *self, unsigned int texture_
|
||||
self->params.egl->glBindFramebuffer(GL_FRAMEBUFFER, self->framebuffers[0]);
|
||||
//cap_xcomp->params.egl->glClear(GL_COLOR_BUFFER_BIT); // TODO: Do this in a separate clear_ function. We want to do that when using multiple drm to create the final image (multiple monitors for example)
|
||||
|
||||
const int shader_index = external_texture ? 2 : 0;
|
||||
const int shader_index = external_texture ? EXTERNAL_TEXTURE_SHADER_OFFSET : 0;
|
||||
gsr_shader_use(&self->shaders[shader_index]);
|
||||
self->params.egl->glUniform1f(self->uniforms[shader_index].rotation, rotation);
|
||||
self->params.egl->glUniform2f(self->uniforms[shader_index].offset, pos_norm.x, pos_norm.y);
|
||||
@@ -430,7 +511,7 @@ void gsr_color_conversion_draw(gsr_color_conversion *self, unsigned int texture_
|
||||
self->params.egl->glBindFramebuffer(GL_FRAMEBUFFER, self->framebuffers[1]);
|
||||
//cap_xcomp->params.egl->glClear(GL_COLOR_BUFFER_BIT);
|
||||
|
||||
const int shader_index = external_texture ? 3 : 1;
|
||||
const int shader_index = external_texture ? EXTERNAL_TEXTURE_SHADER_OFFSET + 1 : 1;
|
||||
gsr_shader_use(&self->shaders[shader_index]);
|
||||
self->params.egl->glUniform1f(self->uniforms[shader_index].rotation, rotation);
|
||||
self->params.egl->glUniform2f(self->uniforms[shader_index].offset, pos_norm.x, pos_norm.y);
|
||||
@@ -458,6 +539,13 @@ void gsr_color_conversion_clear(gsr_color_conversion *self) {
|
||||
color2[3] = 1.0f;
|
||||
break;
|
||||
}
|
||||
case GSR_DESTINATION_COLOR_RGB8: {
|
||||
color2[0] = 0.0f;
|
||||
color2[1] = 0.0f;
|
||||
color2[2] = 0.0f;
|
||||
color2[3] = 1.0f;
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
self->params.egl->glBindFramebuffer(GL_FRAMEBUFFER, self->framebuffers[0]);
|
||||
|
||||
130
src/encoder/video/image.c
Normal file
130
src/encoder/video/image.c
Normal file
@@ -0,0 +1,130 @@
|
||||
#include "../../../include/encoder/video/image.h"
|
||||
#include "../../../include/egl.h"
|
||||
|
||||
#include <libavcodec/avcodec.h>
|
||||
#include <libavutil/frame.h>
|
||||
|
||||
#include <stdlib.h>
|
||||
|
||||
#define LINESIZE_ALIGNMENT 4
|
||||
|
||||
typedef struct {
|
||||
gsr_video_encoder_image_params params;
|
||||
|
||||
unsigned int target_texture;
|
||||
} gsr_video_encoder_image;
|
||||
|
||||
static unsigned int gl_create_texture(gsr_egl *egl, int width, int height, int internal_format, unsigned int format) {
|
||||
unsigned int texture_id = 0;
|
||||
egl->glGenTextures(1, &texture_id);
|
||||
egl->glBindTexture(GL_TEXTURE_2D, texture_id);
|
||||
egl->glTexImage2D(GL_TEXTURE_2D, 0, internal_format, width, height, 0, format, GL_UNSIGNED_BYTE, NULL);
|
||||
|
||||
egl->glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_S, GL_CLAMP_TO_EDGE);
|
||||
egl->glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_T, GL_CLAMP_TO_EDGE);
|
||||
egl->glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MAG_FILTER, GL_NEAREST);
|
||||
egl->glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_NEAREST);
|
||||
|
||||
egl->glBindTexture(GL_TEXTURE_2D, 0);
|
||||
return texture_id;
|
||||
}
|
||||
|
||||
static bool gsr_video_encoder_image_setup_textures(gsr_video_encoder_image *self, AVCodecContext *video_codec_context, AVFrame *frame) {
|
||||
int res = av_frame_get_buffer(frame, LINESIZE_ALIGNMENT);
|
||||
if(res < 0) {
|
||||
fprintf(stderr, "gsr error: gsr_video_encoder_image_setup_textures: av_frame_get_buffer failed: %d\n", res);
|
||||
return false;
|
||||
}
|
||||
|
||||
res = av_frame_make_writable(frame);
|
||||
if(res < 0) {
|
||||
fprintf(stderr, "gsr error: gsr_video_encoder_image_setup_textures: av_frame_make_writable failed: %d\n", res);
|
||||
return false;
|
||||
}
|
||||
|
||||
self->target_texture = gl_create_texture(self->params.egl, video_codec_context->width, video_codec_context->height, self->params.color_depth == GSR_COLOR_DEPTH_8_BITS ? GL_RGB8 : GL_RGB16, GL_RGB);
|
||||
if(self->target_texture == 0) {
|
||||
fprintf(stderr, "gsr error: gsr_capture_kms_setup_cuda_textures: failed to create opengl texture\n");
|
||||
return false;
|
||||
}
|
||||
|
||||
return true;
|
||||
}
|
||||
|
||||
static void gsr_video_encoder_image_stop(gsr_video_encoder_image *self, AVCodecContext *video_codec_context);
|
||||
|
||||
static bool gsr_video_encoder_image_start(gsr_video_encoder *encoder, AVCodecContext *video_codec_context, AVFrame *frame) {
|
||||
gsr_video_encoder_image *self = encoder->priv;
|
||||
|
||||
video_codec_context->width = FFALIGN(video_codec_context->width, LINESIZE_ALIGNMENT);
|
||||
video_codec_context->height = FFALIGN(video_codec_context->height, 2);
|
||||
|
||||
frame->width = video_codec_context->width;
|
||||
frame->height = video_codec_context->height;
|
||||
|
||||
if(!gsr_video_encoder_image_setup_textures(self, video_codec_context, frame)) {
|
||||
gsr_video_encoder_image_stop(self, video_codec_context);
|
||||
return false;
|
||||
}
|
||||
|
||||
return true;
|
||||
}
|
||||
|
||||
void gsr_video_encoder_image_stop(gsr_video_encoder_image *self, AVCodecContext *video_codec_context) {
|
||||
(void)video_codec_context;
|
||||
self->params.egl->glDeleteTextures(1, &self->target_texture);
|
||||
self->target_texture = 0;
|
||||
}
|
||||
|
||||
static void gsr_video_encoder_image_copy_textures_to_frame(gsr_video_encoder *encoder, AVFrame *frame, gsr_color_conversion *color_conversion) {
|
||||
gsr_video_encoder_image *self = encoder->priv;
|
||||
// TODO: hdr support
|
||||
self->params.egl->glBindTexture(GL_TEXTURE_2D, self->target_texture);
|
||||
// We could use glGetTexSubImage and then we wouldn't have to use a specific linesize (LINESIZE_ALIGNMENT) that adds padding,
|
||||
// but glGetTexSubImage is only available starting from opengl 4.5.
|
||||
self->params.egl->glGetTexImage(GL_TEXTURE_2D, 0, GL_RGB, GL_UNSIGNED_BYTE, frame->data[0]);
|
||||
self->params.egl->glBindTexture(GL_TEXTURE_2D, 0);
|
||||
// cap_kms->kms.base.egl->eglSwapBuffers(cap_kms->kms.base.egl->egl_display, cap_kms->kms.base.egl->egl_surface);
|
||||
|
||||
self->params.egl->glFlush();
|
||||
self->params.egl->glFinish();
|
||||
}
|
||||
|
||||
static void gsr_video_encoder_image_get_textures(gsr_video_encoder *encoder, unsigned int *textures, int *num_textures, gsr_destination_color *destination_color) {
|
||||
gsr_video_encoder_image *self = encoder->priv;
|
||||
textures[0] = self->target_texture;
|
||||
*num_textures = 1;
|
||||
// TODO: 10-bit support
|
||||
//*destination_color = self->params.color_depth == GSR_COLOR_DEPTH_10_BITS ? GSR_DESTINATION_COLOR_P010 : GSR_DESTINATION_COLOR_NV12;
|
||||
*destination_color = GSR_DESTINATION_COLOR_RGB8;
|
||||
}
|
||||
|
||||
static void gsr_video_encoder_image_destroy(gsr_video_encoder *encoder, AVCodecContext *video_codec_context) {
|
||||
gsr_video_encoder_image_stop(encoder->priv, video_codec_context);
|
||||
free(encoder->priv);
|
||||
free(encoder);
|
||||
}
|
||||
|
||||
gsr_video_encoder* gsr_video_encoder_image_create(const gsr_video_encoder_image_params *params) {
|
||||
gsr_video_encoder *encoder = calloc(1, sizeof(gsr_video_encoder));
|
||||
if(!encoder)
|
||||
return NULL;
|
||||
|
||||
gsr_video_encoder_image *encoder_image = calloc(1, sizeof(gsr_video_encoder_image));
|
||||
if(!encoder_image) {
|
||||
free(encoder);
|
||||
return NULL;
|
||||
}
|
||||
|
||||
encoder_image->params = *params;
|
||||
|
||||
*encoder = (gsr_video_encoder) {
|
||||
.start = gsr_video_encoder_image_start,
|
||||
.copy_textures_to_frame = gsr_video_encoder_image_copy_textures_to_frame,
|
||||
.get_textures = gsr_video_encoder_image_get_textures,
|
||||
.destroy = gsr_video_encoder_image_destroy,
|
||||
.priv = encoder_image
|
||||
};
|
||||
|
||||
return encoder;
|
||||
}
|
||||
135
src/main.cpp
135
src/main.cpp
@@ -13,6 +13,7 @@ extern "C" {
|
||||
#include "../include/encoder/video/vaapi.h"
|
||||
#include "../include/encoder/video/vulkan.h"
|
||||
#include "../include/encoder/video/software.h"
|
||||
#include "../include/encoder/video/image.h"
|
||||
#include "../include/codec_query/nvenc.h"
|
||||
#include "../include/codec_query/vaapi.h"
|
||||
#include "../include/codec_query/vulkan.h"
|
||||
@@ -112,7 +113,9 @@ enum class VideoCodec {
|
||||
VP8,
|
||||
VP9,
|
||||
H264_VULKAN,
|
||||
HEVC_VULKAN
|
||||
HEVC_VULKAN,
|
||||
JPEG,
|
||||
PNG
|
||||
};
|
||||
|
||||
enum class AudioCodec {
|
||||
@@ -216,6 +219,16 @@ static bool video_codec_is_vulkan(VideoCodec video_codec) {
|
||||
}
|
||||
}
|
||||
|
||||
static bool video_codec_is_image(VideoCodec video_codec) {
|
||||
switch(video_codec) {
|
||||
case VideoCodec::JPEG:
|
||||
case VideoCodec::PNG:
|
||||
return true;
|
||||
default:
|
||||
return false;
|
||||
}
|
||||
}
|
||||
|
||||
struct PacketData {
|
||||
PacketData() {}
|
||||
PacketData(const PacketData&) = delete;
|
||||
@@ -580,7 +593,22 @@ static AVCodecContext *create_video_codec_context(AVPixelFormat pix_fmt,
|
||||
if (codec_context->codec_id == AV_CODEC_ID_MPEG1VIDEO)
|
||||
codec_context->mb_decision = 2;
|
||||
|
||||
if(!use_software_video_encoder && vendor != GSR_GPU_VENDOR_NVIDIA && bitrate_mode != BitrateMode::CBR) {
|
||||
if(video_codec_is_image(video_codec)) {
|
||||
switch(video_quality) {
|
||||
case VideoQuality::MEDIUM:
|
||||
codec_context->compression_level = 8;
|
||||
break;
|
||||
case VideoQuality::HIGH:
|
||||
codec_context->compression_level = 6;
|
||||
break;
|
||||
case VideoQuality::VERY_HIGH:
|
||||
codec_context->compression_level = 4;
|
||||
break;
|
||||
case VideoQuality::ULTRA:
|
||||
codec_context->compression_level = 2;
|
||||
break;
|
||||
}
|
||||
} else if(!use_software_video_encoder && vendor != GSR_GPU_VENDOR_NVIDIA && bitrate_mode != BitrateMode::CBR) {
|
||||
// 8 bit / 10 bit = 80%, and increase it even more
|
||||
const float quality_multiply = hdr ? (8.0f/10.0f * 0.7f) : 1.0f;
|
||||
if(codec_context->codec_id == AV_CODEC_ID_AV1 || codec_context->codec_id == AV_CODEC_ID_H264 || codec_context->codec_id == AV_CODEC_ID_HEVC) {
|
||||
@@ -715,6 +743,15 @@ static AVFrame* create_audio_frame(AVCodecContext *audio_codec_context) {
|
||||
return frame;
|
||||
}
|
||||
|
||||
static void open_video_image(AVCodecContext *codec_context) {
|
||||
AVDictionary *options = nullptr;
|
||||
int ret = avcodec_open2(codec_context, codec_context->codec, &options);
|
||||
if (ret < 0) {
|
||||
fprintf(stderr, "Error: Could not open video codec: %s\n", av_error_to_string(ret));
|
||||
_exit(1);
|
||||
}
|
||||
}
|
||||
|
||||
static void dict_set_profile(AVCodecContext *codec_context, gsr_gpu_vendor vendor, gsr_color_depth color_depth, AVDictionary **options) {
|
||||
#if LIBAVCODEC_VERSION_INT < AV_VERSION_INT(61, 17, 100)
|
||||
if(codec_context->codec_id == AV_CODEC_ID_H264) {
|
||||
@@ -1197,7 +1234,7 @@ static void usage_full() {
|
||||
printf("\n");
|
||||
printf(" --info\n");
|
||||
printf(" List info about the system. Lists the following information (prints them to stdout and exits):\n");
|
||||
printf(" Supported video codecs (h264, h264_software, hevc, hevc_hdr, hevc_10bit, av1, av1_hdr, av1_10bit, vp8, vp9 (if supported)).\n");
|
||||
printf(" Supported video codecs (h264, h264_software, hevc, hevc_hdr, hevc_10bit, av1, av1_hdr, av1_10bit, vp8, vp9) and image codecs (jpeg, png) (if supported).\n");
|
||||
printf(" Supported capture options (window, focused, screen, monitors and portal, if supported by the system).\n");
|
||||
printf(" If opengl initialization fails then the program exits with 22, if no usable drm device is found then it exits with 23. On success it exits with 0.\n");
|
||||
printf("\n");
|
||||
@@ -1257,6 +1294,7 @@ static void usage_full() {
|
||||
printf(" %s -w screen -f 60 -a \"app:firefox|app:csgo\" -o \"$HOME/Videos/video.mp4\"\n", program_name);
|
||||
printf(" %s -w screen -f 60 -a \"app-inverse:firefox|app-inverse:csgo\" -o \"$HOME/Videos/video.mp4\"\n", program_name);
|
||||
printf(" %s -w screen -f 60 -a \"default-input|app-inverse:Brave\" -o \"$HOME/Videos/video.mp4\"\n", program_name);
|
||||
printf(" %s -w screen -f 60 -o \"$HOME/Pictures/image.jpg\"\n", program_name);
|
||||
//fprintf(stderr, " gpu-screen-recorder -w screen -f 60 -q ultra -pixfmt yuv444 -o video.mp4\n");
|
||||
fflush(stdout);
|
||||
_exit(1);
|
||||
@@ -1645,6 +1683,12 @@ static bool string_starts_with(const std::string &str, const char *substr) {
|
||||
return (int)str.size() >= len && memcmp(str.data(), substr, len) == 0;
|
||||
}
|
||||
|
||||
static bool string_ends_with(const char *str, const char *substr) {
|
||||
int str_len = strlen(str);
|
||||
int substr_len = strlen(substr);
|
||||
return str_len >= substr_len && memcmp(str + str_len - substr_len, substr, substr_len) == 0;
|
||||
}
|
||||
|
||||
static const AudioDevice* get_audio_device_by_name(const std::vector<AudioDevice> &audio_devices, const char *name) {
|
||||
for(const auto &audio_device : audio_devices) {
|
||||
if(strcmp(audio_device.name.c_str(), name) == 0)
|
||||
@@ -1850,6 +1894,14 @@ fail:
|
||||
static gsr_video_encoder* create_video_encoder(gsr_egl *egl, bool overclock, gsr_color_depth color_depth, bool use_software_video_encoder, VideoCodec video_codec) {
|
||||
gsr_video_encoder *video_encoder = nullptr;
|
||||
|
||||
if(video_codec_is_image(video_codec)) {
|
||||
gsr_video_encoder_image_params params;
|
||||
params.egl = egl;
|
||||
params.color_depth = color_depth;
|
||||
video_encoder = gsr_video_encoder_image_create(¶ms);
|
||||
return video_encoder;
|
||||
}
|
||||
|
||||
if(use_software_video_encoder) {
|
||||
gsr_video_encoder_software_params params;
|
||||
params.egl = egl;
|
||||
@@ -1999,6 +2051,10 @@ static const AVCodec* get_ffmpeg_video_codec(VideoCodec video_codec, gsr_gpu_ven
|
||||
return avcodec_find_encoder_by_name("h264_vulkan");
|
||||
case VideoCodec::HEVC_VULKAN:
|
||||
return avcodec_find_encoder_by_name("hevc_vulkan");
|
||||
case VideoCodec::JPEG:
|
||||
return avcodec_find_encoder_by_name("libopenjpeg");
|
||||
case VideoCodec::PNG:
|
||||
return avcodec_find_encoder_by_name("png");
|
||||
}
|
||||
return nullptr;
|
||||
}
|
||||
@@ -2069,6 +2125,10 @@ static void list_supported_video_codecs(gsr_egl *egl, bool wayland) {
|
||||
puts("vp8");
|
||||
if(supported_video_codecs.vp9.supported)
|
||||
puts("vp9");
|
||||
if(avcodec_find_encoder_by_name("libopenjpeg"))
|
||||
puts("jpeg");
|
||||
if(avcodec_find_encoder_by_name("png"))
|
||||
puts("png");
|
||||
//if(supported_video_codecs_vulkan.h264.supported)
|
||||
// puts("h264_vulkan");
|
||||
//if(supported_video_codecs_vulkan.hevc.supported)
|
||||
@@ -2344,7 +2404,7 @@ static void validate_monitor_get_valid(const gsr_egl *egl, std::string &window_s
|
||||
}
|
||||
|
||||
static gsr_capture* create_capture_impl(std::string &window_str, vec2i output_resolution, bool wayland, gsr_egl *egl, int fps, VideoCodec video_codec, gsr_color_range color_range,
|
||||
bool record_cursor, bool use_software_video_encoder, bool restore_portal_session, const char *portal_session_token_filepath,
|
||||
bool record_cursor, bool restore_portal_session, const char *portal_session_token_filepath,
|
||||
gsr_color_depth color_depth)
|
||||
{
|
||||
Window src_window_id = None;
|
||||
@@ -2406,7 +2466,6 @@ static gsr_capture* create_capture_impl(std::string &window_str, vec2i output_re
|
||||
nvfbc_params.color_depth = color_depth;
|
||||
nvfbc_params.color_range = color_range;
|
||||
nvfbc_params.record_cursor = record_cursor;
|
||||
nvfbc_params.use_software_video_encoder = use_software_video_encoder;
|
||||
nvfbc_params.output_resolution = output_resolution;
|
||||
capture = gsr_capture_nvfbc_create(&nvfbc_params);
|
||||
if(!capture)
|
||||
@@ -2457,7 +2516,10 @@ static gsr_capture* create_capture_impl(std::string &window_str, vec2i output_re
|
||||
}
|
||||
|
||||
static AVPixelFormat get_pixel_format(VideoCodec video_codec, gsr_gpu_vendor vendor, bool use_software_video_encoder) {
|
||||
if(use_software_video_encoder) {
|
||||
if(video_codec_is_image(video_codec)) {
|
||||
// TODO: hdr
|
||||
return AV_PIX_FMT_RGB24;
|
||||
} else if(use_software_video_encoder) {
|
||||
return AV_PIX_FMT_NV12;
|
||||
} else {
|
||||
if(video_codec_is_vulkan(video_codec))
|
||||
@@ -2681,6 +2743,8 @@ static const char* video_codec_to_string(VideoCodec video_codec) {
|
||||
case VideoCodec::VP9: return "vp9";
|
||||
case VideoCodec::H264_VULKAN: return "h264_vulkan";
|
||||
case VideoCodec::HEVC_VULKAN: return "hevc_vulkan";
|
||||
case VideoCodec::JPEG: return "jpeg";
|
||||
case VideoCodec::PNG: return "png";
|
||||
}
|
||||
return "";
|
||||
}
|
||||
@@ -2698,6 +2762,8 @@ static bool video_codec_only_supports_low_power_mode(const gsr_supported_video_c
|
||||
case VideoCodec::VP9: return supported_video_codecs.vp9.low_power;
|
||||
case VideoCodec::H264_VULKAN: return supported_video_codecs.h264.low_power;
|
||||
case VideoCodec::HEVC_VULKAN: return supported_video_codecs.hevc.low_power; // TODO: hdr, 10 bit
|
||||
case VideoCodec::JPEG: return false;
|
||||
case VideoCodec::PNG: return false;
|
||||
}
|
||||
return false;
|
||||
}
|
||||
@@ -2773,6 +2839,11 @@ static const AVCodec* pick_video_codec(VideoCodec *video_codec, gsr_egl *egl, bo
|
||||
video_codec_f = get_ffmpeg_video_codec(*video_codec, egl->gpu_info.vendor);
|
||||
break;
|
||||
}
|
||||
case VideoCodec::JPEG:
|
||||
case VideoCodec::PNG: {
|
||||
video_codec_f = get_ffmpeg_video_codec(*video_codec, egl->gpu_info.vendor);
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
if(!video_codec_auto && !video_codec_f && !is_flv) {
|
||||
@@ -2834,6 +2905,12 @@ static const AVCodec* pick_video_codec(VideoCodec *video_codec, gsr_egl *egl, bo
|
||||
video_codec_f = get_ffmpeg_video_codec(*video_codec, egl->gpu_info.vendor);
|
||||
break;
|
||||
}
|
||||
case VideoCodec::JPEG:
|
||||
case VideoCodec::PNG: {
|
||||
// TODO:
|
||||
//assert(false);
|
||||
break;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -3007,6 +3084,20 @@ static AudioDeviceData create_application_audio_audio_input(const MergedAudioInp
|
||||
}
|
||||
#endif
|
||||
|
||||
static void set_video_codec_for_image_output(const char *filename, VideoCodec *video_codec, const char **video_codec_to_use) {
|
||||
const bool video_codec_auto = strcmp(*video_codec_to_use, "auto") == 0;
|
||||
if(!video_codec_auto)
|
||||
fprintf(stderr, "Warning: expected -k option to be set to 'auto' (or not specified) for image output\n");
|
||||
|
||||
if(string_ends_with(filename, ".jpg") || string_ends_with(filename, ".jpeg")) {
|
||||
*video_codec = VideoCodec::JPEG;
|
||||
*video_codec_to_use = "jpeg";
|
||||
} else if(string_ends_with(filename, ".png")) {
|
||||
*video_codec = VideoCodec::PNG;
|
||||
*video_codec_to_use = "png";
|
||||
}
|
||||
}
|
||||
|
||||
static bool arg_get_boolean_value(std::map<std::string, Arg> &args, const char *arg_name, bool default_value) {
|
||||
auto it = args.find(arg_name);
|
||||
if(it == args.end() || !it->second.value()) {
|
||||
@@ -3418,6 +3509,11 @@ int main(int argc, char **argv) {
|
||||
_exit(1);
|
||||
}
|
||||
|
||||
if(video_codec_is_hdr(video_codec) && is_portal_capture) {
|
||||
fprintf(stderr, "Warning: portal capture option doesn't support hdr yet (PipeWire doesn't support hdr), the video will be tonemapped from hdr to sdr\n");
|
||||
video_codec = hdr_video_codec_to_sdr_video_codec(video_codec);
|
||||
}
|
||||
|
||||
const bool is_monitor_capture = strcmp(window_str.c_str(), "focused") != 0 && !is_portal_capture && contains_non_hex_number(window_str.c_str());
|
||||
gsr_egl egl;
|
||||
if(!gsr_egl_load(&egl, window, is_monitor_capture, gl_debug)) {
|
||||
@@ -3637,6 +3733,12 @@ int main(int argc, char **argv) {
|
||||
|
||||
const bool is_output_piped = strcmp(filename, "/dev/stdout") == 0;
|
||||
|
||||
set_video_codec_for_image_output(filename, &video_codec, &video_codec_to_use);
|
||||
if(video_codec_is_image(video_codec) && !audio_input_arg.values.empty()) {
|
||||
fprintf(stderr, "Error: can't record audio (-a) when taking a screenshot\n");
|
||||
_exit(1);
|
||||
}
|
||||
|
||||
AVFormatContext *av_format_context;
|
||||
// The output format is automatically guessed by the file extension
|
||||
avformat_alloc_output_context2(&av_format_context, nullptr, container_format, filename);
|
||||
@@ -3662,18 +3764,14 @@ int main(int argc, char **argv) {
|
||||
const bool force_no_audio_offset = is_livestream || is_output_piped || (file_extension != "mp4" && file_extension != "mkv" && file_extension != "webm");
|
||||
const double target_fps = 1.0 / (double)fps;
|
||||
|
||||
if(video_codec_is_hdr(video_codec) && is_portal_capture) {
|
||||
fprintf(stderr, "Warning: portal capture option doesn't support hdr yet (PipeWire doesn't support hdr), the video will be tonemapped from hdr to sdr\n");
|
||||
video_codec = hdr_video_codec_to_sdr_video_codec(video_codec);
|
||||
}
|
||||
|
||||
const bool uses_amix = merged_audio_inputs_should_use_amix(requested_audio_inputs);
|
||||
audio_codec = select_audio_codec_with_fallback(audio_codec, file_extension, uses_amix);
|
||||
if(!video_codec_is_image(video_codec))
|
||||
audio_codec = select_audio_codec_with_fallback(audio_codec, file_extension, uses_amix);
|
||||
bool low_power = false;
|
||||
const AVCodec *video_codec_f = select_video_codec_with_fallback(&video_codec, video_codec_to_use, file_extension.c_str(), use_software_video_encoder, &egl, &low_power);
|
||||
|
||||
const gsr_color_depth color_depth = video_codec_to_bit_depth(video_codec);
|
||||
gsr_capture *capture = create_capture_impl(window_str, output_resolution, wayland, &egl, fps, video_codec, color_range, record_cursor, use_software_video_encoder, restore_portal_session, portal_session_token_filepath, color_depth);
|
||||
gsr_capture *capture = create_capture_impl(window_str, output_resolution, wayland, &egl, fps, video_codec, color_range, record_cursor, restore_portal_session, portal_session_token_filepath, color_depth);
|
||||
|
||||
// (Some?) livestreaming services require at least one audio track to work.
|
||||
// If not audio is provided then create one silent audio track.
|
||||
@@ -3745,7 +3843,9 @@ int main(int argc, char **argv) {
|
||||
|
||||
gsr_color_conversion_clear(&color_conversion);
|
||||
|
||||
if(use_software_video_encoder) {
|
||||
if(video_codec_is_image(video_codec)) {
|
||||
open_video_image(video_codec_context);
|
||||
} else if(use_software_video_encoder) {
|
||||
open_video_software(video_codec_context, quality, pixel_format, hdr, color_depth, bitrate_mode);
|
||||
} else {
|
||||
open_video_hardware(video_codec_context, quality, very_old_gpu, egl.gpu_info.vendor, pixel_format, hdr, color_depth, bitrate_mode, video_codec, low_power);
|
||||
@@ -3835,6 +3935,8 @@ int main(int argc, char **argv) {
|
||||
if(replay_buffer_size_secs == -1) {
|
||||
AVDictionary *options = nullptr;
|
||||
av_dict_set(&options, "strict", "experimental", 0);
|
||||
if(video_codec_is_image(video_codec))
|
||||
av_dict_set(&options, "update", "true", 0);
|
||||
//av_dict_set_int(&av_format_context->metadata, "video_full_range_flag", 1, 0);
|
||||
|
||||
int ret = avformat_write_header(av_format_context, &options);
|
||||
@@ -4084,6 +4186,7 @@ int main(int argc, char **argv) {
|
||||
|
||||
double last_capture_seconds = record_start_time;
|
||||
bool wait_until_frame_time_elapsed = false;
|
||||
const bool is_image_output = video_codec_is_image(video_codec);
|
||||
|
||||
while(running) {
|
||||
const double frame_start = clock_get_monotonic_seconds();
|
||||
@@ -4191,6 +4294,10 @@ int main(int argc, char **argv) {
|
||||
// TODO: Move to separate thread because this could write to network (for example when livestreaming)
|
||||
receive_frames(video_codec_context, VIDEO_STREAM_INDEX, video_stream, video_frame->pts, av_format_context,
|
||||
replay_start_time, frame_data_queue, replay_buffer_size_secs, frames_erased, write_output_mutex, paused_time_offset);
|
||||
if(is_image_output) {
|
||||
running = 0;
|
||||
break;
|
||||
}
|
||||
} else {
|
||||
fprintf(stderr, "Error: avcodec_send_frame failed, error: %s\n", av_error_to_string(ret));
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user