mirror of
https://github.com/libsdl-org/SDL.git
synced 2024-12-12 21:23:28 +08:00
1359 lines
46 KiB
C
1359 lines
46 KiB
C
/*
|
|
Copyright (C) 1997-2024 Sam Lantinga <slouken@libsdl.org>
|
|
|
|
This software is provided 'as-is', without any express or implied
|
|
warranty. In no event will the authors be held liable for any damages
|
|
arising from the use of this software.
|
|
|
|
Permission is granted to anyone to use this software for any purpose,
|
|
including commercial applications, and to alter it and redistribute it
|
|
freely.
|
|
*/
|
|
/* Simple program: Display a video with a sprite bouncing around over it
|
|
*
|
|
* For a more complete video example, see ffplay.c in the ffmpeg sources.
|
|
*/
|
|
|
|
#include <stdlib.h>
|
|
#include <time.h>
|
|
|
|
#include <SDL3/SDL.h>
|
|
#include <SDL3/SDL_main.h>
|
|
#include <SDL3/SDL_test.h>
|
|
|
|
#include <libavcodec/avcodec.h>
|
|
#include <libavformat/avformat.h>
|
|
#include <libavutil/avutil.h>
|
|
#include <libavutil/mastering_display_metadata.h>
|
|
#include <libavutil/pixdesc.h>
|
|
#include <libswscale/swscale.h>
|
|
|
|
#ifdef HAVE_EGL
|
|
#include <SDL3/SDL_opengl.h>
|
|
#include <SDL3/SDL_opengles2.h>
|
|
#include <SDL3/SDL_egl.h>
|
|
|
|
#include <libavutil/hwcontext_drm.h>
|
|
|
|
#ifndef fourcc_code
|
|
#define fourcc_code(a, b, c, d) ((uint32_t)(a) | ((uint32_t)(b) << 8) | ((uint32_t)(c) << 16) | ((uint32_t)(d) << 24))
|
|
#endif
|
|
#ifndef DRM_FORMAT_R8
|
|
#define DRM_FORMAT_R8 fourcc_code('R', '8', ' ', ' ')
|
|
#endif
|
|
#ifndef DRM_FORMAT_GR88
|
|
#define DRM_FORMAT_GR88 fourcc_code('G', 'R', '8', '8')
|
|
#endif
|
|
#endif
|
|
|
|
#ifdef SDL_PLATFORM_APPLE
|
|
#include <CoreVideo/CoreVideo.h>
|
|
#endif
|
|
|
|
#ifdef SDL_PLATFORM_WIN32
|
|
#define COBJMACROS
|
|
#include <libavutil/hwcontext_d3d11va.h>
|
|
#endif /* SDL_PLATFORM_WIN32 */
|
|
|
|
#include "testffmpeg_vulkan.h"
|
|
|
|
#include "icon.h"
|
|
|
|
|
|
static SDL_Texture *sprite;
|
|
static SDL_FRect *positions;
|
|
static SDL_FRect *velocities;
|
|
static int sprite_w, sprite_h;
|
|
static int num_sprites = 0;
|
|
|
|
static SDL_Window *window;
|
|
static SDL_Renderer *renderer;
|
|
static SDL_AudioStream *audio;
|
|
static SDL_Texture *video_texture;
|
|
static Uint64 video_start;
|
|
static SDL_bool software_only;
|
|
static SDL_bool has_eglCreateImage;
|
|
#ifdef HAVE_EGL
|
|
static SDL_bool has_EGL_EXT_image_dma_buf_import;
|
|
static SDL_bool has_EGL_EXT_image_dma_buf_import_modifiers;
|
|
static PFNGLACTIVETEXTUREARBPROC glActiveTextureARBFunc;
|
|
static PFNGLEGLIMAGETARGETTEXTURE2DOESPROC glEGLImageTargetTexture2DOESFunc;
|
|
#endif
|
|
#ifdef SDL_PLATFORM_WIN32
|
|
static ID3D11Device *d3d11_device;
|
|
static ID3D11DeviceContext *d3d11_context;
|
|
static const GUID SDL_IID_ID3D11Resource = { 0xdc8e63f3, 0xd12b, 0x4952, { 0xb4, 0x7b, 0x5e, 0x45, 0x02, 0x6a, 0x86, 0x2d } };
|
|
#endif
|
|
static VulkanVideoContext *vulkan_context;
|
|
struct SwsContextContainer
|
|
{
|
|
struct SwsContext *context;
|
|
};
|
|
static const char *SWS_CONTEXT_CONTAINER_PROPERTY = "SWS_CONTEXT_CONTAINER";
|
|
static int done;
|
|
|
|
static SDL_bool CreateWindowAndRenderer(SDL_WindowFlags window_flags, const char *driver)
|
|
{
|
|
SDL_PropertiesID props;
|
|
SDL_RendererInfo info;
|
|
SDL_bool useOpenGL = (driver && (SDL_strcmp(driver, "opengl") == 0 || SDL_strcmp(driver, "opengles2") == 0));
|
|
SDL_bool useEGL = (driver && SDL_strcmp(driver, "opengles2") == 0);
|
|
SDL_bool useVulkan = (driver && SDL_strcmp(driver, "vulkan") == 0);
|
|
Uint32 flags = SDL_WINDOW_HIDDEN;
|
|
|
|
if (useOpenGL) {
|
|
if (useEGL) {
|
|
SDL_SetHint(SDL_HINT_VIDEO_FORCE_EGL, "1");
|
|
SDL_GL_SetAttribute(SDL_GL_CONTEXT_PROFILE_MASK, SDL_GL_CONTEXT_PROFILE_ES);
|
|
SDL_GL_SetAttribute(SDL_GL_CONTEXT_MAJOR_VERSION, 2);
|
|
SDL_GL_SetAttribute(SDL_GL_CONTEXT_MINOR_VERSION, 0);
|
|
} else {
|
|
SDL_SetHint(SDL_HINT_VIDEO_FORCE_EGL, "0");
|
|
SDL_GL_SetAttribute(SDL_GL_CONTEXT_PROFILE_MASK, 0);
|
|
SDL_GL_SetAttribute(SDL_GL_CONTEXT_MAJOR_VERSION, 2);
|
|
SDL_GL_SetAttribute(SDL_GL_CONTEXT_MINOR_VERSION, 1);
|
|
}
|
|
SDL_GL_SetAttribute(SDL_GL_RED_SIZE, 5);
|
|
SDL_GL_SetAttribute(SDL_GL_GREEN_SIZE, 6);
|
|
SDL_GL_SetAttribute(SDL_GL_BLUE_SIZE, 5);
|
|
|
|
flags |= SDL_WINDOW_OPENGL;
|
|
}
|
|
if (useVulkan) {
|
|
flags |= SDL_WINDOW_VULKAN;
|
|
}
|
|
|
|
/* The window will be resized to the video size when it's loaded, in OpenVideoStream() */
|
|
window = SDL_CreateWindow("testffmpeg", 1920, 1080, flags);
|
|
if (!window) {
|
|
return SDL_FALSE;
|
|
}
|
|
|
|
if (useVulkan) {
|
|
vulkan_context = CreateVulkanVideoContext(window);
|
|
if (!vulkan_context) {
|
|
SDL_DestroyWindow(window);
|
|
window = NULL;
|
|
return SDL_FALSE;
|
|
}
|
|
}
|
|
|
|
props = SDL_CreateProperties();
|
|
SDL_SetStringProperty(props, SDL_PROP_RENDERER_CREATE_NAME_STRING, driver);
|
|
SDL_SetProperty(props, SDL_PROP_RENDERER_CREATE_WINDOW_POINTER, window);
|
|
if (useVulkan) {
|
|
SetupVulkanRenderProperties(vulkan_context, props);
|
|
}
|
|
if (SDL_GetBooleanProperty(SDL_GetDisplayProperties(SDL_GetDisplayForWindow(window)), SDL_PROP_DISPLAY_HDR_ENABLED_BOOLEAN, SDL_FALSE)) {
|
|
/* Try to create an HDR capable renderer */
|
|
SDL_SetNumberProperty(props, SDL_PROP_RENDERER_CREATE_OUTPUT_COLORSPACE_NUMBER, SDL_COLORSPACE_SRGB_LINEAR);
|
|
renderer = SDL_CreateRendererWithProperties(props);
|
|
}
|
|
if (!renderer) {
|
|
/* Try again with the sRGB colorspace */
|
|
SDL_SetNumberProperty(props, SDL_PROP_RENDERER_CREATE_OUTPUT_COLORSPACE_NUMBER, SDL_COLORSPACE_SRGB);
|
|
renderer = SDL_CreateRendererWithProperties(props);
|
|
}
|
|
SDL_DestroyProperties(props);
|
|
if (!renderer) {
|
|
SDL_DestroyWindow(window);
|
|
window = NULL;
|
|
return SDL_FALSE;
|
|
}
|
|
|
|
if (SDL_GetRendererInfo(renderer, &info) == 0) {
|
|
SDL_Log("Created renderer %s\n", info.name);
|
|
}
|
|
|
|
#ifdef HAVE_EGL
|
|
if (useEGL) {
|
|
const char *egl_extensions = eglQueryString(eglGetCurrentDisplay(), EGL_EXTENSIONS);
|
|
if (!egl_extensions) {
|
|
return SDL_FALSE;
|
|
}
|
|
|
|
char *extensions = SDL_strdup(egl_extensions);
|
|
if (!extensions) {
|
|
return SDL_FALSE;
|
|
}
|
|
|
|
char *saveptr, *token;
|
|
token = SDL_strtok_r(extensions, " ", &saveptr);
|
|
if (!token) {
|
|
free(extensions);
|
|
return SDL_FALSE;
|
|
}
|
|
do {
|
|
if (SDL_strcmp(token, "EGL_EXT_image_dma_buf_import") == 0) {
|
|
has_EGL_EXT_image_dma_buf_import = SDL_TRUE;
|
|
} else if (SDL_strcmp(token, "EGL_EXT_image_dma_buf_import_modifiers") == 0) {
|
|
has_EGL_EXT_image_dma_buf_import_modifiers = SDL_TRUE;
|
|
}
|
|
} while ((token = SDL_strtok_r(NULL, " ", &saveptr)) != NULL);
|
|
|
|
free(extensions);
|
|
|
|
if (SDL_GL_ExtensionSupported("GL_OES_EGL_image")) {
|
|
glEGLImageTargetTexture2DOESFunc = (PFNGLEGLIMAGETARGETTEXTURE2DOESPROC)eglGetProcAddress("glEGLImageTargetTexture2DOES");
|
|
}
|
|
|
|
glActiveTextureARBFunc = (PFNGLACTIVETEXTUREARBPROC)SDL_GL_GetProcAddress("glActiveTextureARB");
|
|
|
|
if (has_EGL_EXT_image_dma_buf_import &&
|
|
glEGLImageTargetTexture2DOESFunc &&
|
|
glActiveTextureARBFunc) {
|
|
has_eglCreateImage = SDL_TRUE;
|
|
}
|
|
}
|
|
#endif /* HAVE_EGL */
|
|
|
|
#ifdef SDL_PLATFORM_WIN32
|
|
d3d11_device = (ID3D11Device *)SDL_GetProperty(SDL_GetRendererProperties(renderer), SDL_PROP_RENDERER_D3D11_DEVICE_POINTER, NULL);
|
|
if (d3d11_device) {
|
|
ID3D11Device_AddRef(d3d11_device);
|
|
ID3D11Device_GetImmediateContext(d3d11_device, &d3d11_context);
|
|
}
|
|
#endif
|
|
|
|
return SDL_TRUE;
|
|
}
|
|
|
|
static SDL_Texture *CreateTexture(SDL_Renderer *r, unsigned char *data, unsigned int len, int *w, int *h) {
|
|
SDL_Texture *texture = NULL;
|
|
SDL_Surface *surface;
|
|
SDL_RWops *src = SDL_RWFromConstMem(data, len);
|
|
if (src) {
|
|
surface = SDL_LoadBMP_RW(src, SDL_TRUE);
|
|
if (surface) {
|
|
/* Treat white as transparent */
|
|
SDL_SetSurfaceColorKey(surface, SDL_TRUE, SDL_MapRGB(surface->format, 255, 255, 255));
|
|
|
|
texture = SDL_CreateTextureFromSurface(r, surface);
|
|
*w = surface->w;
|
|
*h = surface->h;
|
|
SDL_DestroySurface(surface);
|
|
}
|
|
}
|
|
return texture;
|
|
}
|
|
|
|
static void MoveSprite(void)
|
|
{
|
|
SDL_Rect viewport;
|
|
SDL_FRect *position, *velocity;
|
|
int i;
|
|
|
|
SDL_GetRenderViewport(renderer, &viewport);
|
|
|
|
for (i = 0; i < num_sprites; ++i) {
|
|
position = &positions[i];
|
|
velocity = &velocities[i];
|
|
position->x += velocity->x;
|
|
if ((position->x < 0) || (position->x >= (viewport.w - sprite_w))) {
|
|
velocity->x = -velocity->x;
|
|
position->x += velocity->x;
|
|
}
|
|
position->y += velocity->y;
|
|
if ((position->y < 0) || (position->y >= (viewport.h - sprite_h))) {
|
|
velocity->y = -velocity->y;
|
|
position->y += velocity->y;
|
|
}
|
|
}
|
|
|
|
/* Blit the sprite onto the screen */
|
|
for (i = 0; i < num_sprites; ++i) {
|
|
position = &positions[i];
|
|
|
|
/* Blit the sprite onto the screen */
|
|
SDL_RenderTexture(renderer, sprite, NULL, position);
|
|
}
|
|
}
|
|
|
|
static Uint32 GetTextureFormat(enum AVPixelFormat format)
|
|
{
|
|
switch (format) {
|
|
case AV_PIX_FMT_RGB8:
|
|
return SDL_PIXELFORMAT_RGB332;
|
|
case AV_PIX_FMT_RGB444:
|
|
return SDL_PIXELFORMAT_RGB444;
|
|
case AV_PIX_FMT_RGB555:
|
|
return SDL_PIXELFORMAT_RGB555;
|
|
case AV_PIX_FMT_BGR555:
|
|
return SDL_PIXELFORMAT_BGR555;
|
|
case AV_PIX_FMT_RGB565:
|
|
return SDL_PIXELFORMAT_RGB565;
|
|
case AV_PIX_FMT_BGR565:
|
|
return SDL_PIXELFORMAT_BGR565;
|
|
case AV_PIX_FMT_RGB24:
|
|
return SDL_PIXELFORMAT_RGB24;
|
|
case AV_PIX_FMT_BGR24:
|
|
return SDL_PIXELFORMAT_BGR24;
|
|
case AV_PIX_FMT_0RGB32:
|
|
return SDL_PIXELFORMAT_XRGB8888;
|
|
case AV_PIX_FMT_0BGR32:
|
|
return SDL_PIXELFORMAT_XBGR8888;
|
|
case AV_PIX_FMT_NE(RGB0, 0BGR):
|
|
return SDL_PIXELFORMAT_RGBX8888;
|
|
case AV_PIX_FMT_NE(BGR0, 0RGB):
|
|
return SDL_PIXELFORMAT_BGRX8888;
|
|
case AV_PIX_FMT_RGB32:
|
|
return SDL_PIXELFORMAT_ARGB8888;
|
|
case AV_PIX_FMT_RGB32_1:
|
|
return SDL_PIXELFORMAT_RGBA8888;
|
|
case AV_PIX_FMT_BGR32:
|
|
return SDL_PIXELFORMAT_ABGR8888;
|
|
case AV_PIX_FMT_BGR32_1:
|
|
return SDL_PIXELFORMAT_BGRA8888;
|
|
case AV_PIX_FMT_YUV420P:
|
|
return SDL_PIXELFORMAT_IYUV;
|
|
case AV_PIX_FMT_YUYV422:
|
|
return SDL_PIXELFORMAT_YUY2;
|
|
case AV_PIX_FMT_UYVY422:
|
|
return SDL_PIXELFORMAT_UYVY;
|
|
case AV_PIX_FMT_NV12:
|
|
return SDL_PIXELFORMAT_NV12;
|
|
case AV_PIX_FMT_NV21:
|
|
return SDL_PIXELFORMAT_NV21;
|
|
case AV_PIX_FMT_P010:
|
|
return SDL_PIXELFORMAT_P010;
|
|
default:
|
|
return SDL_PIXELFORMAT_UNKNOWN;
|
|
}
|
|
}
|
|
|
|
static SDL_bool SupportedPixelFormat(enum AVPixelFormat format)
|
|
{
|
|
if (!software_only) {
|
|
if (has_eglCreateImage &&
|
|
(format == AV_PIX_FMT_VAAPI || format == AV_PIX_FMT_DRM_PRIME)) {
|
|
return SDL_TRUE;
|
|
}
|
|
#ifdef SDL_PLATFORM_APPLE
|
|
if (format == AV_PIX_FMT_VIDEOTOOLBOX) {
|
|
return SDL_TRUE;
|
|
}
|
|
#endif
|
|
#ifdef SDL_PLATFORM_WIN32
|
|
if (d3d11_device && format == AV_PIX_FMT_D3D11) {
|
|
return SDL_TRUE;
|
|
}
|
|
#endif
|
|
if (vulkan_context && format == AV_PIX_FMT_VULKAN) {
|
|
return SDL_TRUE;
|
|
}
|
|
}
|
|
|
|
if (GetTextureFormat(format) != SDL_PIXELFORMAT_UNKNOWN) {
|
|
return SDL_TRUE;
|
|
}
|
|
return SDL_FALSE;
|
|
}
|
|
|
|
static enum AVPixelFormat GetSupportedPixelFormat(AVCodecContext *s, const enum AVPixelFormat *pix_fmts)
|
|
{
|
|
const enum AVPixelFormat *p;
|
|
|
|
for (p = pix_fmts; *p != AV_PIX_FMT_NONE; p++) {
|
|
const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(*p);
|
|
|
|
if (!(desc->flags & AV_PIX_FMT_FLAG_HWACCEL)) {
|
|
/* We support all memory formats using swscale */
|
|
break;
|
|
}
|
|
|
|
if (SupportedPixelFormat(*p)) {
|
|
/* We support this format */
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (*p == AV_PIX_FMT_NONE) {
|
|
SDL_Log("Couldn't find a supported pixel format:\n");
|
|
for (p = pix_fmts; *p != AV_PIX_FMT_NONE; p++) {
|
|
SDL_Log(" %s\n", av_get_pix_fmt_name(*p));
|
|
}
|
|
}
|
|
|
|
return *p;
|
|
}
|
|
|
|
static AVCodecContext *OpenVideoStream(AVFormatContext *ic, int stream, const AVCodec *codec)
|
|
{
|
|
AVStream *st = ic->streams[stream];
|
|
AVCodecParameters *codecpar = st->codecpar;
|
|
AVCodecContext *context;
|
|
const AVCodecHWConfig *config;
|
|
enum AVHWDeviceType type;
|
|
int i;
|
|
int result;
|
|
|
|
SDL_Log("Video stream: %s %dx%d\n", avcodec_get_name(codec->id), codecpar->width, codecpar->height);
|
|
|
|
context = avcodec_alloc_context3(NULL);
|
|
if (!context) {
|
|
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "avcodec_alloc_context3 failed");
|
|
return NULL;
|
|
}
|
|
|
|
result = avcodec_parameters_to_context(context, ic->streams[stream]->codecpar);
|
|
if (result < 0) {
|
|
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "avcodec_parameters_to_context failed: %s\n", av_err2str(result));
|
|
avcodec_free_context(&context);
|
|
return NULL;
|
|
}
|
|
context->pkt_timebase = ic->streams[stream]->time_base;
|
|
|
|
/* Look for supported hardware accelerated configurations */
|
|
i = 0;
|
|
while (!context->hw_device_ctx &&
|
|
(config = avcodec_get_hw_config(codec, i++)) != NULL) {
|
|
#if 0
|
|
SDL_Log("Found %s hardware acceleration with pixel format %s\n", av_hwdevice_get_type_name(config->device_type), av_get_pix_fmt_name(config->pix_fmt));
|
|
#endif
|
|
|
|
if (!(config->methods & AV_CODEC_HW_CONFIG_METHOD_HW_DEVICE_CTX) ||
|
|
!SupportedPixelFormat(config->pix_fmt)) {
|
|
continue;
|
|
}
|
|
|
|
type = AV_HWDEVICE_TYPE_NONE;
|
|
while (!context->hw_device_ctx &&
|
|
(type = av_hwdevice_iterate_types(type)) != AV_HWDEVICE_TYPE_NONE) {
|
|
if (type != config->device_type) {
|
|
continue;
|
|
}
|
|
|
|
#ifdef SDL_PLATFORM_WIN32
|
|
if (d3d11_device && type == AV_HWDEVICE_TYPE_D3D11VA) {
|
|
AVD3D11VADeviceContext *device_context;
|
|
|
|
context->hw_device_ctx = av_hwdevice_ctx_alloc(type);
|
|
|
|
device_context = (AVD3D11VADeviceContext *)((AVHWDeviceContext *)context->hw_device_ctx->data)->hwctx;
|
|
device_context->device = d3d11_device;
|
|
ID3D11Device_AddRef(device_context->device);
|
|
device_context->device_context = d3d11_context;
|
|
ID3D11DeviceContext_AddRef(device_context->device_context);
|
|
|
|
result = av_hwdevice_ctx_init(context->hw_device_ctx);
|
|
if (result < 0) {
|
|
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Couldn't create hardware device context: %s", av_err2str(result));
|
|
} else {
|
|
SDL_Log("Using %s hardware acceleration with pixel format %s\n", av_hwdevice_get_type_name(config->device_type), av_get_pix_fmt_name(config->pix_fmt));
|
|
}
|
|
} else
|
|
#endif
|
|
if (vulkan_context && type == AV_HWDEVICE_TYPE_VULKAN) {
|
|
AVVulkanDeviceContext *device_context;
|
|
|
|
context->hw_device_ctx = av_hwdevice_ctx_alloc(type);
|
|
|
|
device_context = (AVVulkanDeviceContext *)((AVHWDeviceContext *)context->hw_device_ctx->data)->hwctx;
|
|
SetupVulkanDeviceContextData(vulkan_context, device_context);
|
|
|
|
result = av_hwdevice_ctx_init(context->hw_device_ctx);
|
|
if (result < 0) {
|
|
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Couldn't create hardware device context: %s", av_err2str(result));
|
|
} else {
|
|
SDL_Log("Using %s hardware acceleration with pixel format %s\n", av_hwdevice_get_type_name(config->device_type), av_get_pix_fmt_name(config->pix_fmt));
|
|
}
|
|
} else {
|
|
result = av_hwdevice_ctx_create(&context->hw_device_ctx, type, NULL, NULL, 0);
|
|
if (result < 0) {
|
|
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Couldn't create hardware device context: %s", av_err2str(result));
|
|
} else {
|
|
SDL_Log("Using %s hardware acceleration with pixel format %s\n", av_hwdevice_get_type_name(config->device_type), av_get_pix_fmt_name(config->pix_fmt));
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Allow supported hardware accelerated pixel formats */
|
|
context->get_format = GetSupportedPixelFormat;
|
|
|
|
result = avcodec_open2(context, codec, NULL);
|
|
if (result < 0) {
|
|
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Couldn't open codec %s: %s", avcodec_get_name(context->codec_id), av_err2str(result));
|
|
avcodec_free_context(&context);
|
|
return NULL;
|
|
}
|
|
|
|
SDL_SetWindowSize(window, codecpar->width, codecpar->height);
|
|
SDL_SetWindowPosition(window, SDL_WINDOWPOS_CENTERED, SDL_WINDOWPOS_CENTERED);
|
|
|
|
return context;
|
|
}
|
|
|
|
static SDL_Colorspace GetFrameColorspace(AVFrame *frame)
|
|
{
|
|
SDL_Colorspace colorspace = SDL_COLORSPACE_SRGB;
|
|
|
|
if (frame && frame->colorspace != AVCOL_SPC_RGB) {
|
|
#ifdef DEBUG_COLORSPACE
|
|
SDL_Log("Frame colorspace: range: %d, primaries: %d, trc: %d, colorspace: %d, chroma_location: %d\n", frame->color_range, frame->color_primaries, frame->color_trc, frame->colorspace, frame->chroma_location);
|
|
#endif
|
|
colorspace = SDL_DEFINE_COLORSPACE(SDL_COLOR_TYPE_YCBCR,
|
|
frame->color_range,
|
|
frame->color_primaries,
|
|
frame->color_trc,
|
|
frame->colorspace,
|
|
frame->chroma_location);
|
|
}
|
|
return colorspace;
|
|
}
|
|
|
|
static SDL_PropertiesID CreateVideoTextureProperties(AVFrame *frame, Uint32 format, int access)
|
|
{
|
|
AVFrameSideData *pSideData;
|
|
SDL_PropertiesID props;
|
|
int width = frame->width;
|
|
int height = frame->height;
|
|
SDL_Colorspace colorspace = GetFrameColorspace(frame);
|
|
|
|
/* ITU-R BT.2408-6 recommends using an SDR white point of 203 nits, which is more likely for game content */
|
|
static const float k_flSDRWhitePoint = 203.0f;
|
|
float flMaxLuminance = k_flSDRWhitePoint;
|
|
|
|
if (frame->hw_frames_ctx) {
|
|
AVHWFramesContext *frames = (AVHWFramesContext *)(frame->hw_frames_ctx->data);
|
|
|
|
width = frames->width;
|
|
height = frames->height;
|
|
if (format == SDL_PIXELFORMAT_UNKNOWN) {
|
|
format = GetTextureFormat(frames->sw_format);
|
|
}
|
|
} else {
|
|
if (format == SDL_PIXELFORMAT_UNKNOWN) {
|
|
format = GetTextureFormat(frame->format);
|
|
}
|
|
}
|
|
|
|
props = SDL_CreateProperties();
|
|
SDL_SetNumberProperty(props, SDL_PROP_TEXTURE_CREATE_COLORSPACE_NUMBER, colorspace);
|
|
pSideData = av_frame_get_side_data(frame, AV_FRAME_DATA_MASTERING_DISPLAY_METADATA);
|
|
if (pSideData) {
|
|
AVMasteringDisplayMetadata *pMasteringDisplayMetadata = (AVMasteringDisplayMetadata *)pSideData->data;
|
|
flMaxLuminance = (float)pMasteringDisplayMetadata->max_luminance.num / pMasteringDisplayMetadata->max_luminance.den;
|
|
} else if (SDL_COLORSPACETRANSFER(colorspace) == SDL_TRANSFER_CHARACTERISTICS_PQ) {
|
|
/* The official definition is 10000, but PQ game content is often mastered for 400 or 1000 nits */
|
|
flMaxLuminance = 1000.0f;
|
|
}
|
|
if (flMaxLuminance > k_flSDRWhitePoint) {
|
|
SDL_SetFloatProperty(props, SDL_PROP_TEXTURE_CREATE_SDR_WHITE_POINT_FLOAT, k_flSDRWhitePoint);
|
|
SDL_SetFloatProperty(props, SDL_PROP_TEXTURE_CREATE_HDR_HEADROOM_FLOAT, flMaxLuminance / k_flSDRWhitePoint);
|
|
}
|
|
SDL_SetNumberProperty(props, SDL_PROP_TEXTURE_CREATE_FORMAT_NUMBER, format);
|
|
SDL_SetNumberProperty(props, SDL_PROP_TEXTURE_CREATE_ACCESS_NUMBER, access);
|
|
SDL_SetNumberProperty(props, SDL_PROP_TEXTURE_CREATE_WIDTH_NUMBER, width);
|
|
SDL_SetNumberProperty(props, SDL_PROP_TEXTURE_CREATE_HEIGHT_NUMBER, height);
|
|
|
|
return props;
|
|
}
|
|
|
|
static void SDLCALL FreeSwsContextContainer(void *userdata, void *value)
|
|
{
|
|
struct SwsContextContainer *sws_container = (struct SwsContextContainer *)value;
|
|
if (sws_container->context) {
|
|
sws_freeContext(sws_container->context);
|
|
}
|
|
SDL_free(sws_container);
|
|
}
|
|
|
|
static SDL_bool GetTextureForMemoryFrame(AVFrame *frame, SDL_Texture **texture)
|
|
{
|
|
int texture_width = 0, texture_height = 0;
|
|
Uint32 texture_format = SDL_PIXELFORMAT_UNKNOWN;
|
|
Uint32 frame_format = GetTextureFormat(frame->format);
|
|
|
|
if (*texture) {
|
|
SDL_QueryTexture(*texture, &texture_format, NULL, &texture_width, &texture_height);
|
|
}
|
|
if (!*texture || texture_width != frame->width || texture_height != frame->height ||
|
|
(frame_format != SDL_PIXELFORMAT_UNKNOWN && texture_format != frame_format) ||
|
|
(frame_format == SDL_PIXELFORMAT_UNKNOWN && texture_format != SDL_PIXELFORMAT_ARGB8888)) {
|
|
if (*texture) {
|
|
SDL_DestroyTexture(*texture);
|
|
}
|
|
|
|
SDL_PropertiesID props;
|
|
if (frame_format == SDL_PIXELFORMAT_UNKNOWN) {
|
|
props = CreateVideoTextureProperties(frame, SDL_PIXELFORMAT_ARGB8888, SDL_TEXTUREACCESS_STREAMING);
|
|
} else {
|
|
props = CreateVideoTextureProperties(frame, frame_format, SDL_TEXTUREACCESS_STREAMING);
|
|
}
|
|
*texture = SDL_CreateTextureWithProperties(renderer, props);
|
|
SDL_DestroyProperties(props);
|
|
if (!*texture) {
|
|
return SDL_FALSE;
|
|
}
|
|
|
|
if (frame_format == SDL_PIXELFORMAT_UNKNOWN || SDL_ISPIXELFORMAT_ALPHA(frame_format)) {
|
|
SDL_SetTextureBlendMode(*texture, SDL_BLENDMODE_BLEND);
|
|
} else {
|
|
SDL_SetTextureBlendMode(*texture, SDL_BLENDMODE_NONE);
|
|
}
|
|
SDL_SetTextureScaleMode(*texture, SDL_SCALEMODE_LINEAR);
|
|
}
|
|
|
|
switch (frame_format) {
|
|
case SDL_PIXELFORMAT_UNKNOWN:
|
|
{
|
|
SDL_PropertiesID props = SDL_GetTextureProperties(*texture);
|
|
struct SwsContextContainer *sws_container = (struct SwsContextContainer *)SDL_GetProperty(props, SWS_CONTEXT_CONTAINER_PROPERTY, NULL);
|
|
if (!sws_container) {
|
|
sws_container = (struct SwsContextContainer *)SDL_calloc(1, sizeof(*sws_container));
|
|
if (!sws_container) {
|
|
return SDL_FALSE;
|
|
}
|
|
SDL_SetPropertyWithCleanup(props, SWS_CONTEXT_CONTAINER_PROPERTY, sws_container, FreeSwsContextContainer, NULL);
|
|
}
|
|
sws_container->context = sws_getCachedContext(sws_container->context, frame->width, frame->height, frame->format, frame->width, frame->height, AV_PIX_FMT_BGRA, SWS_POINT, NULL, NULL, NULL);
|
|
if (sws_container->context) {
|
|
uint8_t *pixels[4];
|
|
int pitch[4];
|
|
if (SDL_LockTexture(*texture, NULL, (void **)&pixels[0], &pitch[0]) == 0) {
|
|
sws_scale(sws_container->context, (const uint8_t * const *)frame->data, frame->linesize, 0, frame->height, pixels, pitch);
|
|
SDL_UnlockTexture(*texture);
|
|
}
|
|
} else {
|
|
SDL_SetError("Can't initialize the conversion context");
|
|
return SDL_FALSE;
|
|
}
|
|
break;
|
|
}
|
|
case SDL_PIXELFORMAT_IYUV:
|
|
if (frame->linesize[0] > 0 && frame->linesize[1] > 0 && frame->linesize[2] > 0) {
|
|
SDL_UpdateYUVTexture(*texture, NULL, frame->data[0], frame->linesize[0],
|
|
frame->data[1], frame->linesize[1],
|
|
frame->data[2], frame->linesize[2]);
|
|
} else if (frame->linesize[0] < 0 && frame->linesize[1] < 0 && frame->linesize[2] < 0) {
|
|
SDL_UpdateYUVTexture(*texture, NULL, frame->data[0] + frame->linesize[0] * (frame->height - 1), -frame->linesize[0],
|
|
frame->data[1] + frame->linesize[1] * (AV_CEIL_RSHIFT(frame->height, 1) - 1), -frame->linesize[1],
|
|
frame->data[2] + frame->linesize[2] * (AV_CEIL_RSHIFT(frame->height, 1) - 1), -frame->linesize[2]);
|
|
}
|
|
break;
|
|
default:
|
|
if (frame->linesize[0] < 0) {
|
|
SDL_UpdateTexture(*texture, NULL, frame->data[0] + frame->linesize[0] * (frame->height - 1), -frame->linesize[0]);
|
|
} else {
|
|
SDL_UpdateTexture(*texture, NULL, frame->data[0], frame->linesize[0]);
|
|
}
|
|
break;
|
|
}
|
|
return SDL_TRUE;
|
|
}
|
|
|
|
static SDL_bool GetTextureForDRMFrame(AVFrame *frame, SDL_Texture **texture)
|
|
{
|
|
#ifdef HAVE_EGL
|
|
AVHWFramesContext *frames = (AVHWFramesContext *)(frame->hw_frames_ctx->data);
|
|
const AVDRMFrameDescriptor *desc = (const AVDRMFrameDescriptor *)frame->data[0];
|
|
int i, j, image_index, num_planes;
|
|
EGLDisplay display = eglGetCurrentDisplay();
|
|
SDL_PropertiesID props;
|
|
GLuint textures[2];
|
|
|
|
/* FIXME: Assuming NV12 data format */
|
|
num_planes = 0;
|
|
for (i = 0; i < desc->nb_layers; ++i) {
|
|
num_planes += desc->layers[i].nb_planes;
|
|
}
|
|
if (num_planes != 2) {
|
|
SDL_SetError("Expected NV12 frames with 2 planes, instead got %d planes", num_planes);
|
|
return SDL_FALSE;
|
|
}
|
|
|
|
if (*texture) {
|
|
/* Free the previous texture now that we're about to render a new one */
|
|
SDL_DestroyTexture(*texture);
|
|
} else {
|
|
/* First time set up for NV12 textures */
|
|
SDL_SetHint("SDL_RENDER_OPENGL_NV12_RG_SHADER", "1");
|
|
}
|
|
|
|
props = CreateVideoTextureProperties(frame, SDL_PIXELFORMAT_UNKNOWN, SDL_TEXTUREACCESS_STATIC);
|
|
*texture = SDL_CreateTextureWithProperties(renderer, props);
|
|
SDL_DestroyProperties(props);
|
|
if (!*texture) {
|
|
return SDL_FALSE;
|
|
}
|
|
SDL_SetTextureBlendMode(*texture, SDL_BLENDMODE_NONE);
|
|
SDL_SetTextureScaleMode(*texture, SDL_SCALEMODE_LINEAR);
|
|
|
|
props = SDL_GetTextureProperties(*texture);
|
|
textures[0] = (GLuint)SDL_GetNumberProperty(props, SDL_PROP_TEXTURE_OPENGLES2_TEXTURE_NUMBER, 0);
|
|
textures[1] = (GLuint)SDL_GetNumberProperty(props, SDL_PROP_TEXTURE_OPENGLES2_TEXTURE_UV_NUMBER, 0);
|
|
if (!textures[0] || !textures[1]) {
|
|
SDL_SetError("Couldn't get NV12 OpenGL textures");
|
|
return SDL_FALSE;
|
|
}
|
|
|
|
/* import the frame into OpenGL */
|
|
image_index = 0;
|
|
for (i = 0; i < desc->nb_layers; ++i) {
|
|
const AVDRMLayerDescriptor *layer = &desc->layers[i];
|
|
for (j = 0; j < layer->nb_planes; ++j) {
|
|
static const uint32_t formats[ 2 ] = { DRM_FORMAT_R8, DRM_FORMAT_GR88 };
|
|
const AVDRMPlaneDescriptor *plane = &layer->planes[j];
|
|
const AVDRMObjectDescriptor *object = &desc->objects[plane->object_index];
|
|
|
|
EGLAttrib attr[32];
|
|
size_t k = 0;
|
|
|
|
attr[k++] = EGL_LINUX_DRM_FOURCC_EXT;
|
|
attr[k++] = formats[i];
|
|
|
|
attr[k++] = EGL_WIDTH;
|
|
attr[k++] = frames->width / ( image_index + 1 ); /* half size for chroma */
|
|
|
|
attr[k++] = EGL_HEIGHT;
|
|
attr[k++] = frames->height / ( image_index + 1 );
|
|
|
|
attr[k++] = EGL_DMA_BUF_PLANE0_FD_EXT;
|
|
attr[k++] = object->fd;
|
|
|
|
attr[k++] = EGL_DMA_BUF_PLANE0_OFFSET_EXT;
|
|
attr[k++] = plane->offset;
|
|
|
|
attr[k++] = EGL_DMA_BUF_PLANE0_PITCH_EXT;
|
|
attr[k++] = plane->pitch;
|
|
|
|
if (has_EGL_EXT_image_dma_buf_import_modifiers) {
|
|
attr[k++] = EGL_DMA_BUF_PLANE0_MODIFIER_LO_EXT;
|
|
attr[k++] = (object->format_modifier >> 0) & 0xFFFFFFFF;
|
|
|
|
attr[k++] = EGL_DMA_BUF_PLANE0_MODIFIER_HI_EXT;
|
|
attr[k++] = (object->format_modifier >> 32) & 0xFFFFFFFF;
|
|
}
|
|
|
|
attr[k++] = EGL_NONE;
|
|
|
|
EGLImage pImage = eglCreateImage(display, EGL_NO_CONTEXT, EGL_LINUX_DMA_BUF_EXT, NULL, attr);
|
|
|
|
glActiveTextureARBFunc(GL_TEXTURE0_ARB + image_index);
|
|
glBindTexture(GL_TEXTURE_2D, textures[image_index]);
|
|
glEGLImageTargetTexture2DOESFunc(GL_TEXTURE_2D, pImage);
|
|
++image_index;
|
|
}
|
|
}
|
|
|
|
return SDL_TRUE;
|
|
#else
|
|
return SDL_FALSE;
|
|
#endif
|
|
}
|
|
|
|
static SDL_bool GetTextureForVAAPIFrame(AVFrame *frame, SDL_Texture **texture)
|
|
{
|
|
AVFrame *drm_frame;
|
|
SDL_bool result = SDL_FALSE;
|
|
|
|
drm_frame = av_frame_alloc();
|
|
if (drm_frame) {
|
|
drm_frame->format = AV_PIX_FMT_DRM_PRIME;
|
|
if (av_hwframe_map(drm_frame, frame, 0) == 0) {
|
|
result = GetTextureForDRMFrame(drm_frame, texture);
|
|
} else {
|
|
SDL_SetError("Couldn't map hardware frame");
|
|
}
|
|
av_frame_free(&drm_frame);
|
|
}
|
|
return result;
|
|
}
|
|
|
|
static SDL_bool GetTextureForD3D11Frame(AVFrame *frame, SDL_Texture **texture)
|
|
{
|
|
#ifdef SDL_PLATFORM_WIN32
|
|
AVHWFramesContext *frames = (AVHWFramesContext *)(frame->hw_frames_ctx->data);
|
|
int texture_width = 0, texture_height = 0;
|
|
ID3D11Texture2D *pTexture = (ID3D11Texture2D *)frame->data[0];
|
|
UINT iSliceIndex = (UINT)(uintptr_t)frame->data[1];
|
|
|
|
if (*texture) {
|
|
SDL_QueryTexture(*texture, NULL, NULL, &texture_width, &texture_height);
|
|
}
|
|
if (!*texture || texture_width != frames->width || texture_height != frames->height) {
|
|
if (*texture) {
|
|
SDL_DestroyTexture(*texture);
|
|
}
|
|
|
|
SDL_PropertiesID props = CreateVideoTextureProperties(frame, SDL_PIXELFORMAT_UNKNOWN, SDL_TEXTUREACCESS_STATIC);
|
|
*texture = SDL_CreateTextureWithProperties(renderer, props);
|
|
SDL_DestroyProperties(props);
|
|
if (!*texture) {
|
|
return SDL_FALSE;
|
|
}
|
|
}
|
|
|
|
ID3D11Resource *dx11_resource = SDL_GetProperty(SDL_GetTextureProperties(*texture), SDL_PROP_TEXTURE_D3D11_TEXTURE_POINTER, NULL);
|
|
if (!dx11_resource) {
|
|
SDL_SetError("Couldn't get texture ID3D11Resource interface");
|
|
return SDL_FALSE;
|
|
}
|
|
ID3D11DeviceContext_CopySubresourceRegion(d3d11_context, dx11_resource, 0, 0, 0, 0, (ID3D11Resource *)pTexture, iSliceIndex, NULL);
|
|
|
|
return SDL_TRUE;
|
|
#else
|
|
return SDL_FALSE;
|
|
#endif
|
|
}
|
|
|
|
static SDL_bool GetTextureForVideoToolboxFrame(AVFrame *frame, SDL_Texture **texture)
|
|
{
|
|
#ifdef SDL_PLATFORM_APPLE
|
|
CVPixelBufferRef pPixelBuffer = (CVPixelBufferRef)frame->data[3];
|
|
SDL_PropertiesID props;
|
|
|
|
if (*texture) {
|
|
/* Free the previous texture now that we're about to render a new one */
|
|
/* FIXME: We can actually keep a cache of textures that map to pixel buffers */
|
|
SDL_DestroyTexture(*texture);
|
|
}
|
|
|
|
props = CreateVideoTextureProperties(frame, SDL_PIXELFORMAT_UNKNOWN, SDL_TEXTUREACCESS_STATIC);
|
|
SDL_SetProperty(props, SDL_PROP_TEXTURE_CREATE_METAL_PIXELBUFFER_POINTER, pPixelBuffer);
|
|
*texture = SDL_CreateTextureWithProperties(renderer, props);
|
|
SDL_DestroyProperties(props);
|
|
if (!*texture) {
|
|
return SDL_FALSE;
|
|
}
|
|
|
|
return SDL_TRUE;
|
|
#else
|
|
return SDL_FALSE;
|
|
#endif
|
|
}
|
|
|
|
static SDL_bool GetTextureForVulkanFrame(AVFrame *frame, SDL_Texture **texture)
|
|
{
|
|
SDL_PropertiesID props;
|
|
|
|
if (*texture) {
|
|
SDL_DestroyTexture(*texture);
|
|
}
|
|
|
|
props = CreateVideoTextureProperties(frame, SDL_PIXELFORMAT_UNKNOWN, SDL_TEXTUREACCESS_STATIC);
|
|
*texture = CreateVulkanVideoTexture(vulkan_context, frame, renderer, props);
|
|
SDL_DestroyProperties(props);
|
|
if (!*texture) {
|
|
return SDL_FALSE;
|
|
}
|
|
return SDL_TRUE;
|
|
}
|
|
|
|
static SDL_bool GetTextureForFrame(AVFrame *frame, SDL_Texture **texture)
|
|
{
|
|
switch (frame->format) {
|
|
case AV_PIX_FMT_VAAPI:
|
|
return GetTextureForVAAPIFrame(frame, texture);
|
|
case AV_PIX_FMT_DRM_PRIME:
|
|
return GetTextureForDRMFrame(frame, texture);
|
|
case AV_PIX_FMT_D3D11:
|
|
return GetTextureForD3D11Frame(frame, texture);
|
|
case AV_PIX_FMT_VIDEOTOOLBOX:
|
|
return GetTextureForVideoToolboxFrame(frame, texture);
|
|
case AV_PIX_FMT_VULKAN:
|
|
return GetTextureForVulkanFrame(frame, texture);
|
|
default:
|
|
return GetTextureForMemoryFrame(frame, texture);
|
|
}
|
|
}
|
|
|
|
static int BeginFrameRendering(AVFrame *frame)
|
|
{
|
|
if (frame->format == AV_PIX_FMT_VULKAN) {
|
|
return BeginVulkanFrameRendering(vulkan_context, frame, renderer);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int FinishFrameRendering(AVFrame *frame)
|
|
{
|
|
if (frame->format == AV_PIX_FMT_VULKAN) {
|
|
return FinishVulkanFrameRendering(vulkan_context, frame, renderer);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static void DisplayVideoTexture(AVFrame *frame)
|
|
{
|
|
/* Update the video texture */
|
|
if (!GetTextureForFrame(frame, &video_texture)) {
|
|
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Couldn't get texture for frame: %s\n", SDL_GetError());
|
|
return;
|
|
}
|
|
|
|
SDL_FRect src;
|
|
src.x = 0.0f;
|
|
src.y = 0.0f;
|
|
src.w = (float)frame->width;
|
|
src.h = (float)frame->height;
|
|
if (frame->linesize[0] < 0) {
|
|
SDL_RenderTextureRotated(renderer, video_texture, &src, NULL, 0.0, NULL, SDL_FLIP_VERTICAL);
|
|
} else {
|
|
SDL_RenderTexture(renderer, video_texture, &src, NULL);
|
|
}
|
|
}
|
|
|
|
static void DisplayVideoFrame(AVFrame *frame)
|
|
{
|
|
DisplayVideoTexture(frame);
|
|
}
|
|
|
|
static void HandleVideoFrame(AVFrame *frame, double pts)
|
|
{
|
|
/* Quick and dirty PTS handling */
|
|
if (!video_start) {
|
|
video_start = SDL_GetTicks();
|
|
}
|
|
double now = (double)(SDL_GetTicks() - video_start) / 1000.0;
|
|
while (now < pts - 0.001) {
|
|
SDL_Delay(1);
|
|
now = (double)(SDL_GetTicks() - video_start) / 1000.0;
|
|
}
|
|
|
|
if (BeginFrameRendering(frame) < 0) {
|
|
return;
|
|
}
|
|
|
|
SDL_SetRenderDrawColor(renderer, 0, 0, 0, 255);
|
|
SDL_RenderClear(renderer);
|
|
|
|
DisplayVideoFrame(frame);
|
|
|
|
/* Render any bouncing balls */
|
|
MoveSprite();
|
|
|
|
SDL_RenderPresent(renderer);
|
|
|
|
FinishFrameRendering(frame);
|
|
}
|
|
|
|
static AVCodecContext *OpenAudioStream(AVFormatContext *ic, int stream, const AVCodec *codec)
|
|
{
|
|
AVStream *st = ic->streams[stream];
|
|
AVCodecParameters *codecpar = st->codecpar;
|
|
AVCodecContext *context;
|
|
int result;
|
|
|
|
SDL_Log("Audio stream: %s %d channels, %d Hz\n", avcodec_get_name(codec->id), codecpar->ch_layout.nb_channels, codecpar->sample_rate);
|
|
|
|
context = avcodec_alloc_context3(NULL);
|
|
if (!context) {
|
|
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "avcodec_alloc_context3 failed\n");
|
|
return NULL;
|
|
}
|
|
|
|
result = avcodec_parameters_to_context(context, ic->streams[stream]->codecpar);
|
|
if (result < 0) {
|
|
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "avcodec_parameters_to_context failed: %s\n", av_err2str(result));
|
|
avcodec_free_context(&context);
|
|
return NULL;
|
|
}
|
|
context->pkt_timebase = ic->streams[stream]->time_base;
|
|
|
|
result = avcodec_open2(context, codec, NULL);
|
|
if (result < 0) {
|
|
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Couldn't open codec %s: %s", avcodec_get_name(context->codec_id), av_err2str(result));
|
|
avcodec_free_context(&context);
|
|
return NULL;
|
|
}
|
|
|
|
SDL_AudioSpec spec = { SDL_AUDIO_F32, codecpar->ch_layout.nb_channels, codecpar->sample_rate };
|
|
audio = SDL_OpenAudioDeviceStream(SDL_AUDIO_DEVICE_DEFAULT_OUTPUT, &spec, NULL, NULL);
|
|
if (audio) {
|
|
SDL_ResumeAudioDevice(SDL_GetAudioStreamDevice(audio));
|
|
} else {
|
|
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Couldn't open audio: %s", SDL_GetError());
|
|
}
|
|
return context;
|
|
}
|
|
|
|
static SDL_AudioFormat GetAudioFormat(enum AVSampleFormat format)
|
|
{
|
|
switch (format) {
|
|
case AV_SAMPLE_FMT_U8:
|
|
case AV_SAMPLE_FMT_U8P:
|
|
return SDL_AUDIO_U8;
|
|
case AV_SAMPLE_FMT_S16:
|
|
case AV_SAMPLE_FMT_S16P:
|
|
return SDL_AUDIO_S16;
|
|
case AV_SAMPLE_FMT_S32:
|
|
case AV_SAMPLE_FMT_S32P:
|
|
return SDL_AUDIO_S32;
|
|
case AV_SAMPLE_FMT_FLT:
|
|
case AV_SAMPLE_FMT_FLTP:
|
|
return SDL_AUDIO_F32;
|
|
default:
|
|
/* Unsupported */
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
static SDL_bool IsPlanarAudioFormat(enum AVSampleFormat format)
|
|
{
|
|
switch (format) {
|
|
case AV_SAMPLE_FMT_U8P:
|
|
case AV_SAMPLE_FMT_S16P:
|
|
case AV_SAMPLE_FMT_S32P:
|
|
case AV_SAMPLE_FMT_FLTP:
|
|
case AV_SAMPLE_FMT_DBLP:
|
|
case AV_SAMPLE_FMT_S64P:
|
|
return SDL_TRUE;
|
|
default:
|
|
return SDL_FALSE;
|
|
}
|
|
}
|
|
|
|
static void InterleaveAudio(AVFrame *frame, const SDL_AudioSpec *spec)
|
|
{
|
|
int c, n;
|
|
int samplesize = SDL_AUDIO_BYTESIZE(spec->format);
|
|
int framesize = SDL_AUDIO_FRAMESIZE(*spec);
|
|
Uint8 *data = (Uint8 *)SDL_malloc(frame->nb_samples * framesize);
|
|
if (!data) {
|
|
return;
|
|
}
|
|
|
|
/* This could be optimized with SIMD and not allocating memory each time */
|
|
for (c = 0; c < spec->channels; ++c) {
|
|
const Uint8 *src = frame->data[c];
|
|
Uint8 *dst = data + c * samplesize;
|
|
for (n = frame->nb_samples; n--; ) {
|
|
SDL_memcpy(dst, src, samplesize);
|
|
src += samplesize;
|
|
dst += framesize;
|
|
}
|
|
}
|
|
SDL_PutAudioStreamData(audio, data, frame->nb_samples * framesize);
|
|
SDL_free(data);
|
|
}
|
|
|
|
static void HandleAudioFrame(AVFrame *frame)
|
|
{
|
|
if (audio) {
|
|
SDL_AudioSpec spec = { GetAudioFormat(frame->format), frame->ch_layout.nb_channels, frame->sample_rate };
|
|
SDL_SetAudioStreamFormat(audio, &spec, NULL);
|
|
|
|
if (frame->ch_layout.nb_channels > 1 && IsPlanarAudioFormat(frame->format)) {
|
|
InterleaveAudio(frame, &spec);
|
|
} else {
|
|
SDL_PutAudioStreamData(audio, frame->data[0], frame->nb_samples * SDL_AUDIO_FRAMESIZE(spec));
|
|
}
|
|
}
|
|
}
|
|
|
|
static void print_usage(SDLTest_CommonState *state, const char *argv0) {
|
|
static const char *options[] = { "[--sprites N]", "[--audio-codec codec]", "[--video-codec codec]", "[--software]", "video_file", NULL };
|
|
SDLTest_CommonLogUsage(state, argv0, options);
|
|
}
|
|
|
|
int main(int argc, char *argv[])
|
|
{
|
|
const char *file = NULL;
|
|
AVFormatContext *ic = NULL;
|
|
int audio_stream = -1;
|
|
int video_stream = -1;
|
|
const char *audio_codec_name = NULL;
|
|
const char *video_codec_name = NULL;
|
|
const AVCodec *audio_codec = NULL;
|
|
const AVCodec *video_codec = NULL;
|
|
AVCodecContext *audio_context = NULL;
|
|
AVCodecContext *video_context = NULL;
|
|
AVPacket *pkt = NULL;
|
|
AVFrame *frame = NULL;
|
|
double first_pts = -1.0;
|
|
int i;
|
|
int result;
|
|
int return_code = -1;
|
|
SDL_WindowFlags window_flags;
|
|
SDL_bool flushing = SDL_FALSE;
|
|
SDL_bool decoded = SDL_FALSE;
|
|
SDLTest_CommonState *state;
|
|
|
|
/* Initialize test framework */
|
|
state = SDLTest_CommonCreateState(argv, 0);
|
|
|
|
/* Enable standard application logging */
|
|
SDL_LogSetPriority(SDL_LOG_CATEGORY_APPLICATION, SDL_LOG_PRIORITY_INFO);
|
|
|
|
|
|
/* Parse commandline */
|
|
for (i = 1; i < argc;) {
|
|
int consumed;
|
|
|
|
consumed = SDLTest_CommonArg(state, i);
|
|
if (!consumed) {
|
|
if (SDL_strcmp(argv[i], "--sprites") == 0 && argv[i+1]) {
|
|
num_sprites = SDL_atoi(argv[i+1]);
|
|
consumed = 2;
|
|
} else if (SDL_strcmp(argv[i], "--audio-codec") == 0 && argv[i+1]) {
|
|
audio_codec_name = argv[i+1];
|
|
consumed = 2;
|
|
} else if (SDL_strcmp(argv[i], "--video-codec") == 0 && argv[i+1]) {
|
|
video_codec_name = argv[i+1];
|
|
consumed = 2;
|
|
} else if (SDL_strcmp(argv[i], "--software") == 0) {
|
|
software_only = SDL_TRUE;
|
|
consumed = 1;
|
|
} else if (!file) {
|
|
/* We'll try to open this as a media file */
|
|
file = argv[i];
|
|
consumed = 1;
|
|
}
|
|
}
|
|
if (consumed <= 0) {
|
|
print_usage(state, argv[0]);
|
|
return_code = 1;
|
|
goto quit;
|
|
}
|
|
|
|
i += consumed;
|
|
}
|
|
|
|
if (!file) {
|
|
print_usage(state, argv[0]);
|
|
return_code = 1;
|
|
goto quit;
|
|
}
|
|
|
|
if (SDL_Init(SDL_INIT_AUDIO|SDL_INIT_VIDEO) < 0) {
|
|
return_code = 2;
|
|
goto quit;
|
|
}
|
|
|
|
window_flags = SDL_WINDOW_HIDDEN | SDL_WINDOW_RESIZABLE | SDL_WINDOW_HIGH_PIXEL_DENSITY;
|
|
#ifdef SDL_PLATFORM_APPLE
|
|
window_flags |= SDL_WINDOW_METAL;
|
|
#elif !defined(SDL_PLATFORM_WIN32)
|
|
window_flags |= SDL_WINDOW_OPENGL;
|
|
#endif
|
|
if (SDL_GetHint(SDL_HINT_RENDER_DRIVER) != NULL) {
|
|
CreateWindowAndRenderer(window_flags, SDL_GetHint(SDL_HINT_RENDER_DRIVER));
|
|
}
|
|
#ifdef HAVE_EGL
|
|
/* Try to create an EGL compatible window for DRM hardware frame support */
|
|
if (!window) {
|
|
CreateWindowAndRenderer(window_flags, "opengles2");
|
|
}
|
|
#endif
|
|
#ifdef SDL_PLATFORM_APPLE
|
|
if (!window) {
|
|
CreateWindowAndRenderer(window_flags, "metal");
|
|
}
|
|
#endif
|
|
#ifdef SDL_PLATFORM_WIN32
|
|
if (!window) {
|
|
CreateWindowAndRenderer(window_flags, "direct3d11");
|
|
}
|
|
#endif
|
|
if (!window) {
|
|
if (!CreateWindowAndRenderer(window_flags, NULL)) {
|
|
return_code = 2;
|
|
goto quit;
|
|
}
|
|
}
|
|
|
|
if (SDL_SetWindowTitle(window, file) < 0) {
|
|
SDL_Log("SDL_SetWindowTitle: %s", SDL_GetError());
|
|
}
|
|
|
|
/* Open the media file */
|
|
result = avformat_open_input(&ic, file, NULL, NULL);
|
|
if (result < 0) {
|
|
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Couldn't open %s: %d", argv[1], result);
|
|
return_code = 4;
|
|
goto quit;
|
|
}
|
|
video_stream = av_find_best_stream(ic, AVMEDIA_TYPE_VIDEO, -1, -1, &video_codec, 0);
|
|
if (video_stream >= 0) {
|
|
if (video_codec_name) {
|
|
video_codec = avcodec_find_decoder_by_name(video_codec_name);
|
|
if (!video_codec) {
|
|
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Couldn't find codec '%s'", video_codec_name);
|
|
return_code = 4;
|
|
goto quit;
|
|
}
|
|
}
|
|
video_context = OpenVideoStream(ic, video_stream, video_codec);
|
|
if (!video_context) {
|
|
return_code = 4;
|
|
goto quit;
|
|
}
|
|
}
|
|
audio_stream = av_find_best_stream(ic, AVMEDIA_TYPE_AUDIO, -1, video_stream, &audio_codec, 0);
|
|
if (audio_stream >= 0) {
|
|
if (audio_codec_name) {
|
|
audio_codec = avcodec_find_decoder_by_name(audio_codec_name);
|
|
if (!audio_codec) {
|
|
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Couldn't find codec '%s'", audio_codec_name);
|
|
return_code = 4;
|
|
goto quit;
|
|
}
|
|
}
|
|
audio_context = OpenAudioStream(ic, audio_stream, audio_codec);
|
|
if (!audio_context) {
|
|
return_code = 4;
|
|
goto quit;
|
|
}
|
|
}
|
|
pkt = av_packet_alloc();
|
|
if (!pkt) {
|
|
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "av_packet_alloc failed");
|
|
return_code = 4;
|
|
goto quit;
|
|
}
|
|
frame = av_frame_alloc();
|
|
if (!frame) {
|
|
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "av_frame_alloc failed");
|
|
return_code = 4;
|
|
goto quit;
|
|
}
|
|
|
|
/* Create the sprite */
|
|
sprite = CreateTexture(renderer, icon_bmp, icon_bmp_len, &sprite_w, &sprite_h);
|
|
|
|
if (!sprite) {
|
|
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Couldn't create texture (%s)", SDL_GetError());
|
|
return_code = 3;
|
|
goto quit;
|
|
}
|
|
|
|
/* Allocate memory for the sprite info */
|
|
positions = (SDL_FRect *)SDL_malloc(num_sprites * sizeof(*positions));
|
|
velocities = (SDL_FRect *)SDL_malloc(num_sprites * sizeof(*velocities));
|
|
if (!positions || !velocities) {
|
|
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Out of memory!\n");
|
|
return_code = 3;
|
|
goto quit;
|
|
}
|
|
|
|
/* Position sprites and set their velocities */
|
|
SDL_Rect viewport;
|
|
SDL_GetRenderViewport(renderer, &viewport);
|
|
srand((unsigned int)time(NULL));
|
|
for (i = 0; i < num_sprites; ++i) {
|
|
positions[i].x = (float)(rand() % (viewport.w - sprite_w));
|
|
positions[i].y = (float)(rand() % (viewport.h - sprite_h));
|
|
positions[i].w = (float)sprite_w;
|
|
positions[i].h = (float)sprite_h;
|
|
velocities[i].x = 0.0f;
|
|
velocities[i].y = 0.0f;
|
|
while (!velocities[i].x || !velocities[i].y) {
|
|
velocities[i].x = (float)((rand() % (2 + 1)) - 1);
|
|
velocities[i].y = (float)((rand() % (2 + 1)) - 1);
|
|
}
|
|
}
|
|
|
|
/* We're ready to go! */
|
|
SDL_ShowWindow(window);
|
|
|
|
/* Main render loop */
|
|
done = 0;
|
|
|
|
while (!done) {
|
|
SDL_Event event;
|
|
|
|
/* Check for events */
|
|
while (SDL_PollEvent(&event)) {
|
|
if (event.type == SDL_EVENT_QUIT ||
|
|
(event.type == SDL_EVENT_KEY_DOWN && event.key.keysym.sym == SDLK_ESCAPE)) {
|
|
done = 1;
|
|
}
|
|
}
|
|
|
|
if (!flushing) {
|
|
result = av_read_frame(ic, pkt);
|
|
if (result < 0) {
|
|
SDL_Log("End of stream, finishing decode\n");
|
|
if (audio_context) {
|
|
avcodec_flush_buffers(audio_context);
|
|
}
|
|
if (video_context) {
|
|
avcodec_flush_buffers(video_context);
|
|
}
|
|
flushing = SDL_TRUE;
|
|
} else {
|
|
if (pkt->stream_index == audio_stream) {
|
|
result = avcodec_send_packet(audio_context, pkt);
|
|
if (result < 0) {
|
|
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "avcodec_send_packet(audio_context) failed: %s", av_err2str(result));
|
|
}
|
|
} else if (pkt->stream_index == video_stream) {
|
|
result = avcodec_send_packet(video_context, pkt);
|
|
if (result < 0) {
|
|
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "avcodec_send_packet(video_context) failed: %s", av_err2str(result));
|
|
}
|
|
}
|
|
av_packet_unref(pkt);
|
|
}
|
|
}
|
|
|
|
decoded = SDL_FALSE;
|
|
if (audio_context) {
|
|
while (avcodec_receive_frame(audio_context, frame) >= 0) {
|
|
HandleAudioFrame(frame);
|
|
decoded = SDL_TRUE;
|
|
}
|
|
if (flushing) {
|
|
/* Let SDL know we're done sending audio */
|
|
SDL_FlushAudioStream(audio);
|
|
}
|
|
}
|
|
if (video_context) {
|
|
while (avcodec_receive_frame(video_context, frame) >= 0) {
|
|
double pts = ((double)frame->pts * video_context->pkt_timebase.num) / video_context->pkt_timebase.den;
|
|
if (first_pts < 0.0) {
|
|
first_pts = pts;
|
|
}
|
|
pts -= first_pts;
|
|
|
|
HandleVideoFrame(frame, pts);
|
|
decoded = SDL_TRUE;
|
|
}
|
|
} else {
|
|
/* Update video rendering */
|
|
SDL_SetRenderDrawColor(renderer, 0xA0, 0xA0, 0xA0, 0xFF);
|
|
SDL_RenderClear(renderer);
|
|
MoveSprite();
|
|
SDL_RenderPresent(renderer);
|
|
}
|
|
|
|
if (flushing && !decoded) {
|
|
if (SDL_GetAudioStreamQueued(audio) > 0) {
|
|
/* Wait a little bit for the audio to finish */
|
|
SDL_Delay(10);
|
|
} else {
|
|
done = 1;
|
|
}
|
|
}
|
|
}
|
|
return_code = 0;
|
|
quit:
|
|
#ifdef SDL_PLATFORM_WIN32
|
|
if (d3d11_context) {
|
|
ID3D11DeviceContext_Release(d3d11_context);
|
|
d3d11_context = NULL;
|
|
}
|
|
if (d3d11_device) {
|
|
ID3D11Device_Release(d3d11_device);
|
|
d3d11_device = NULL;
|
|
}
|
|
#endif
|
|
SDL_free(positions);
|
|
SDL_free(velocities);
|
|
av_frame_free(&frame);
|
|
av_packet_free(&pkt);
|
|
avcodec_free_context(&audio_context);
|
|
avcodec_free_context(&video_context);
|
|
avformat_close_input(&ic);
|
|
SDL_DestroyRenderer(renderer);
|
|
if (vulkan_context) {
|
|
DestroyVulkanVideoContext(vulkan_context);
|
|
}
|
|
SDL_DestroyWindow(window);
|
|
SDL_Quit();
|
|
SDLTest_CommonDestroyState(state);
|
|
return return_code;
|
|
}
|