X-Git-Url: http://de.git.xonotic.org/?p=xonotic%2Fdarkplaces.git;a=blobdiff_plain;f=cap_ogg.c;h=3318c356c4119bb34e0465ed7307d3c4de7aa568;hp=87ff6847b1aad7bd2769b0519b94d46b7bf20a59;hb=1739bc4c7b59ff54fd2f3c0492ab4ca6ed6563ac;hpb=c7c115abcc0d036e0ad5b77f6e147ab560f4c3b9 diff --git a/cap_ogg.c b/cap_ogg.c index 87ff6847..3318c356 100644 --- a/cap_ogg.c +++ b/cap_ogg.c @@ -1,3 +1,6 @@ +#ifndef _MSC_VER +#include +#endif #include #include "quakedef.h" @@ -8,19 +11,27 @@ static cvar_t cl_capturevideo_ogg_theora_quality = {CVAR_SAVE, "cl_capturevideo_ogg_theora_quality", "32", "video quality factor (0 to 63), or -1 to use bitrate only; higher is better"}; static cvar_t cl_capturevideo_ogg_theora_bitrate = {CVAR_SAVE, "cl_capturevideo_ogg_theora_bitrate", "-1", "video bitrate (45 to 2000 kbps), or -1 to use quality only; higher is better"}; static cvar_t cl_capturevideo_ogg_theora_keyframe_bitrate_multiplier = {CVAR_SAVE, "cl_capturevideo_ogg_theora_keyframe_bitrate_multiplier", "1.5", "how much more bit rate to use for keyframes, specified as a factor of at least 1"}; -static cvar_t cl_capturevideo_ogg_theora_keyframe_frequency = {CVAR_SAVE, "cl_capturevideo_ogg_theora_keyframe_frequency", "64", "maximum number of frames between two key frames (1 to 1000)"}; -static cvar_t cl_capturevideo_ogg_theora_keyframe_mindistance = {CVAR_SAVE, "cl_capturevideo_ogg_theora_keyframe_mindistance", "8", "minimum number of frames between two key frames (1 to 1000)"}; +static cvar_t cl_capturevideo_ogg_theora_keyframe_maxinterval = {CVAR_SAVE, "cl_capturevideo_ogg_theora_keyframe_maxinterval", "64", "maximum keyframe interval (1 to 1000)"}; +static cvar_t cl_capturevideo_ogg_theora_keyframe_mininterval = {CVAR_SAVE, "cl_capturevideo_ogg_theora_keyframe_mininterval", "8", "minimum keyframe interval (1 to 1000)"}; static cvar_t cl_capturevideo_ogg_theora_keyframe_auto_threshold = {CVAR_SAVE, "cl_capturevideo_ogg_theora_keyframe_auto_threshold", "80", "threshold for key frame decision (0 to 100)"}; static cvar_t cl_capturevideo_ogg_theora_noise_sensitivity = {CVAR_SAVE, "cl_capturevideo_ogg_theora_noise_sensitivity", "1", "video noise sensitivity (0 to 6); lower is better"}; static cvar_t cl_capturevideo_ogg_theora_sharpness = {CVAR_SAVE, "cl_capturevideo_ogg_theora_sharpness", "0", "sharpness (0 to 2); lower is sharper"}; -static cvar_t cl_capturevideo_ogg_vorbis_quality = {CVAR_SAVE, "cl_capturevideo_ogg_vorbis_quality", "1", "audio quality (-1 to 10); higher is better"}; +static cvar_t cl_capturevideo_ogg_vorbis_quality = {CVAR_SAVE, "cl_capturevideo_ogg_vorbis_quality", "3", "audio quality (-1 to 10); higher is better"}; // ogg.h stuff +#ifdef _MSC_VER +typedef __int16 ogg_int16_t; +typedef unsigned __int16 ogg_uint16_t; +typedef __int32 ogg_int32_t; +typedef unsigned __int32 ogg_uint32_t; +typedef __int64 ogg_int64_t; +#else typedef int16_t ogg_int16_t; -typedef u_int16_t ogg_uint16_t; +typedef uint16_t ogg_uint16_t; typedef int32_t ogg_int32_t; -typedef u_int32_t ogg_uint32_t; +typedef uint32_t ogg_uint32_t; typedef int64_t ogg_int64_t; +#endif typedef struct { long endbyte; @@ -116,6 +127,7 @@ static int (*qogg_stream_flush) (ogg_stream_state *os, ogg_page *og); static int (*qogg_stream_init) (ogg_stream_state *os,int serialno); static int (*qogg_stream_clear) (ogg_stream_state *os); +static ogg_int64_t (*qogg_page_granulepos) (ogg_page *og); // end of ogg.h stuff @@ -265,6 +277,8 @@ static void (*qvorbis_comment_clear) (vorbis_comment *vc); static int (*qvorbis_block_init) (vorbis_dsp_state *v, vorbis_block *vb); static int (*qvorbis_block_clear) (vorbis_block *vb); static void (*qvorbis_dsp_clear) (vorbis_dsp_state *v); +static double (*qvorbis_granule_time) (vorbis_dsp_state *v, + ogg_int64_t granulepos); /* Vorbis PRIMITIVES: analysis/DSP layer ****************************/ @@ -331,7 +345,7 @@ typedef enum { OC_PF_420, /**< Chroma subsampling by 2 in each direction (4:2:0) */ OC_PF_RSVD, /**< Reserved value */ OC_PF_422, /**< Horizonatal chroma subsampling by 2 (4:2:2) */ - OC_PF_444, /**< No chroma subsampling at all (4:4:4) */ + OC_PF_444 /**< No chroma subsampling at all (4:4:4) */ } theora_pixelformat; /** * Theora bitstream info. @@ -442,6 +456,7 @@ static void (*qtheora_info_clear) (theora_info *c); static void (*qtheora_clear) (theora_state *t); static void (*qtheora_comment_init) (theora_comment *tc); static void (*qtheora_comment_clear) (theora_comment *tc); +static double (*qtheora_granule_time) (theora_state *th,ogg_int64_t granulepos); // end of theora.h stuff static dllfunction_t oggfuncs[] = @@ -451,6 +466,7 @@ static dllfunction_t oggfuncs[] = {"ogg_stream_flush", (void **) &qogg_stream_flush}, {"ogg_stream_init", (void **) &qogg_stream_init}, {"ogg_stream_clear", (void **) &qogg_stream_clear}, + {"ogg_page_granulepos", (void **) &qogg_page_granulepos}, {NULL, NULL} }; @@ -478,6 +494,7 @@ static dllfunction_t vorbisfuncs[] = {"vorbis_analysis", (void **) &qvorbis_analysis}, {"vorbis_bitrate_addblock", (void **) &qvorbis_bitrate_addblock}, {"vorbis_bitrate_flushpacket", (void **) &qvorbis_bitrate_flushpacket}, + {"vorbis_granule_time", (void **) &qvorbis_granule_time}, {NULL, NULL} }; @@ -494,18 +511,18 @@ static dllfunction_t theorafuncs[] = {"theora_encode_comment", (void **) &qtheora_encode_comment}, {"theora_encode_tables", (void **) &qtheora_encode_tables}, {"theora_clear", (void **) &qtheora_clear}, + {"theora_granule_time", (void **) &qtheora_granule_time}, {NULL, NULL} }; static dllhandle_t og_dll = NULL, vo_dll = NULL, ve_dll = NULL, th_dll = NULL; -qboolean SCR_CaptureVideo_Ogg_OpenLibrary() +qboolean SCR_CaptureVideo_Ogg_OpenLibrary(void) { const char* dllnames_og [] = { -#if defined(WIN64) - "libogg64.dll", -#elif defined(WIN32) +#if defined(WIN32) + "libogg-0.dll", "libogg.dll", "ogg.dll", #elif defined(MACOSX) @@ -518,9 +535,8 @@ qboolean SCR_CaptureVideo_Ogg_OpenLibrary() }; const char* dllnames_vo [] = { -#if defined(WIN64) - "libvorbis64.dll", -#elif defined(WIN32) +#if defined(WIN32) + "libvorbis-0.dll", "libvorbis.dll", "vorbis.dll", #elif defined(MACOSX) @@ -533,9 +549,8 @@ qboolean SCR_CaptureVideo_Ogg_OpenLibrary() }; const char* dllnames_ve [] = { -#if defined(WIN64) - "libvorbisenc64.dll", -#elif defined(WIN32) +#if defined(WIN32) + "libvorbisenc-2.dll", "libvorbisenc.dll", "vorbisenc.dll", #elif defined(MACOSX) @@ -548,9 +563,8 @@ qboolean SCR_CaptureVideo_Ogg_OpenLibrary() }; const char* dllnames_th [] = { -#if defined(WIN64) - "libtheora64.dll", -#elif defined(WIN32) +#if defined(WIN32) + "libtheora-0.dll", "libtheora.dll", "theora.dll", #elif defined(MACOSX) @@ -572,26 +586,26 @@ qboolean SCR_CaptureVideo_Ogg_OpenLibrary() Sys_LoadLibrary (dllnames_ve, &ve_dll, vorbisencfuncs); } -void SCR_CaptureVideo_Ogg_Init() +void SCR_CaptureVideo_Ogg_Init(void) { SCR_CaptureVideo_Ogg_OpenLibrary(); Cvar_RegisterVariable(&cl_capturevideo_ogg_theora_quality); Cvar_RegisterVariable(&cl_capturevideo_ogg_theora_bitrate); Cvar_RegisterVariable(&cl_capturevideo_ogg_theora_keyframe_bitrate_multiplier); - Cvar_RegisterVariable(&cl_capturevideo_ogg_theora_keyframe_frequency); - Cvar_RegisterVariable(&cl_capturevideo_ogg_theora_keyframe_mindistance); + Cvar_RegisterVariable(&cl_capturevideo_ogg_theora_keyframe_maxinterval); + Cvar_RegisterVariable(&cl_capturevideo_ogg_theora_keyframe_mininterval); Cvar_RegisterVariable(&cl_capturevideo_ogg_theora_keyframe_auto_threshold); Cvar_RegisterVariable(&cl_capturevideo_ogg_theora_noise_sensitivity); Cvar_RegisterVariable(&cl_capturevideo_ogg_vorbis_quality); } -qboolean SCR_CaptureVideo_Ogg_Available() +qboolean SCR_CaptureVideo_Ogg_Available(void) { return og_dll && th_dll && vo_dll && ve_dll; } -void SCR_CaptureVideo_Ogg_CloseDLL() +void SCR_CaptureVideo_Ogg_CloseDLL(void) { Sys_UnloadLibrary (&ve_dll); Sys_UnloadLibrary (&vo_dll); @@ -599,6 +613,21 @@ void SCR_CaptureVideo_Ogg_CloseDLL() Sys_UnloadLibrary (&og_dll); } +// this struct should not be needed +// however, libogg appears to pull the ogg_page's data element away from our +// feet before we get to write the data due to interleaving +// so this struct is used to keep the page data around until it actually gets +// written +typedef struct allocatedoggpage_s +{ + size_t len; + double time; + unsigned char data[65307]; + // this number is from RFC 3533. In case libogg writes more, we'll have to increase this + // but we'll get a Host_Error in this case so we can track it down +} +allocatedoggpage_t; + typedef struct capturevideostate_ogg_formatspecific_s { ogg_stream_state to, vo; @@ -607,22 +636,111 @@ typedef struct capturevideostate_ogg_formatspecific_s vorbis_dsp_state vd; vorbis_block vb; vorbis_info vi; - yuv_buffer yuv; + yuv_buffer yuv[2]; + int yuvi; + int lastnum; int channels; + + allocatedoggpage_t videopage, audiopage; } capturevideostate_ogg_formatspecific_t; #define LOAD_FORMATSPECIFIC_OGG() capturevideostate_ogg_formatspecific_t *format = (capturevideostate_ogg_formatspecific_t *) cls.capturevideo.formatspecific -static void SCR_CaptureVideo_Ogg_EndVideo() +static void SCR_CaptureVideo_Ogg_Interleave(void) +{ + LOAD_FORMATSPECIFIC_OGG(); + ogg_page pg; + + if(!cls.capturevideo.soundrate) + { + while(qogg_stream_pageout(&format->to, &pg) > 0) + { + FS_Write(cls.capturevideo.videofile, pg.header, pg.header_len); + FS_Write(cls.capturevideo.videofile, pg.body, pg.body_len); + } + return; + } + + for(;;) + { + // first: make sure we have a page of both types + if(!format->videopage.len) + if(qogg_stream_pageout(&format->to, &pg) > 0) + { + format->videopage.len = pg.header_len + pg.body_len; + format->videopage.time = qtheora_granule_time(&format->ts, qogg_page_granulepos(&pg)); + if(format->videopage.len > sizeof(format->videopage.data)) + Host_Error("video page too long"); + memcpy(format->videopage.data, pg.header, pg.header_len); + memcpy(format->videopage.data + pg.header_len, pg.body, pg.body_len); + } + if(!format->audiopage.len) + if(qogg_stream_pageout(&format->vo, &pg) > 0) + { + format->audiopage.len = pg.header_len + pg.body_len; + format->audiopage.time = qvorbis_granule_time(&format->vd, qogg_page_granulepos(&pg)); + if(format->audiopage.len > sizeof(format->audiopage.data)) + Host_Error("audio page too long"); + memcpy(format->audiopage.data, pg.header, pg.header_len); + memcpy(format->audiopage.data + pg.header_len, pg.body, pg.body_len); + } + + if(format->videopage.len && format->audiopage.len) + { + // output the page that ends first + if(format->videopage.time < format->audiopage.time) + { + FS_Write(cls.capturevideo.videofile, format->videopage.data, format->videopage.len); + format->videopage.len = 0; + } + else + { + FS_Write(cls.capturevideo.videofile, format->audiopage.data, format->audiopage.len); + format->audiopage.len = 0; + } + } + else + break; + } +} + +static void SCR_CaptureVideo_Ogg_FlushInterleaving(void) +{ + LOAD_FORMATSPECIFIC_OGG(); + + if(cls.capturevideo.soundrate) + if(format->audiopage.len) + { + FS_Write(cls.capturevideo.videofile, format->audiopage.data, format->audiopage.len); + format->audiopage.len = 0; + } + + if(format->videopage.len) + { + FS_Write(cls.capturevideo.videofile, format->videopage.data, format->videopage.len); + format->videopage.len = 0; + } +} + +static void SCR_CaptureVideo_Ogg_EndVideo(void) { LOAD_FORMATSPECIFIC_OGG(); ogg_page pg; ogg_packet pt; - // repeat the last frame so we can set the end-of-stream flag - qtheora_encode_YUVin(&format->ts, &format->yuv); - qtheora_encode_packetout(&format->ts, true, &pt); - qogg_stream_packetin(&format->to, &pt); + if(format->yuvi >= 0) + { + // send the previous (and last) frame + while(format->lastnum-- > 0) + { + qtheora_encode_YUVin(&format->ts, &format->yuv[format->yuvi]); + + while(qtheora_encode_packetout(&format->ts, !format->lastnum, &pt)) + qogg_stream_packetin(&format->to, &pt); + + SCR_CaptureVideo_Ogg_Interleave(); + } + } if(cls.capturevideo.soundrate) { @@ -633,9 +751,12 @@ static void SCR_CaptureVideo_Ogg_EndVideo() qvorbis_bitrate_addblock(&format->vb); while(qvorbis_bitrate_flushpacket(&format->vd, &pt)) qogg_stream_packetin(&format->vo, &pt); + SCR_CaptureVideo_Ogg_Interleave(); } } + SCR_CaptureVideo_Ogg_FlushInterleaving(); + while(qogg_stream_pageout(&format->to, &pg) > 0) { FS_Write(cls.capturevideo.videofile, pg.header, pg.header_len); @@ -682,18 +803,22 @@ static void SCR_CaptureVideo_Ogg_EndVideo() qtheora_clear(&format->ts); qvorbis_info_clear(&format->vi); - Mem_Free(format->yuv.y); - Mem_Free(format->yuv.u); - Mem_Free(format->yuv.v); + Mem_Free(format->yuv[0].y); + Mem_Free(format->yuv[0].u); + Mem_Free(format->yuv[0].v); + Mem_Free(format->yuv[1].y); + Mem_Free(format->yuv[1].u); + Mem_Free(format->yuv[1].v); Mem_Free(format); FS_Close(cls.capturevideo.videofile); cls.capturevideo.videofile = NULL; } -static void SCR_CaptureVideo_Ogg_ConvertFrame_BGRA_to_YUV() +static void SCR_CaptureVideo_Ogg_ConvertFrame_BGRA_to_YUV(void) { LOAD_FORMATSPECIFIC_OGG(); + yuv_buffer *yuv; int x, y; int blockr, blockg, blockb; unsigned char *b = cls.capturevideo.outbuffer; @@ -701,6 +826,8 @@ static void SCR_CaptureVideo_Ogg_ConvertFrame_BGRA_to_YUV() int h = cls.capturevideo.height; int inpitch = w*4; + yuv = &format->yuv[format->yuvi]; + for(y = 0; y < h; ++y) { for(b = cls.capturevideo.outbuffer + (h-1-y)*w*4, x = 0; x < w; ++x) @@ -708,7 +835,7 @@ static void SCR_CaptureVideo_Ogg_ConvertFrame_BGRA_to_YUV() blockr = b[2]; blockg = b[1]; blockb = b[0]; - format->yuv.y[x + format->yuv.y_stride * y] = + yuv->y[x + yuv->y_stride * y] = cls.capturevideo.yuvnormalizetable[0][cls.capturevideo.rgbtoyuvscaletable[0][0][blockr] + cls.capturevideo.rgbtoyuvscaletable[0][1][blockg] + cls.capturevideo.rgbtoyuvscaletable[0][2][blockb]]; b += 4; } @@ -720,9 +847,9 @@ static void SCR_CaptureVideo_Ogg_ConvertFrame_BGRA_to_YUV() blockr = (b[2] + b[6] + b[inpitch+2] + b[inpitch+6]) >> 2; blockg = (b[1] + b[5] + b[inpitch+1] + b[inpitch+5]) >> 2; blockb = (b[0] + b[4] + b[inpitch+0] + b[inpitch+4]) >> 2; - format->yuv.u[x + format->yuv.uv_stride * (y/2)] = + yuv->u[x + yuv->uv_stride * (y/2)] = cls.capturevideo.yuvnormalizetable[1][cls.capturevideo.rgbtoyuvscaletable[1][0][blockr] + cls.capturevideo.rgbtoyuvscaletable[1][1][blockg] + cls.capturevideo.rgbtoyuvscaletable[1][2][blockb] + 128]; - format->yuv.v[x + format->yuv.uv_stride * (y/2)] = + yuv->v[x + yuv->uv_stride * (y/2)] = cls.capturevideo.yuvnormalizetable[2][cls.capturevideo.rgbtoyuvscaletable[2][0][blockr] + cls.capturevideo.rgbtoyuvscaletable[2][1][blockg] + cls.capturevideo.rgbtoyuvscaletable[2][2][blockb] + 128]; b += 8; } @@ -733,41 +860,59 @@ static void SCR_CaptureVideo_Ogg_ConvertFrame_BGRA_to_YUV() static void SCR_CaptureVideo_Ogg_VideoFrames(int num) { LOAD_FORMATSPECIFIC_OGG(); - ogg_page pg; ogg_packet pt; // data is in cls.capturevideo.outbuffer as BGRA and has size width*height - SCR_CaptureVideo_Ogg_ConvertFrame_BGRA_to_YUV(); - - while(num-- > 0) + if(format->yuvi >= 0) { - qtheora_encode_YUVin(&format->ts, &format->yuv); - qtheora_encode_packetout(&format->ts, false, &pt); - qogg_stream_packetin(&format->to, &pt); - - while(qogg_stream_pageout(&format->to, &pg) > 0) + // send the previous frame + while(format->lastnum-- > 0) { - FS_Write(cls.capturevideo.videofile, pg.header, pg.header_len); - FS_Write(cls.capturevideo.videofile, pg.body, pg.body_len); + qtheora_encode_YUVin(&format->ts, &format->yuv[format->yuvi]); + + while(qtheora_encode_packetout(&format->ts, false, &pt)) + qogg_stream_packetin(&format->to, &pt); + + SCR_CaptureVideo_Ogg_Interleave(); } } + + format->yuvi = (format->yuvi + 1) % 2; + SCR_CaptureVideo_Ogg_ConvertFrame_BGRA_to_YUV(); + format->lastnum = num; + + // TODO maybe send num-1 frames from here already } +typedef int channelmapping_t[8]; +channelmapping_t mapping[8] = +{ + { 0, -1, -1, -1, -1, -1, -1, -1 }, // mono + { 0, 1, -1, -1, -1, -1, -1, -1 }, // stereo + { 0, 1, 2, -1, -1, -1, -1, -1 }, // L C R + { 0, 1, 2, 3, -1, -1, -1, -1 }, // surround40 + { 0, 2, 3, 4, 1, -1, -1, -1 }, // FL FC FR RL RR + { 0, 2, 3, 4, 1, 5, -1, -1 }, // surround51 + { 0, 2, 3, 4, 1, 5, 6, -1 }, // (not defined by vorbis spec) + { 0, 2, 3, 4, 1, 5, 6, 7 } // surround71 (not defined by vorbis spec) +}; + static void SCR_CaptureVideo_Ogg_SoundFrame(const portable_sampleframe_t *paintbuffer, size_t length) { LOAD_FORMATSPECIFIC_OGG(); float **vorbis_buffer; size_t i; int j; - ogg_page pg; ogg_packet pt; + int *map = mapping[bound(1, cls.capturevideo.soundchannels, 8) - 1]; vorbis_buffer = qvorbis_analysis_buffer(&format->vd, length); - for(i = 0; i < length; ++i) + for(j = 0; j < cls.capturevideo.soundchannels; ++j) { - for(j = 0; j < cls.capturevideo.soundchannels; ++j) - vorbis_buffer[j][i] = paintbuffer[i].sample[j] / 32768.0f; + float *b = vorbis_buffer[map[j]]; + for(i = 0; i < length; ++i) + b[i] = paintbuffer[i].sample[j] / 32768.0f; } qvorbis_analysis_wrote(&format->vd, length); @@ -780,24 +925,21 @@ static void SCR_CaptureVideo_Ogg_SoundFrame(const portable_sampleframe_t *paintb qogg_stream_packetin(&format->vo, &pt); } - while(qogg_stream_pageout(&format->vo, &pg) > 0) - { - FS_Write(cls.capturevideo.videofile, pg.header, pg.header_len); - FS_Write(cls.capturevideo.videofile, pg.body, pg.body_len); - } + SCR_CaptureVideo_Ogg_Interleave(); } -void SCR_CaptureVideo_Ogg_BeginVideo() +void SCR_CaptureVideo_Ogg_BeginVideo(void) { cls.capturevideo.format = CAPTUREVIDEOFORMAT_OGG_VORBIS_THEORA; - cls.capturevideo.videofile = FS_OpenRealFile(va("%s.ogv", cls.capturevideo.basename), "wb", false); + cls.capturevideo.formatextension = "ogv"; + cls.capturevideo.videofile = FS_OpenRealFile(va("%s.%s", cls.capturevideo.basename, cls.capturevideo.formatextension), "wb", false); cls.capturevideo.endvideo = SCR_CaptureVideo_Ogg_EndVideo; cls.capturevideo.videoframes = SCR_CaptureVideo_Ogg_VideoFrames; cls.capturevideo.soundframe = SCR_CaptureVideo_Ogg_SoundFrame; cls.capturevideo.formatspecific = Mem_Alloc(tempmempool, sizeof(capturevideostate_ogg_formatspecific_t)); { LOAD_FORMATSPECIFIC_OGG(); - int num, denom; + int num, denom, i; ogg_page pg; ogg_packet pt, pt2, pt3; theora_comment tc; @@ -817,6 +959,8 @@ void SCR_CaptureVideo_Ogg_BeginVideo() qogg_stream_init(&format->vo, format->serial2); } + format->videopage.len = format->audiopage.len = 0; + qtheora_info_init(&ti); ti.frame_width = cls.capturevideo.width; ti.frame_height = cls.capturevideo.height; @@ -825,19 +969,21 @@ void SCR_CaptureVideo_Ogg_BeginVideo() //ti.offset_x = ((ti.width - ti.frame_width) / 2) & ~1; //ti.offset_y = ((ti.height - ti.frame_height) / 2) & ~1; - format->yuv.y_width = ti.width; - format->yuv.y_height = ti.height; - format->yuv.y_stride = ti.width; - - format->yuv.uv_width = ti.width / 2; - format->yuv.uv_height = ti.height / 2; - format->yuv.uv_stride = ti.width / 2; - - format->yuv.y = Mem_Alloc(tempmempool, format->yuv.y_stride * format->yuv.y_height); - format->yuv.u = Mem_Alloc(tempmempool, format->yuv.uv_stride * format->yuv.uv_height); - format->yuv.v = Mem_Alloc(tempmempool, format->yuv.uv_stride * format->yuv.uv_height); + for(i = 0; i < 2; ++i) + { + format->yuv[i].y_width = ti.width; + format->yuv[i].y_height = ti.height; + format->yuv[i].y_stride = ti.width; + format->yuv[i].uv_width = ti.width / 2; + format->yuv[i].uv_height = ti.height / 2; + format->yuv[i].uv_stride = ti.width / 2; + format->yuv[i].y = (unsigned char *) Mem_Alloc(tempmempool, format->yuv[i].y_stride * format->yuv[i].y_height); + format->yuv[i].u = (unsigned char *) Mem_Alloc(tempmempool, format->yuv[i].uv_stride * format->yuv[i].uv_height); + format->yuv[i].v = (unsigned char *) Mem_Alloc(tempmempool, format->yuv[i].uv_stride * format->yuv[i].uv_height); + } + format->yuvi = -1; // -1: no frame valid yet, write into 0 - FindFraction(cls.capturevideo.framerate, &num, &denom, 1001); + FindFraction(cls.capturevideo.framerate / cls.capturevideo.framestep, &num, &denom, 1001); ti.fps_numerator = num; ti.fps_denominator = denom; @@ -859,13 +1005,13 @@ void SCR_CaptureVideo_Ogg_BeginVideo() if(ti.quality < 0) { ti.target_bitrate = -1; - ti.keyframe_data_target_bitrate = -1; + ti.keyframe_data_target_bitrate = (unsigned int)-1; ti.quality = 63; } else { ti.target_bitrate = -1; - ti.keyframe_data_target_bitrate = -1; + ti.keyframe_data_target_bitrate = (unsigned int)-1; ti.quality = bound(0, ti.quality, 63); } } @@ -874,25 +1020,26 @@ void SCR_CaptureVideo_Ogg_BeginVideo() if(ti.quality < 0) { ti.target_bitrate = bound(45000, ti.target_bitrate, 2000000); - ti.keyframe_data_target_bitrate = ti.target_bitrate * max(1, cl_capturevideo_ogg_theora_keyframe_bitrate_multiplier.value); + ti.keyframe_data_target_bitrate = (int) (ti.target_bitrate * max(1, cl_capturevideo_ogg_theora_keyframe_bitrate_multiplier.value)); ti.quality = -1; } else { ti.target_bitrate = bound(45000, ti.target_bitrate, 2000000); - ti.keyframe_data_target_bitrate = ti.target_bitrate * max(1, cl_capturevideo_ogg_theora_keyframe_bitrate_multiplier.value); + ti.keyframe_data_target_bitrate = (int) (ti.target_bitrate * max(1, cl_capturevideo_ogg_theora_keyframe_bitrate_multiplier.value)); ti.quality = -1; } } - ti.keyframe_frequency = bound(1, cl_capturevideo_ogg_theora_keyframe_frequency.integer, 1000); - ti.keyframe_mindistance = bound(1, cl_capturevideo_ogg_theora_keyframe_mindistance.integer, (int) ti.keyframe_frequency); + // this -1 magic is because ti.keyframe_frequency and ti.keyframe_mindistance use different metrics + ti.keyframe_frequency = bound(1, cl_capturevideo_ogg_theora_keyframe_maxinterval.integer, 1000); + ti.keyframe_mindistance = bound(1, cl_capturevideo_ogg_theora_keyframe_mininterval.integer, (int) ti.keyframe_frequency) - 1; ti.noise_sensitivity = bound(0, cl_capturevideo_ogg_theora_noise_sensitivity.integer, 6); ti.sharpness = bound(0, cl_capturevideo_ogg_theora_sharpness.integer, 2); ti.keyframe_auto_threshold = bound(0, cl_capturevideo_ogg_theora_keyframe_auto_threshold.integer, 100); ti.keyframe_frequency_force = ti.keyframe_frequency; - ti.keyframe_auto_p = (ti.keyframe_frequency != ti.keyframe_mindistance); + ti.keyframe_auto_p = (ti.keyframe_frequency != ti.keyframe_mindistance + 1); qtheora_encode_init(&format->ts, &ti); qtheora_info_clear(&ti);