static cvar_t cl_capturevideo_ogg_theora_keyframe_auto_threshold = {CVAR_SAVE, "cl_capturevideo_ogg_theora_keyframe_auto_threshold", "80", "threshold for key frame decision (0 to 100)"};
static cvar_t cl_capturevideo_ogg_theora_noise_sensitivity = {CVAR_SAVE, "cl_capturevideo_ogg_theora_noise_sensitivity", "1", "video noise sensitivity (0 to 6); lower is better"};
static cvar_t cl_capturevideo_ogg_theora_sharpness = {CVAR_SAVE, "cl_capturevideo_ogg_theora_sharpness", "0", "sharpness (0 to 2); lower is sharper"};
-static cvar_t cl_capturevideo_ogg_vorbis_quality = {CVAR_SAVE, "cl_capturevideo_ogg_vorbis_quality", "1", "audio quality (-1 to 10); higher is better"};
+static cvar_t cl_capturevideo_ogg_vorbis_quality = {CVAR_SAVE, "cl_capturevideo_ogg_vorbis_quality", "3", "audio quality (-1 to 10); higher is better"};
// ogg.h stuff
#ifdef _MSC_VER
static dllhandle_t og_dll = NULL, vo_dll = NULL, ve_dll = NULL, th_dll = NULL;
-qboolean SCR_CaptureVideo_Ogg_OpenLibrary()
+qboolean SCR_CaptureVideo_Ogg_OpenLibrary(void)
{
const char* dllnames_og [] =
{
Sys_LoadLibrary (dllnames_ve, &ve_dll, vorbisencfuncs);
}
-void SCR_CaptureVideo_Ogg_Init()
+void SCR_CaptureVideo_Ogg_Init(void)
{
SCR_CaptureVideo_Ogg_OpenLibrary();
Cvar_RegisterVariable(&cl_capturevideo_ogg_vorbis_quality);
}
-qboolean SCR_CaptureVideo_Ogg_Available()
+qboolean SCR_CaptureVideo_Ogg_Available(void)
{
return og_dll && th_dll && vo_dll && ve_dll;
}
-void SCR_CaptureVideo_Ogg_CloseDLL()
+void SCR_CaptureVideo_Ogg_CloseDLL(void)
{
Sys_UnloadLibrary (&ve_dll);
Sys_UnloadLibrary (&vo_dll);
capturevideostate_ogg_formatspecific_t;
#define LOAD_FORMATSPECIFIC_OGG() capturevideostate_ogg_formatspecific_t *format = (capturevideostate_ogg_formatspecific_t *) cls.capturevideo.formatspecific
-static void SCR_CaptureVideo_Ogg_Interleave()
+static void SCR_CaptureVideo_Ogg_Interleave(void)
{
LOAD_FORMATSPECIFIC_OGG();
ogg_page pg;
}
}
-static void SCR_CaptureVideo_Ogg_FlushInterleaving()
+static void SCR_CaptureVideo_Ogg_FlushInterleaving(void)
{
LOAD_FORMATSPECIFIC_OGG();
}
}
-static void SCR_CaptureVideo_Ogg_EndVideo()
+static void SCR_CaptureVideo_Ogg_EndVideo(void)
{
LOAD_FORMATSPECIFIC_OGG();
ogg_page pg;
cls.capturevideo.videofile = NULL;
}
-static void SCR_CaptureVideo_Ogg_ConvertFrame_BGRA_to_YUV()
+static void SCR_CaptureVideo_Ogg_ConvertFrame_BGRA_to_YUV(void)
{
LOAD_FORMATSPECIFIC_OGG();
yuv_buffer *yuv;
// TODO maybe send num-1 frames from here already
}
+typedef int channelmapping_t[8];
+channelmapping_t mapping[8] =
+{
+ { 0, -1, -1, -1, -1, -1, -1, -1 }, // mono
+ { 0, 1, -1, -1, -1, -1, -1, -1 }, // stereo
+ { 0, 1, 2, -1, -1, -1, -1, -1 }, // L C R
+ { 0, 1, 2, 3, -1, -1, -1, -1 }, // surround40
+ { 0, 2, 3, 4, 1, -1, -1, -1 }, // FL FC FR RL RR
+ { 0, 2, 3, 4, 1, 5, -1, -1 }, // surround51
+ { 0, 2, 3, 4, 1, 5, 6, -1 }, // (not defined by vorbis spec)
+ { 0, 2, 3, 4, 1, 5, 6, 7 } // surround71 (not defined by vorbis spec)
+};
+
static void SCR_CaptureVideo_Ogg_SoundFrame(const portable_sampleframe_t *paintbuffer, size_t length)
{
LOAD_FORMATSPECIFIC_OGG();
size_t i;
int j;
ogg_packet pt;
+ int *map = mapping[bound(1, cls.capturevideo.soundchannels, 8) - 1];
vorbis_buffer = qvorbis_analysis_buffer(&format->vd, length);
- for(i = 0; i < length; ++i)
+ for(j = 0; j < cls.capturevideo.soundchannels; ++j)
{
- for(j = 0; j < cls.capturevideo.soundchannels; ++j)
- vorbis_buffer[j][i] = paintbuffer[i].sample[j] / 32768.0f;
+ float *b = vorbis_buffer[map[j]];
+ for(i = 0; i < length; ++i)
+ b[i] = paintbuffer[i].sample[j] / 32768.0f;
}
qvorbis_analysis_wrote(&format->vd, length);
SCR_CaptureVideo_Ogg_Interleave();
}
-void SCR_CaptureVideo_Ogg_BeginVideo()
+void SCR_CaptureVideo_Ogg_BeginVideo(void)
{
cls.capturevideo.format = CAPTUREVIDEOFORMAT_OGG_VORBIS_THEORA;
cls.capturevideo.formatextension = "ogv";
format->yuv[i].uv_width = ti.width / 2;
format->yuv[i].uv_height = ti.height / 2;
format->yuv[i].uv_stride = ti.width / 2;
- format->yuv[i].y = Mem_Alloc(tempmempool, format->yuv[i].y_stride * format->yuv[i].y_height);
- format->yuv[i].u = Mem_Alloc(tempmempool, format->yuv[i].uv_stride * format->yuv[i].uv_height);
- format->yuv[i].v = Mem_Alloc(tempmempool, format->yuv[i].uv_stride * format->yuv[i].uv_height);
+ format->yuv[i].y = (unsigned char *) Mem_Alloc(tempmempool, format->yuv[i].y_stride * format->yuv[i].y_height);
+ format->yuv[i].u = (unsigned char *) Mem_Alloc(tempmempool, format->yuv[i].uv_stride * format->yuv[i].uv_height);
+ format->yuv[i].v = (unsigned char *) Mem_Alloc(tempmempool, format->yuv[i].uv_stride * format->yuv[i].uv_height);
}
format->yuvi = -1; // -1: no frame valid yet, write into 0
- FindFraction(cls.capturevideo.framerate, &num, &denom, 1001);
+ FindFraction(cls.capturevideo.framerate / cls.capturevideo.framestep, &num, &denom, 1001);
ti.fps_numerator = num;
ti.fps_denominator = denom;
if(ti.quality < 0)
{
ti.target_bitrate = -1;
- ti.keyframe_data_target_bitrate = -1;
+ ti.keyframe_data_target_bitrate = (unsigned int)-1;
ti.quality = 63;
}
else
{
ti.target_bitrate = -1;
- ti.keyframe_data_target_bitrate = -1;
+ ti.keyframe_data_target_bitrate = (unsigned int)-1;
ti.quality = bound(0, ti.quality, 63);
}
}
if(ti.quality < 0)
{
ti.target_bitrate = bound(45000, ti.target_bitrate, 2000000);
- ti.keyframe_data_target_bitrate = ti.target_bitrate * max(1, cl_capturevideo_ogg_theora_keyframe_bitrate_multiplier.value);
+ ti.keyframe_data_target_bitrate = (int) (ti.target_bitrate * max(1, cl_capturevideo_ogg_theora_keyframe_bitrate_multiplier.value));
ti.quality = -1;
}
else
{
ti.target_bitrate = bound(45000, ti.target_bitrate, 2000000);
- ti.keyframe_data_target_bitrate = ti.target_bitrate * max(1, cl_capturevideo_ogg_theora_keyframe_bitrate_multiplier.value);
+ ti.keyframe_data_target_bitrate = (int) (ti.target_bitrate * max(1, cl_capturevideo_ogg_theora_keyframe_bitrate_multiplier.value));
ti.quality = -1;
}
}