godot/drivers/theora/video_stream_theora.cpp
2014-02-09 22:10:30 -03:00

681 lines
16 KiB
C++

#ifdef THEORA_ENABLED
#include "video_stream_theora.h"
#include "os/os.h"
#include "yuv2rgb.h"
AudioStream::UpdateMode VideoStreamTheora::get_update_mode() const {
return UPDATE_IDLE;
};
int VideoStreamTheora:: buffer_data() {
char *buffer=ogg_sync_buffer(&oy,4096);
int bytes=file->get_buffer((uint8_t*)buffer, 4096);
ogg_sync_wrote(&oy,bytes);
return(bytes);
}
int VideoStreamTheora::queue_page(ogg_page *page){
if(theora_p)ogg_stream_pagein(&to,page);
if(vorbis_p)ogg_stream_pagein(&vo,page);
return 0;
}
Image VideoStreamTheora::peek_frame() const {
if (frames_pending == 0)
return Image();
return Image(size.x, size.y, 0, format, frame_data);
};
Image VideoStreamTheora::pop_frame() {
Image ret = peek_frame();
frames_pending = 0;
return ret;
};
int VideoStreamTheora::get_pending_frame_count() const {
return frames_pending;
};
void VideoStreamTheora::video_write(void){
th_ycbcr_buffer yuv;
int y_offset, uv_offset;
th_decode_ycbcr_out(td,yuv);
y_offset=(ti.pic_x&~1)+yuv[0].stride*(ti.pic_y&~1);
/*
{
int pixels = size.x * size.y;
frame_data.resize(pixels * 4);
DVector<uint8_t>::Write w = frame_data.write();
char* dst = (char*)w.ptr();
int p = 0;
for (int i=0; i<size.y; i++) {
char *in_y = (char *)yuv[0].data+y_offset+yuv[0].stride*i;
char *out = dst + (int)size.x * 4 * i;
for (int j=0;j<size.x;j++) {
dst[p++] = in_y[j];
dst[p++] = in_y[j];
dst[p++] = in_y[j];
dst[p++] = 255;
};
}
format = Image::FORMAT_RGBA;
}
// */
//*
int pitch = 4;
frame_data.resize(size.x * size.y * pitch);
DVector<uint8_t>::Write w = frame_data.write();
char* dst = (char*)w.ptr();
uv_offset=(ti.pic_x/2)+(yuv[1].stride)*(ti.pic_y/2);
if (px_fmt == TH_PF_444) {
yuv444_2_rgb8888((uint8_t*)dst, (uint8_t*)yuv[0].data, (uint8_t*)yuv[1].data, (uint8_t*)yuv[2].data, size.x, size.y, yuv[0].stride, yuv[1].stride, size.x<<2, 0);
} else if (px_fmt == TH_PF_422) {
yuv422_2_rgb8888((uint8_t*)dst, (uint8_t*)yuv[0].data, (uint8_t*)yuv[1].data, (uint8_t*)yuv[2].data, size.x, size.y, yuv[0].stride, yuv[1].stride, size.x<<2, 0);
} else if (px_fmt == TH_PF_420) {
yuv420_2_rgb8888((uint8_t*)dst, (uint8_t*)yuv[0].data, (uint8_t*)yuv[2].data, (uint8_t*)yuv[1].data, size.x, size.y, yuv[0].stride, yuv[1].stride, size.x<<2, 0);
};
format = Image::FORMAT_RGBA;
/*
if (px_fmt == TH_PF_444) {
int pitch = 3;
frame_data.resize(size.x * size.y * pitch);
DVector<uint8_t>::Write w = frame_data.write();
char* dst = (char*)w.ptr();
for(int i=0;i<size.y;i++) {
char *in_y = (char *)yuv[0].data+y_offset+yuv[0].stride*i;
char *out = dst + (int)size.x * pitch * i;
char *in_u = (char *)yuv[1].data+uv_offset+yuv[1].stride*i;
char *in_v = (char *)yuv[2].data+uv_offset+yuv[2].stride*i;
for (int j=0;j<size.x;j++) {
out[j*3+0] = in_y[j];
out[j*3+1] = in_u[j];
out[j*3+2] = in_v[j];
};
}
format = Image::FORMAT_YUV_444;
} else {
int div;
if (px_fmt!=TH_PF_422) {
div = 2;
}
bool rgba = true;
if (rgba) {
int pitch = 4;
frame_data.resize(size.x * size.y * pitch);
DVector<uint8_t>::Write w = frame_data.write();
char* dst = (char*)w.ptr();
uv_offset=(ti.pic_x/2)+(yuv[1].stride)*(ti.pic_y / div);
for(int i=0;i<size.y;i++) {
char *in_y = (char *)yuv[0].data+y_offset+yuv[0].stride*i;
char *in_u = (char *)yuv[1].data+uv_offset+yuv[1].stride*(i/div);
char *in_v = (char *)yuv[2].data+uv_offset+yuv[2].stride*(i/div);
uint8_t *out = (uint8_t*)dst + (int)size.x * pitch * i;
int ofs = 0;
for (int j=0;j<size.x;j++) {
uint8_t y, u, v;
y = in_y[j];
u = in_u[j/2];
v = in_v[j/2];
int32_t r = Math::fast_ftoi(1.164 * (y - 16) + 1.596 * (v - 128));
int32_t g = Math::fast_ftoi(1.164 * (y - 16) - 0.813 * (v - 128) - 0.391 * (u - 128));
int32_t b = Math::fast_ftoi(1.164 * (y - 16) + 2.018 * (u - 128));
out[ofs++] = CLAMP(r, 0, 255);
out[ofs++] = CLAMP(g, 0, 255);
out[ofs++] = CLAMP(b, 0, 255);
out[ofs++] = 255;
}
}
format = Image::FORMAT_RGBA;
} else {
int pitch = 2;
frame_data.resize(size.x * size.y * pitch);
DVector<uint8_t>::Write w = frame_data.write();
char* dst = (char*)w.ptr();
uv_offset=(ti.pic_x/2)+(yuv[1].stride)*(ti.pic_y / div);
for(int i=0;i<size.y;i++) {
char *in_y = (char *)yuv[0].data+y_offset+yuv[0].stride*i;
char *out = dst + (int)size.x * pitch * i;
for (int j=0;j<size.x;j++)
out[j*2] = in_y[j];
char *in_u = (char *)yuv[1].data+uv_offset+yuv[1].stride*(i/div);
char *in_v = (char *)yuv[2].data+uv_offset+yuv[2].stride*(i/div);
for (int j=0;j<(int)size.x>>1;j++) {
out[j*4+1] = in_u[j];
out[j*4+3] = in_v[j];
}
}
format = Image::FORMAT_YUV_422;
};
};
// */
frames_pending = 1;
}
void VideoStreamTheora::clear() {
if (file_name == "")
return;
if(vorbis_p){
ogg_stream_clear(&vo);
if (vorbis_p >= 3) {
vorbis_block_clear(&vb);
vorbis_dsp_clear(&vd);
};
vorbis_comment_clear(&vc);
vorbis_info_clear(&vi);
vorbis_p = 0;
}
if(theora_p){
ogg_stream_clear(&to);
th_decode_free(td);
th_comment_clear(&tc);
th_info_clear(&ti);
theora_p = 0;
}
ogg_sync_clear(&oy);
file_name = "";
theora_p = 0;
vorbis_p = 0;
videobuf_ready = 0;
frames_pending = 0;
videobuf_time = 0;
playing = false;
};
void VideoStreamTheora::set_file(const String& p_file) {
ogg_packet op;
th_setup_info *ts = NULL;
file_name = p_file;
if (file) {
memdelete(file);
}
file = FileAccess::open(p_file, FileAccess::READ);
ERR_FAIL_COND(!file);
audio_frames_wrote = 0;
ogg_sync_init(&oy);
/* init supporting Vorbis structures needed in header parsing */
vorbis_info_init(&vi);
vorbis_comment_init(&vc);
/* init supporting Theora structures needed in header parsing */
th_comment_init(&tc);
th_info_init(&ti);
/* Ogg file open; parse the headers */
/* Only interested in Vorbis/Theora streams */
int stateflag = 0;
while(!stateflag){
int ret=buffer_data();
if(ret==0)break;
while(ogg_sync_pageout(&oy,&og)>0){
ogg_stream_state test;
/* is this a mandated initial header? If not, stop parsing */
if(!ogg_page_bos(&og)){
/* don't leak the page; get it into the appropriate stream */
queue_page(&og);
stateflag=1;
break;
}
ogg_stream_init(&test,ogg_page_serialno(&og));
ogg_stream_pagein(&test,&og);
ogg_stream_packetout(&test,&op);
/* identify the codec: try theora */
if(!theora_p && th_decode_headerin(&ti,&tc,&ts,&op)>=0){
/* it is theora */
copymem(&to,&test,sizeof(test));
theora_p=1;
}else if(!vorbis_p && vorbis_synthesis_headerin(&vi,&vc,&op)>=0){
/* it is vorbis */
copymem(&vo,&test,sizeof(test));
vorbis_p=1;
}else{
/* whatever it is, we don't care about it */
ogg_stream_clear(&test);
}
}
/* fall through to non-bos page parsing */
}
/* we're expecting more header packets. */
while((theora_p && theora_p<3) || (vorbis_p && vorbis_p<3)){
int ret;
/* look for further theora headers */
while(theora_p && (theora_p<3) && (ret=ogg_stream_packetout(&to,&op))){
if(ret<0){
fprintf(stderr,"Error parsing Theora stream headers; "
"corrupt stream?\n");
clear();
return;
}
if(!th_decode_headerin(&ti,&tc,&ts,&op)){
fprintf(stderr,"Error parsing Theora stream headers; "
"corrupt stream?\n");
clear();
return;
}
theora_p++;
}
/* look for more vorbis header packets */
while(vorbis_p && (vorbis_p<3) && (ret=ogg_stream_packetout(&vo,&op))){
if(ret<0){
fprintf(stderr,"Error parsing Vorbis stream headers; corrupt stream?\n");
clear();
return;
}
ret = vorbis_synthesis_headerin(&vi,&vc,&op);
if(ret){
fprintf(stderr,"Error parsing Vorbis stream headers; corrupt stream?\n");
clear();
return;
}
vorbis_p++;
if(vorbis_p==3)break;
}
/* The header pages/packets will arrive before anything else we
care about, or the stream is not obeying spec */
if(ogg_sync_pageout(&oy,&og)>0){
queue_page(&og); /* demux into the appropriate stream */
}else{
int ret=buffer_data(); /* someone needs more data */
if(ret==0){
fprintf(stderr,"End of file while searching for codec headers.\n");
clear();
return;
}
}
}
/* and now we have it all. initialize decoders */
if(theora_p){
td=th_decode_alloc(&ti,ts);
printf("Ogg logical stream %lx is Theora %dx%d %.02f fps",
to.serialno,ti.pic_width,ti.pic_height,
(double)ti.fps_numerator/ti.fps_denominator);
px_fmt=ti.pixel_fmt;
switch(ti.pixel_fmt){
case TH_PF_420: printf(" 4:2:0 video\n"); break;
case TH_PF_422: printf(" 4:2:2 video\n"); break;
case TH_PF_444: printf(" 4:4:4 video\n"); break;
case TH_PF_RSVD:
default:
printf(" video\n (UNKNOWN Chroma sampling!)\n");
break;
}
if(ti.pic_width!=ti.frame_width || ti.pic_height!=ti.frame_height)
printf(" Frame content is %dx%d with offset (%d,%d).\n",
ti.frame_width, ti.frame_height, ti.pic_x, ti.pic_y);
th_decode_ctl(td,TH_DECCTL_GET_PPLEVEL_MAX,&pp_level_max,
sizeof(pp_level_max));
pp_level=pp_level_max;
th_decode_ctl(td,TH_DECCTL_SET_PPLEVEL,&pp_level,sizeof(pp_level));
pp_inc=0;
/*{
int arg = 0xffff;
th_decode_ctl(td,TH_DECCTL_SET_TELEMETRY_MBMODE,&arg,sizeof(arg));
th_decode_ctl(td,TH_DECCTL_SET_TELEMETRY_MV,&arg,sizeof(arg));
th_decode_ctl(td,TH_DECCTL_SET_TELEMETRY_QI,&arg,sizeof(arg));
arg=10;
th_decode_ctl(td,TH_DECCTL_SET_TELEMETRY_BITS,&arg,sizeof(arg));
}*/
int w;
int h;
w=(ti.pic_x+ti.frame_width+1&~1)-(ti.pic_x&~1);
h=(ti.pic_y+ti.frame_height+1&~1)-(ti.pic_y&~1);
size.x = w;
size.y = h;
}else{
/* tear down the partial theora setup */
th_info_clear(&ti);
th_comment_clear(&tc);
}
th_setup_free(ts);
if(vorbis_p){
vorbis_synthesis_init(&vd,&vi);
vorbis_block_init(&vd,&vb);
fprintf(stderr,"Ogg logical stream %lx is Vorbis %d channel %ld Hz audio.\n",
vo.serialno,vi.channels,vi.rate);
_setup(vi.channels, vi.rate);
}else{
/* tear down the partial vorbis setup */
vorbis_info_clear(&vi);
vorbis_comment_clear(&vc);
}
playing = false;
buffering=true;
time=0;
};
float VideoStreamTheora::get_time() const {
//print_line("total: "+itos(get_total())+" todo: "+itos(get_todo()));
//return MAX(0,time-((get_total())/(float)vi.rate));
return time-((get_total())/(float)vi.rate);
};
void VideoStreamTheora::update() {
if (!playing) {
//printf("not playing\n");
return;
};
double ctime =AudioServer::get_singleton()->get_mix_time();
if (last_update_time) {
double delta = (ctime-last_update_time);
time+=delta;
//print_line("delta: "+rtos(delta));
}
last_update_time=ctime;
int audio_todo = get_todo();
ogg_packet op;
int audio_pending = 0;
while (vorbis_p && audio_todo) {
int ret;
float **pcm;
/* if there's pending, decoded audio, grab it */
if ((ret=vorbis_synthesis_pcmout(&vd,&pcm))>0) {
audio_pending = ret;
int16_t* out = get_write_buffer();
int count = 0;
int to_read = MIN(ret, audio_todo);
for (int i=0; i<to_read; i++) {
for(int j=0;j<vi.channels;j++){
int val=Math::fast_ftoi(pcm[j][i]*32767.f);
if(val>32767)val=32767;
if(val<-32768)val=-32768;
out[count++] = val;
};
};
int tr = vorbis_synthesis_read(&vd, to_read);
audio_todo -= to_read;
audio_frames_wrote += to_read;
write(to_read);
audio_pending -= to_read;
if (audio_todo==0)
buffering=false;
} else {
/* no pending audio; is there a pending packet to decode? */
if (ogg_stream_packetout(&vo,&op)>0){
if(vorbis_synthesis(&vb,&op)==0) { /* test for success! */
vorbis_synthesis_blockin(&vd,&vb);
}
} else { /* we need more data; break out to suck in another page */
//printf("need moar data\n");
break;
};
}
}
while(theora_p && !videobuf_ready){
/* theora is one in, one out... */
if(ogg_stream_packetout(&to,&op)>0){
if(pp_inc){
pp_level+=pp_inc;
th_decode_ctl(td,TH_DECCTL_SET_PPLEVEL,&pp_level,
sizeof(pp_level));
pp_inc=0;
}
/*HACK: This should be set after a seek or a gap, but we might not have
a granulepos for the first packet (we only have them for the last
packet on a page), so we just set it as often as we get it.
To do this right, we should back-track from the last packet on the
page and compute the correct granulepos for the first packet after
a seek or a gap.*/
if(op.granulepos>=0){
th_decode_ctl(td,TH_DECCTL_SET_GRANPOS,&op.granulepos,
sizeof(op.granulepos));
}
ogg_int64_t videobuf_granulepos;
if(th_decode_packetin(td,&op,&videobuf_granulepos)==0){
videobuf_time=th_granule_time(td,videobuf_granulepos);
//printf("frame time %f, play time %f, ready %i\n", (float)videobuf_time, get_time(), videobuf_ready);
/* is it already too old to be useful? This is only actually
useful cosmetically after a SIGSTOP. Note that we have to
decode the frame even if we don't show it (for now) due to
keyframing. Soon enough libtheora will be able to deal
with non-keyframe seeks. */
if(videobuf_time>=get_time())
videobuf_ready=1;
else{
/*If we are too slow, reduce the pp level.*/
pp_inc=pp_level>0?-1:0;
}
}
} else
break;
}
if (/*!videobuf_ready && */ audio_pending == 0 && file->eof_reached()) {
printf("video done, stopping\n");
stop();
return;
};
if (!videobuf_ready || audio_todo > 0){
/* no data yet for somebody. Grab another page */
buffer_data();
while(ogg_sync_pageout(&oy,&og)>0){
queue_page(&og);
}
}
/* If playback has begun, top audio buffer off immediately. */
//if(stateflag) audio_write_nonblocking();
/* are we at or past time for this video frame? */
if(videobuf_ready && videobuf_time<=get_time()){
video_write();
videobuf_ready=0;
} else {
//printf("frame at %f not ready (time %f), ready %i\n", (float)videobuf_time, get_time(), videobuf_ready);
}
float tdiff=videobuf_time-get_time();
/*If we have lots of extra time, increase the post-processing level.*/
if(tdiff>ti.fps_denominator*0.25/ti.fps_numerator){
pp_inc=pp_level<pp_level_max?1:0;
}
else if(tdiff<ti.fps_denominator*0.05/ti.fps_numerator){
pp_inc=pp_level>0?-1:0;
}
};
bool VideoStreamTheora::_can_mix() const {
return !buffering;
};
void VideoStreamTheora::play() {
if (!playing)
last_update_time=0;
playing = true;
};
void VideoStreamTheora::stop() {
playing = false;
last_update_time=0;
};
bool VideoStreamTheora::is_playing() const {
return playing;
};
void VideoStreamTheora::set_paused(bool p_paused) {
playing = !p_paused;
};
bool VideoStreamTheora::is_paused(bool p_paused) const {
return playing;
};
void VideoStreamTheora::set_loop(bool p_enable) {
};
bool VideoStreamTheora::has_loop() const {
return false;
};
float VideoStreamTheora::get_length() const {
return 0;
};
String VideoStreamTheora::get_stream_name() const {
return "";
};
int VideoStreamTheora::get_loop_count() const {
return 0;
};
float VideoStreamTheora::get_pos() const {
return get_time();
};
void VideoStreamTheora::seek_pos(float p_time) {
// no
};
VideoStreamTheora::VideoStreamTheora() {
file = NULL;
theora_p = 0;
vorbis_p = 0;
videobuf_ready = 0;
playing = false;
frames_pending = 0;
videobuf_time = 0;
last_update_time =0;
buffering=false;
};
VideoStreamTheora::~VideoStreamTheora() {
clear();
if (file)
memdelete(file);
};
RES ResourceFormatLoaderVideoStreamTheora::load(const String &p_path,const String& p_original_path) {
VideoStreamTheora *stream = memnew(VideoStreamTheora);
stream->set_file(p_path);
return Ref<VideoStreamTheora>(stream);
}
void ResourceFormatLoaderVideoStreamTheora::get_recognized_extensions(List<String> *p_extensions) const {
p_extensions->push_back("ogm");
p_extensions->push_back("ogv");
}
bool ResourceFormatLoaderVideoStreamTheora::handles_type(const String& p_type) const {
return (p_type=="AudioStream" || p_type=="VideoStreamTheora");
}
String ResourceFormatLoaderVideoStreamTheora::get_resource_type(const String &p_path) const {
String exl=p_path.extension().to_lower();
if (exl=="ogm" || exl=="ogv")
return "AudioStreamTheora";
return "";
}
#endif