#include <stdio.h>
#include <stdlib.h>
#include <string.h>
+#include <limits.h>
#include <pthread.h>
#include <sys/time.h>
#include <math.h>
if ( this != NULL )
{
// Get properties from the consumer
- mlt_properties properties = mlt_consumer_properties( this );
+ mlt_properties properties = MLT_CONSUMER_PROPERTIES( this );
// Assign close callback
this->close = consumer_close;
mlt_properties_set_int( properties, "audio_bit_rate", 128000 );
mlt_properties_set_int( properties, "video_bit_rate", 200 * 1000 );
mlt_properties_set_int( properties, "video_bit_rate_tolerance", 4000 * 1000 );
- mlt_properties_set_int( properties, "frame_rate_base", 1 );
mlt_properties_set_int( properties, "gop_size", 12 );
mlt_properties_set_int( properties, "b_frames", 0 );
mlt_properties_set_int( properties, "mb_decision", FF_MB_DECISION_SIMPLE );
static int consumer_start( mlt_consumer this )
{
// Get the properties
- mlt_properties properties = mlt_consumer_properties( this );
+ mlt_properties properties = MLT_CONSUMER_PROPERTIES( this );
// Check that we're not already running
if ( !mlt_properties_get_int( properties, "running" ) )
{
// Allocate a thread
pthread_t *thread = calloc( 1, sizeof( pthread_t ) );
- pthread_attr_t thread_attributes;
// Get the width and height
int width = mlt_properties_get_int( properties, "width" );
// Set the running state
mlt_properties_set_int( properties, "running", 1 );
- // Inherit the scheduling priority
- pthread_attr_init( &thread_attributes );
- pthread_attr_setinheritsched( &thread_attributes, PTHREAD_INHERIT_SCHED );
-
// Create the thread
- pthread_create( thread, &thread_attributes, consumer_thread, this );
+ pthread_create( thread, NULL, consumer_thread, this );
}
return 0;
}
static int consumer_stop( mlt_consumer this )
{
// Get the properties
- mlt_properties properties = mlt_consumer_properties( this );
+ mlt_properties properties = MLT_CONSUMER_PROPERTIES( this );
// Check that we're running
if ( mlt_properties_get_int( properties, "running" ) )
static int consumer_is_stopped( mlt_consumer this )
{
// Get the properties
- mlt_properties properties = mlt_consumer_properties( this );
+ mlt_properties properties = MLT_CONSUMER_PROPERTIES( this );
return !mlt_properties_get_int( properties, "running" );
}
static AVStream *add_audio_stream( mlt_consumer this, AVFormatContext *oc, int codec_id )
{
// Get the properties
- mlt_properties properties = mlt_consumer_properties( this );
+ mlt_properties properties = MLT_CONSUMER_PROPERTIES( this );
// Create a new stream
AVStream *st = av_new_stream( oc, 1 );
c->bit_rate = mlt_properties_get_int( properties, "audio_bit_rate" );
c->sample_rate = mlt_properties_get_int( properties, "frequency" );
c->channels = mlt_properties_get_int( properties, "channels" );
+
+ // Allow the user to override the audio fourcc
+ if ( mlt_properties_get( properties, "afourcc" ) )
+ {
+ char *tail = NULL;
+ char *arg = mlt_properties_get( properties, "afourcc" );
+ int tag = strtol( arg, &tail, 0);
+ if( !tail || *tail )
+ tag = arg[ 0 ] + ( arg[ 1 ] << 8 ) + ( arg[ 2 ] << 16 ) + ( arg[ 3 ] << 24 );
+ c->codec_tag = tag;
+ }
}
else
{
static AVStream *add_video_stream( mlt_consumer this, AVFormatContext *oc, int codec_id )
{
// Get the properties
- mlt_properties properties = mlt_consumer_properties( this );
+ mlt_properties properties = MLT_CONSUMER_PROPERTIES( this );
// Create a new stream
AVStream *st = av_new_stream( oc, 0 );
c->bit_rate_tolerance = mlt_properties_get_int( properties, "video_bit_rate_tolerance" );
c->width = mlt_properties_get_int( properties, "width" );
c->height = mlt_properties_get_int( properties, "height" );
- c->frame_rate = mlt_properties_get_double( properties, "fps" );
- c->frame_rate_base = mlt_properties_get_double( properties, "frame_rate_base" );
- c->frame_rate_base = 1;
+ c->time_base.num = mlt_properties_get_int( properties, "frame_rate_den" );
+ c->time_base.den = mlt_properties_get_int( properties, "frame_rate_num" );
c->gop_size = mlt_properties_get_int( properties, "gop_size" );
+ c->pix_fmt = PIX_FMT_YUV420P;
if ( mlt_properties_get_int( properties, "b_frames" ) )
{
}
c->mb_decision = mlt_properties_get_int( properties, "mb_decision" );
- c->sample_aspect_ratio = av_d2q( mlt_properties_get_double( properties, "aspect_ratio" ), 255 );
+ c->sample_aspect_ratio = av_d2q( mlt_properties_get_double( properties, "display_ratio" ) * c->height / c->width, 255 );
c->mb_cmp = mlt_properties_get_int( properties, "mb_cmp" );
c->ildct_cmp = mlt_properties_get_int( properties, "ildct_cmp" );
c->me_sub_cmp = mlt_properties_get_int( properties, "sub_cmp" );
st->quality = FF_QP2LAMBDA * mlt_properties_get_double( properties, "qscale" );
}
- // Some formats want stream headers to be seperate (hmm)
- if( !strcmp( oc->oformat->name, "mp4" ) ||
- !strcmp( oc->oformat->name, "mov" ) ||
- !strcmp( oc->oformat->name, "3gp" ) )
+ // Allow the user to override the video fourcc
+ if ( mlt_properties_get( properties, "vfourcc" ) )
+ {
+ char *tail = NULL;
+ const char *arg = mlt_properties_get( properties, "vfourcc" );
+ int tag = strtol( arg, &tail, 0);
+ if( !tail || *tail )
+ tag = arg[ 0 ] + ( arg[ 1 ] << 8 ) + ( arg[ 2 ] << 16 ) + ( arg[ 3 ] << 24 );
+ c->codec_tag = tag;
+ }
+
+ // Some formats want stream headers to be seperate
+ if ( oc->oformat->flags & AVFMT_GLOBALHEADER )
c->flags |= CODEC_FLAG_GLOBAL_HEADER;
c->rc_max_rate = mlt_properties_get_int( properties, "video_rc_max_rate" );
static int open_video(AVFormatContext *oc, AVStream *st)
{
// Get the codec
- AVCodecContext *c = &st->codec;
+ AVCodecContext *video_enc = &st->codec;
// find the video encoder
- AVCodec *codec = avcodec_find_encoder(c->codec_id);
+ AVCodec *codec = avcodec_find_encoder( video_enc->codec_id );
+
+ if( codec && codec->pix_fmts )
+ {
+ const enum PixelFormat *p = codec->pix_fmts;
+ for( ; *p!=-1; p++ )
+ {
+ if( *p == video_enc->pix_fmt )
+ break;
+ }
+ if( *p == -1 )
+ video_enc->pix_fmt = codec->pix_fmts[ 0 ];
+ }
// Open the codec safely
- return codec != NULL && avcodec_open(c, codec) >= 0;
+ return codec != NULL && avcodec_open( video_enc, codec ) >= 0;
}
void close_video(AVFormatContext *oc, AVStream *st)
mlt_consumer this = arg;
// Get the properties
- mlt_properties properties = mlt_consumer_properties( this );
+ mlt_properties properties = MLT_CONSUMER_PROPERTIES( this );
// Get the terminate on pause property
int terminate_on_pause = mlt_properties_get_int( properties, "terminate_on_pause" );
sample_fifo fifo = mlt_properties_get_data( properties, "sample_fifo", NULL );
// Need two av pictures for converting
- AVFrame *output = alloc_picture( PIX_FMT_YUV420P, width, height );
+ AVFrame *output = NULL;
AVFrame *input = alloc_picture( PIX_FMT_YUV422, width, height );
// For receiving images from an mlt_frame
AVStream *video_st = NULL;
// Time stamps
- double audio_pts, video_pts;
+ double audio_pts = 0;
+ double video_pts = 0;
// Loop variable
int i;
mlt_properties_set_int( properties, "running", 0 );
}
+ // Allocate picture
+ if ( video_st )
+ output = alloc_picture( video_st->codec.pix_fmt, width, height );
+
// Last check - need at least one stream
if ( audio_st == NULL && video_st == NULL )
mlt_properties_set_int( properties, "running", 0 );
frames ++;
// Default audio args
- frame_properties = mlt_frame_properties( frame );
+ frame_properties = MLT_FRAME_PROPERTIES( frame );
// Check for the terminated condition
terminated = terminate_on_pause && mlt_properties_get_double( frame_properties, "_speed" ) == 0.0;
// Get audio and append to the fifo
- if ( audio_st )
+ if ( !terminated && audio_st )
{
samples = mlt_sample_calculator( fps, frequency, count );
mlt_frame_get_audio( frame, &pcm, &aud_fmt, &frequency, &channels, &samples );
mlt_properties_set_data( properties, "sample_fifo", fifo, 0, ( mlt_destructor )sample_fifo_close, NULL );
}
+ if ( mlt_properties_get_double( frame_properties, "_speed" ) != 1.0 )
+ memset( pcm, 0, samples * channels * 2 );
+
// Append the samples
sample_fifo_append( fifo, pcm, samples * channels );
total_time += ( samples * 1000000 ) / frequency;
}
// Encode the image
- if ( video_st )
+ if ( !terminated && video_st )
mlt_deque_push_back( queue, frame );
else
mlt_frame_close( frame );
// While we have stuff to process, process...
while ( 1 )
{
- // Compute current audio and video time
- if (audio_st)
- audio_pts = (double)audio_st->pts.val * audio_st->time_base.num / audio_st->time_base.den;
+ if (audio_st)
+ audio_pts = (double)audio_st->pts.val * audio_st->time_base.num / audio_st->time_base.den;
else
- audio_pts = 0.0;
-
+ audio_pts = 0.0;
+
if (video_st)
- video_pts = (double)video_st->pts.val * video_st->time_base.num / video_st->time_base.den;
- else
- video_pts = 0.0;
+ video_pts = (double)video_st->pts.val * video_st->time_base.num / video_st->time_base.den;
+ else
+ video_pts = 0.0;
// Write interleaved audio and video frames
if ( !video_st || ( video_st && audio_st && audio_pts < video_pts ) )
pkt.size = avcodec_encode_audio( c, audio_outbuf, audio_outbuf_size, buffer );
// Write the compressed frame in the media file
- pkt.pts= c->coded_frame->pts;
+ if ( c->coded_frame )
+ pkt.pts = av_rescale_q( c->coded_frame->pts, c->time_base, audio_st->time_base );
pkt.flags |= PKT_FLAG_KEY;
pkt.stream_index= audio_st->index;
pkt.data= audio_outbuf;
- if ( av_write_frame( oc, &pkt ) != 0)
+ if ( av_interleaved_write_frame( oc, &pkt ) != 0)
fprintf(stderr, "Error while writing audio frame\n");
+
+ audio_pts += c->frame_size;
}
else
{
AVCodecContext *c;
frame = mlt_deque_pop_front( queue );
- frame_properties = mlt_frame_properties( frame );
+ frame_properties = MLT_FRAME_PROPERTIES( frame );
c = &video_st->codec;
uint8_t *p;
uint8_t *q;
+ mlt_events_fire( properties, "consumer-frame-show", frame, NULL );
+
+ // This will cause some fx to go awry....
+ if ( mlt_properties_get_int( properties, "transcode" ) )
+ {
+ mlt_properties_set_int( MLT_FRAME_PROPERTIES( frame ), "normalised_width", img_height * 4.0 / 3.0 );
+ mlt_properties_set_int( MLT_FRAME_PROPERTIES( frame ), "normalised_height", img_height );
+ }
+
mlt_frame_get_image( frame, &image, &img_fmt, &img_width, &img_height, 0 );
q = image;
}
}
- img_convert( ( AVPicture * )output, PIX_FMT_YUV420P, ( AVPicture * )input, PIX_FMT_YUV422, width, height );
+ img_convert( ( AVPicture * )output, video_st->codec.pix_fmt, ( AVPicture * )input, PIX_FMT_YUV422, width, height );
}
if (oc->oformat->flags & AVFMT_RAWPICTURE)
pkt.size= sizeof(AVPicture);
ret = av_write_frame(oc, &pkt);
+ video_pts += c->frame_size;
}
else
{
AVPacket pkt;
av_init_packet( &pkt );
- pkt.pts= c->coded_frame->pts;
- if(c->coded_frame->key_frame)
+ if ( c->coded_frame )
+ pkt.pts= av_rescale_q( c->coded_frame->pts, c->time_base, video_st->time_base );
+ if( c->coded_frame && c->coded_frame->key_frame )
pkt.flags |= PKT_FLAG_KEY;
pkt.stream_index= video_st->index;
pkt.data= video_outbuf;
pkt.size= out_size;
// write the compressed frame in the media file
- ret = av_write_frame(oc, &pkt);
+ ret = av_interleaved_write_frame(oc, &pkt);
+ video_pts += c->frame_size;
}
}
frame_count++;
}
}
- if ( real_time_output && frames % 25 == 0 )
+ if ( real_time_output && frames % 12 == 0 )
{
long passed = time_difference( &ante );
if ( fifo != NULL )
url_fclose(&oc->pb);
// Clean up input and output frames
- av_free( output->data[0] );
+ if ( output )
+ av_free( output->data[0] );
av_free( output );
av_free( input->data[0] );
av_free( input );