*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
#include <stdlib.h>
#include <string.h>
#include <pthread.h>
-
+#include <sys/time.h>
#include <math.h>
// avformat header files
-#include <ffmpeg/avformat.h>
+#include <avformat.h>
+
+//
+// This structure should be extended and made globally available in mlt
+//
typedef struct
{
int16_t *buffer;
int size;
int used;
+ double time;
+ int frequency;
+ int channels;
}
*sample_fifo, sample_fifo_s;
-sample_fifo sample_fifo_init( )
+sample_fifo sample_fifo_init( int frequency, int channels )
+{
+ sample_fifo this = calloc( 1, sizeof( sample_fifo_s ) );
+ this->frequency = frequency;
+ this->channels = channels;
+ return this;
+}
+
+// sample_fifo_clear and check are temporarily aborted (not working as intended)
+
+void sample_fifo_clear( sample_fifo this, double time )
+{
+ int words = ( float )( time - this->time ) * this->frequency * this->channels;
+ if ( ( int )( ( float )time * 100 ) < ( int )( ( float )this->time * 100 ) && this->used > words && words > 0 )
+ {
+ memmove( this->buffer, &this->buffer[ words ], ( this->used - words ) * sizeof( int16_t ) );
+ this->used -= words;
+ this->time = time;
+ }
+ else if ( ( int )( ( float )time * 100 ) != ( int )( ( float )this->time * 100 ) )
+ {
+ this->used = 0;
+ this->time = time;
+ }
+}
+
+void sample_fifo_check( sample_fifo this, double time )
{
- return calloc( 1, sizeof( sample_fifo_s ) );
+ if ( this->used == 0 )
+ {
+ if ( ( int )( ( float )time * 100 ) < ( int )( ( float )this->time * 100 ) )
+ this->time = time;
+ }
}
void sample_fifo_append( sample_fifo this, int16_t *samples, int count )
this->used -= count;
memmove( this->buffer, &this->buffer[ count ], this->used * sizeof( int16_t ) );
+ this->time += ( double )count / this->channels / this->frequency;
+
return count;
}
mlt_consumer consumer_avformat_init( char *arg )
{
// Allocate the consumer
- mlt_consumer this = calloc( 1, sizeof( struct mlt_consumer_s ) );
+ mlt_consumer this = mlt_consumer_new( );
// If memory allocated and initialises without error
- if ( this != NULL && mlt_consumer_init( this, NULL ) == 0 )
+ if ( this != NULL )
{
// Get properties from the consumer
mlt_properties properties = mlt_consumer_properties( this );
mlt_properties_set( properties, "target", arg );
// sample and frame queue
- mlt_properties_set_data( properties, "sample_fifo", sample_fifo_init( ), 0, ( mlt_destructor )sample_fifo_close, NULL );
mlt_properties_set_data( properties, "frame_queue", mlt_deque_init( ), 0, ( mlt_destructor )mlt_deque_close, NULL );
- // Set avformat defaults
+ // Set avformat defaults (all lifted from ffmpeg.c)
mlt_properties_set_int( properties, "audio_bit_rate", 128000 );
- mlt_properties_set_int( properties, "video_bit_rate", 400000 );
+ mlt_properties_set_int( properties, "video_bit_rate", 200 * 1000 );
+ mlt_properties_set_int( properties, "video_bit_rate_tolerance", 4000 * 1000 );
+ mlt_properties_set_int( properties, "frame_rate_base", 1 );
mlt_properties_set_int( properties, "gop_size", 12 );
- mlt_properties_set_int( properties, "max_b_frames", 0 );
- mlt_properties_set_int( properties, "mb_decision", 0 );
+ mlt_properties_set_int( properties, "b_frames", 0 );
+ mlt_properties_set_int( properties, "mb_decision", FF_MB_DECISION_SIMPLE );
+ mlt_properties_set_double( properties, "qscale", 0 );
+ mlt_properties_set_int( properties, "me_method", ME_EPZS );
+ mlt_properties_set_int( properties, "mb_cmp", FF_CMP_SAD );
+ mlt_properties_set_int( properties, "ildct_cmp", FF_CMP_VSAD );
+ mlt_properties_set_int( properties, "sub_cmp", FF_CMP_SAD );
+ mlt_properties_set_int( properties, "cmp", FF_CMP_SAD );
+ mlt_properties_set_int( properties, "pre_cmp", FF_CMP_SAD );
+ mlt_properties_set_int( properties, "pre_me", 0 );
+ mlt_properties_set_double( properties, "lumi_mask", 0 );
+ mlt_properties_set_double( properties, "dark_mask", 0 );
+ mlt_properties_set_double( properties, "scplx_mask", 0 );
+ mlt_properties_set_double( properties, "tcplx_mask", 0 );
+ mlt_properties_set_double( properties, "p_mask", 0 );
+ mlt_properties_set_int( properties, "qns", 0 );
+ mlt_properties_set_int( properties, "video_qmin", 2 );
+ mlt_properties_set_int( properties, "video_qmax", 31 );
+ mlt_properties_set_int( properties, "video_lmin", 2*FF_QP2LAMBDA );
+ mlt_properties_set_int( properties, "video_lmax", 31*FF_QP2LAMBDA );
+ mlt_properties_set_int( properties, "video_mb_qmin", 2 );
+ mlt_properties_set_int( properties, "video_mb_qmax", 31 );
+ mlt_properties_set_int( properties, "video_qdiff", 3 );
+ mlt_properties_set_double( properties, "video_qblur", 0.5 );
+ mlt_properties_set_double( properties, "video_qcomp", 0.5 );
+ mlt_properties_set_int( properties, "video_rc_max_rate", 0 );
+ mlt_properties_set_int( properties, "video_rc_min_rate", 0 );
+ mlt_properties_set_int( properties, "video_rc_buffer_size", 0 );
+ mlt_properties_set_double( properties, "video_rc_buffer_aggressivity", 1.0 );
+ mlt_properties_set_double( properties, "video_rc_initial_cplx", 0 );
+ mlt_properties_set_double( properties, "video_i_qfactor", 1.25 );
+ mlt_properties_set_double( properties, "video_b_qfactor", 1.25 );
+ mlt_properties_set_double( properties, "video_i_qoffset", -0.8 );
+ mlt_properties_set_double( properties, "video_b_qoffset", 0 );
+ mlt_properties_set_int( properties, "video_intra_quant_bias", FF_DEFAULT_QUANT_BIAS );
+ mlt_properties_set_int( properties, "video_inter_quant_bias", FF_DEFAULT_QUANT_BIAS );
+ mlt_properties_set_int( properties, "dct_algo", 0 );
+ mlt_properties_set_int( properties, "idct_algo", 0 );
+ mlt_properties_set_int( properties, "me_threshold", 0 );
+ mlt_properties_set_int( properties, "mb_threshold", 0 );
+ mlt_properties_set_int( properties, "intra_dc_precision", 0 );
+ mlt_properties_set_int( properties, "strict", 0 );
+ mlt_properties_set_int( properties, "error_rate", 0 );
+ mlt_properties_set_int( properties, "noise_reduction", 0 );
+ mlt_properties_set_int( properties, "sc_threshold", 0 );
+ mlt_properties_set_int( properties, "me_range", 0 );
+ mlt_properties_set_int( properties, "coder", 0 );
+ mlt_properties_set_int( properties, "context", 0 );
+ mlt_properties_set_int( properties, "predictor", 0 );
// Ensure termination at end of the stream
mlt_properties_set_int( properties, "terminate_on_pause", 1 );
this->stop = consumer_stop;
this->is_stopped = consumer_is_stopped;
}
- else
- {
- // Clean up in case of init failure
- free( this );
- this = NULL;
- }
// Return this
return this;
{
// Allocate a thread
pthread_t *thread = calloc( 1, sizeof( pthread_t ) );
+ pthread_attr_t thread_attributes;
// Get the width and height
int width = mlt_properties_get_int( properties, "width" );
int height = mlt_properties_get_int( properties, "height" );
// Obtain the size property
- char *size = mlt_properties_get( properties, "size" );
+ char *size = mlt_properties_get( properties, "size" );
// Interpret it
if ( size != NULL )
// Set the running state
mlt_properties_set_int( properties, "running", 1 );
+ // Inherit the scheduling priority
+ pthread_attr_init( &thread_attributes );
+ pthread_attr_setinheritsched( &thread_attributes, PTHREAD_INHERIT_SCHED );
+
// Create the thread
- pthread_create( thread, NULL, consumer_thread, this );
+ pthread_create( thread, &thread_attributes, consumer_thread, this );
}
return 0;
}
mlt_properties properties = mlt_consumer_properties( this );
// Create a new stream
- AVStream *st = av_new_stream( oc, 1 );
+ AVStream *st = av_new_stream( oc, 1 );
// If created, then initialise from properties
- if ( st != NULL )
+ if ( st != NULL )
{
- AVCodecContext *c = &st->codec;
- c->codec_id = codec_id;
- c->codec_type = CODEC_TYPE_AUDIO;
-
- // Put sample parameters
- c->bit_rate = mlt_properties_get_int( properties, "audio_bit_rate" );
- c->sample_rate = mlt_properties_get_int( properties, "frequency" );
- c->channels = mlt_properties_get_int( properties, "channels" );
+ AVCodecContext *c = &st->codec;
+ c->codec_id = codec_id;
+ c->codec_type = CODEC_TYPE_AUDIO;
+
+ // Put sample parameters
+ c->bit_rate = mlt_properties_get_int( properties, "audio_bit_rate" );
+ c->sample_rate = mlt_properties_get_int( properties, "frequency" );
+ c->channels = mlt_properties_get_int( properties, "channels" );
}
else
{
- fprintf( stderr, "Could not allocate a stream for audio\n" );
- }
+ fprintf( stderr, "Could not allocate a stream for audio\n" );
+ }
- return st;
+ return st;
}
static int open_audio( AVFormatContext *oc, AVStream *st, int audio_outbuf_size )
int audio_input_frame_size = 0;
// Get the context
- AVCodecContext *c = &st->codec;
+ AVCodecContext *c = &st->codec;
// Find the encoder
- AVCodec *codec = avcodec_find_encoder( c->codec_id );
+ AVCodec *codec = avcodec_find_encoder( c->codec_id );
// Continue if codec found and we can open it
if ( codec != NULL && avcodec_open(c, codec) >= 0 )
{
- // ugly hack for PCM codecs (will be removed ASAP with new PCM
- // support to compute the input frame size in samples
- if ( c->frame_size <= 1 )
+ // ugly hack for PCM codecs (will be removed ASAP with new PCM
+ // support to compute the input frame size in samples
+ if ( c->frame_size <= 1 )
{
- audio_input_frame_size = audio_outbuf_size / c->channels;
- switch(st->codec.codec_id)
+ audio_input_frame_size = audio_outbuf_size / c->channels;
+ switch(st->codec.codec_id)
{
- case CODEC_ID_PCM_S16LE:
- case CODEC_ID_PCM_S16BE:
- case CODEC_ID_PCM_U16LE:
- case CODEC_ID_PCM_U16BE:
- audio_input_frame_size >>= 1;
- break;
- default:
- break;
- }
- }
+ case CODEC_ID_PCM_S16LE:
+ case CODEC_ID_PCM_S16BE:
+ case CODEC_ID_PCM_U16LE:
+ case CODEC_ID_PCM_U16BE:
+ audio_input_frame_size >>= 1;
+ break;
+ default:
+ break;
+ }
+ }
else
{
- audio_input_frame_size = c->frame_size;
- }
+ audio_input_frame_size = c->frame_size;
+ }
+
+ // Some formats want stream headers to be seperate (hmm)
+ if( !strcmp( oc->oformat->name, "mp4" ) ||
+ !strcmp( oc->oformat->name, "mov" ) ||
+ !strcmp( oc->oformat->name, "3gp" ) )
+ c->flags |= CODEC_FLAG_GLOBAL_HEADER;
}
else
{
static void close_audio( AVFormatContext *oc, AVStream *st )
{
- avcodec_close( &st->codec );
+ avcodec_close( &st->codec );
}
/** Add a video output stream
static AVStream *add_video_stream( mlt_consumer this, AVFormatContext *oc, int codec_id )
{
- // Get the properties
+ // Get the properties
mlt_properties properties = mlt_consumer_properties( this );
// Create a new stream
- AVStream *st = av_new_stream( oc, 0 );
+ AVStream *st = av_new_stream( oc, 0 );
- if ( st != NULL )
+ if ( st != NULL )
{
AVCodecContext *c = &st->codec;
- c->codec_id = codec_id;
- c->codec_type = CODEC_TYPE_VIDEO;
-
- // put sample parameters
- c->bit_rate = mlt_properties_get_int( properties, "video_bit_rate" );
- c->width = mlt_properties_get_int( properties, "width" );
- c->height = mlt_properties_get_int( properties, "height" );
- c->frame_rate = mlt_properties_get_double( properties, "fps" );
- c->frame_rate_base = 1;
- c->gop_size = mlt_properties_get_int( properties, "gop_size" );
- c->max_b_frames = mlt_properties_get_int( properties, "max_b_frames" );
- c->mb_decision = mlt_properties_get_int( properties, "mb_decision" );
-
- // Some formats want stream headers to be seperate (hmm)
- if( !strcmp( oc->oformat->name, "mp4" ) ||
+ c->codec_id = codec_id;
+ c->codec_type = CODEC_TYPE_VIDEO;
+
+ // put sample parameters
+ c->bit_rate = mlt_properties_get_int( properties, "video_bit_rate" );
+ c->bit_rate_tolerance = mlt_properties_get_int( properties, "video_bit_rate_tolerance" );
+ c->width = mlt_properties_get_int( properties, "width" );
+ c->height = mlt_properties_get_int( properties, "height" );
+ c->frame_rate = mlt_properties_get_double( properties, "fps" );
+ c->frame_rate_base = mlt_properties_get_double( properties, "frame_rate_base" );
+ c->frame_rate_base = 1;
+ c->gop_size = mlt_properties_get_int( properties, "gop_size" );
+
+ if ( mlt_properties_get_int( properties, "b_frames" ) )
+ {
+ c->max_b_frames = mlt_properties_get_int( properties, "b_frames" );
+ c->b_frame_strategy = 0;
+ c->b_quant_factor = 2.0;
+ }
+
+ c->mb_decision = mlt_properties_get_int( properties, "mb_decision" );
+ c->sample_aspect_ratio = av_d2q( mlt_properties_get_double( properties, "aspect_ratio" ), 255 );
+ c->mb_cmp = mlt_properties_get_int( properties, "mb_cmp" );
+ c->ildct_cmp = mlt_properties_get_int( properties, "ildct_cmp" );
+ c->me_sub_cmp = mlt_properties_get_int( properties, "sub_cmp" );
+ c->me_cmp = mlt_properties_get_int( properties, "cmp" );
+ c->me_pre_cmp = mlt_properties_get_int( properties, "pre_cmp" );
+ c->pre_me = mlt_properties_get_int( properties, "pre_me" );
+ c->lumi_masking = mlt_properties_get_double( properties, "lumi_mask" );
+ c->dark_masking = mlt_properties_get_double( properties, "dark_mask" );
+ c->spatial_cplx_masking = mlt_properties_get_double( properties, "scplx_mask" );
+ c->temporal_cplx_masking = mlt_properties_get_double( properties, "tcplx_mask" );
+ c->p_masking = mlt_properties_get_double( properties, "p_mask" );
+ c->quantizer_noise_shaping= mlt_properties_get_int( properties, "qns" );
+ c->qmin = mlt_properties_get_int( properties, "video_qmin" );
+ c->qmax = mlt_properties_get_int( properties, "video_qmax" );
+ c->lmin = mlt_properties_get_int( properties, "video_lmin" );
+ c->lmax = mlt_properties_get_int( properties, "video_lmax" );
+ c->mb_qmin = mlt_properties_get_int( properties, "video_mb_qmin" );
+ c->mb_qmax = mlt_properties_get_int( properties, "video_mb_qmax" );
+ c->max_qdiff = mlt_properties_get_int( properties, "video_qdiff" );
+ c->qblur = mlt_properties_get_double( properties, "video_qblur" );
+ c->qcompress = mlt_properties_get_double( properties, "video_qcomp" );
+
+ if ( mlt_properties_get_double( properties, "qscale" ) > 0 )
+ {
+ c->flags |= CODEC_FLAG_QSCALE;
+ st->quality = FF_QP2LAMBDA * mlt_properties_get_double( properties, "qscale" );
+ }
+
+ // Some formats want stream headers to be seperate (hmm)
+ if( !strcmp( oc->oformat->name, "mp4" ) ||
!strcmp( oc->oformat->name, "mov" ) ||
!strcmp( oc->oformat->name, "3gp" ) )
- c->flags |= CODEC_FLAG_GLOBAL_HEADER;
+ c->flags |= CODEC_FLAG_GLOBAL_HEADER;
+
+ c->rc_max_rate = mlt_properties_get_int( properties, "video_rc_max_rate" );
+ c->rc_min_rate = mlt_properties_get_int( properties, "video_rc_min_rate" );
+ c->rc_buffer_size = mlt_properties_get_int( properties, "video_rc_buffer_size" );
+ c->rc_buffer_aggressivity= mlt_properties_get_double( properties, "video_rc_buffer_aggressivity" );
+ c->rc_initial_cplx= mlt_properties_get_double( properties, "video_rc_initial_cplx" );
+ c->i_quant_factor = mlt_properties_get_double( properties, "video_i_qfactor" );
+ c->b_quant_factor = mlt_properties_get_double( properties, "video_b_qfactor" );
+ c->i_quant_offset = mlt_properties_get_double( properties, "video_i_qoffset" );
+ c->b_quant_offset = mlt_properties_get_double( properties, "video_b_qoffset" );
+ c->intra_quant_bias = mlt_properties_get_int( properties, "video_intra_quant_bias" );
+ c->inter_quant_bias = mlt_properties_get_int( properties, "video_inter_quant_bias" );
+ c->dct_algo = mlt_properties_get_int( properties, "dct_algo" );
+ c->idct_algo = mlt_properties_get_int( properties, "idct_algo" );
+ c->me_threshold= mlt_properties_get_int( properties, "me_threshold" );
+ c->mb_threshold= mlt_properties_get_int( properties, "mb_threshold" );
+ c->intra_dc_precision= mlt_properties_get_int( properties, "intra_dc_precision" );
+ c->strict_std_compliance = mlt_properties_get_int( properties, "strict" );
+ c->error_rate = mlt_properties_get_int( properties, "error_rate" );
+ c->noise_reduction= mlt_properties_get_int( properties, "noise_reduction" );
+ c->scenechange_threshold= mlt_properties_get_int( properties, "sc_threshold" );
+ c->me_range = mlt_properties_get_int( properties, "me_range" );
+ c->coder_type= mlt_properties_get_int( properties, "coder" );
+ c->context_model= mlt_properties_get_int( properties, "context" );
+ c->prediction_method= mlt_properties_get_int( properties, "predictor" );
+ c->me_method = mlt_properties_get_int( properties, "me_method" );
}
else
{
- fprintf( stderr, "Could not allocate a stream for video\n" );
- }
-
- return st;
+ fprintf( stderr, "Could not allocate a stream for video\n" );
+ }
+
+ return st;
}
static AVFrame *alloc_picture( int pix_fmt, int width, int height )
{
// Allocate a frame
- AVFrame *picture = avcodec_alloc_frame();
+ AVFrame *picture = avcodec_alloc_frame();
// Determine size of the
- int size = avpicture_get_size(pix_fmt, width, height);
+ int size = avpicture_get_size(pix_fmt, width, height);
// Allocate the picture buf
- uint8_t *picture_buf = av_malloc(size);
+ uint8_t *picture_buf = av_malloc(size);
// If we have both, then fill the image
if ( picture != NULL && picture_buf != NULL )
{
// Fill the frame with the allocated buffer
- avpicture_fill((AVPicture *)picture, picture_buf, pix_fmt, width, height);
+ avpicture_fill( (AVPicture *)picture, picture_buf, pix_fmt, width, height);
}
else
{
// Something failed - clean up what we can
- av_free(picture);
- av_free(picture_buf);
- picture = NULL;
+ av_free( picture );
+ av_free( picture_buf );
+ picture = NULL;
}
- return picture;
+ return picture;
}
-
+
static int open_video(AVFormatContext *oc, AVStream *st)
{
// Get the codec
- AVCodecContext *c = &st->codec;
+ AVCodecContext *c = &st->codec;
- // find the video encoder
- AVCodec *codec = avcodec_find_encoder(c->codec_id);
+ // find the video encoder
+ AVCodec *codec = avcodec_find_encoder(c->codec_id);
// Open the codec safely
return codec != NULL && avcodec_open(c, codec) >= 0;
void close_video(AVFormatContext *oc, AVStream *st)
{
- avcodec_close(&st->codec);
+ avcodec_close(&st->codec);
+}
+
+static inline long time_difference( struct timeval *time1 )
+{
+ struct timeval time2;
+ gettimeofday( &time2, NULL );
+ return time2.tv_sec * 1000000 + time2.tv_usec - time1->tv_sec * 1000000 - time1->tv_usec;
}
/** The main thread - the argument is simply the consumer.
// Get the terminate on pause property
int terminate_on_pause = mlt_properties_get_int( properties, "terminate_on_pause" );
+ int terminated = 0;
+
+ // Determine if feed is slow (for realtime stuff)
+ int real_time_output = mlt_properties_get_int( properties, "real_time" );
+
+ // Time structures
+ struct timeval ante;
// Get the frame rate
int fps = mlt_properties_get_double( properties, "fps" );
int samples = 0;
// AVFormat audio buffer and frame size
- int audio_outbuf_size = 10000;
- uint8_t *audio_outbuf = av_malloc( audio_outbuf_size );
+ int audio_outbuf_size = 10000;
+ uint8_t *audio_outbuf = av_malloc( audio_outbuf_size );
int audio_input_frame_size = 0;
// AVFormat video buffer and frame count
int frame_count = 0;
- int video_outbuf_size = 200000;
- uint8_t *video_outbuf = av_malloc(video_outbuf_size);
+ int video_outbuf_size = ( 1024 * 1024 );
+ uint8_t *video_outbuf = av_malloc( video_outbuf_size );
// Used for the frame properties
mlt_frame frame = NULL;
uint8_t *image;
mlt_image_format img_fmt = mlt_image_yuv422;
- // Fo receiving audio samples back from the fifo
- int16_t buffer[ 48000 * 2 ];
+ // For receiving audio samples back from the fifo
+ int16_t *buffer = av_malloc( 48000 * 2 );
int count = 0;
// Allocate the context
- AVFormatContext *oc = av_alloc_format_context();
+ AVFormatContext *oc = av_alloc_format_context( );
// Streams
- AVStream *audio_st = NULL;
+ AVStream *audio_st = NULL;
AVStream *video_st = NULL;
// Time stamps
- double audio_pts, video_pts;
+ double audio_pts, video_pts;
// Loop variable
int i;
+ // Frames despatched
+ long int frames = 0;
+ long int total_time = 0;
+
// Determine the format
- AVOutputFormat *fmt = NULL;
+ AVOutputFormat *fmt = NULL;
char *filename = mlt_properties_get( properties, "target" );
char *format = mlt_properties_get( properties, "format" );
- //char *vcodec = mlt_properties_get( properties, "vcodec" );
- //char *acodec = mlt_properties_get( properties, "acodec" );
+ char *vcodec = mlt_properties_get( properties, "vcodec" );
+ char *acodec = mlt_properties_get( properties, "acodec" );
+
+ // Used to store and override codec ids
+ int audio_codec_id;
+ int video_codec_id;
// Check for user selected format first
if ( format != NULL )
fmt = guess_format( "mpeg", NULL, NULL );
// We need a filename - default to stdout?
- if ( filename == NULL )
+ if ( filename == NULL || !strcmp( filename, "" ) )
filename = "pipe:";
+ // Get the codec ids selected
+ audio_codec_id = fmt->audio_codec;
+ video_codec_id = fmt->video_codec;
+
+ // Check for audio codec overides
+ if ( acodec != NULL )
+ {
+ AVCodec *p = first_avcodec;
+ while( p != NULL )
+ {
+ if ( !strcmp( p->name, acodec ) && p->type == CODEC_TYPE_AUDIO )
+ break;
+ p = p->next;
+ }
+ if ( p != NULL )
+ audio_codec_id = p->id;
+ else
+ fprintf( stderr, "consumer_avcodec: audio codec %s unrecognised - ignoring\n", acodec );
+ }
+
+ // Check for video codec overides
+ if ( vcodec != NULL )
+ {
+ AVCodec *p = first_avcodec;
+ while( p != NULL )
+ {
+ if ( !strcmp( p->name, vcodec ) && p->type == CODEC_TYPE_VIDEO )
+ break;
+ p = p->next;
+ }
+ if ( p != NULL )
+ video_codec_id = p->id;
+ else
+ fprintf( stderr, "consumer_avcodec: video codec %s unrecognised - ignoring\n", vcodec );
+ }
+
// Update the output context
oc->oformat = fmt;
- snprintf( oc->filename, sizeof(oc->filename), "%s", filename );
+ snprintf( oc->filename, sizeof(oc->filename), "%s", filename );
// Add audio and video streams
- if ( fmt->video_codec != CODEC_ID_NONE )
- video_st = add_video_stream( this, oc, fmt->video_codec );
- if ( fmt->audio_codec != CODEC_ID_NONE )
- audio_st = add_audio_stream( this, oc, fmt->audio_codec );
+ if ( fmt->video_codec != CODEC_ID_NONE )
+ video_st = add_video_stream( this, oc, video_codec_id );
+ if ( fmt->audio_codec != CODEC_ID_NONE )
+ audio_st = add_audio_stream( this, oc, audio_codec_id );
// Set the parameters (even though we have none...)
- if ( av_set_parameters(oc, NULL) >= 0 )
+ if ( av_set_parameters(oc, NULL) >= 0 )
{
- if ( video_st && !open_video( oc, video_st ) )
+ if ( video_st && !open_video( oc, video_st ) )
video_st = NULL;
- if ( audio_st )
- audio_input_frame_size = open_audio( oc, audio_st, audio_outbuf_size );
+ if ( audio_st )
+ audio_input_frame_size = open_audio( oc, audio_st, audio_outbuf_size );
- // Open the output file, if needed
- if ( !( fmt->flags & AVFMT_NOFILE ) )
+ // Open the output file, if needed
+ if ( !( fmt->flags & AVFMT_NOFILE ) )
{
- if (url_fopen(&oc->pb, filename, URL_WRONLY) < 0)
+ if (url_fopen(&oc->pb, filename, URL_WRONLY) < 0)
{
- fprintf(stderr, "Could not open '%s'\n", filename);
+ fprintf(stderr, "Could not open '%s'\n", filename);
mlt_properties_set_int( properties, "running", 0 );
- }
- }
-
- // Write the stream header, if any
+ }
+ }
+
+ // Write the stream header, if any
if ( mlt_properties_get_int( properties, "running" ) )
- av_write_header( oc );
+ av_write_header( oc );
}
else
{
- fprintf(stderr, "Invalid output format parameters\n");
+ fprintf(stderr, "Invalid output format parameters\n");
mlt_properties_set_int( properties, "running", 0 );
- }
+ }
// Last check - need at least one stream
if ( audio_st == NULL && video_st == NULL )
mlt_properties_set_int( properties, "running", 0 );
+ // Get the starting time (can ignore the times above)
+ gettimeofday( &ante, NULL );
+
// Loop while running
- while( mlt_properties_get_int( properties, "running" ) )
+ while( mlt_properties_get_int( properties, "running" ) && !terminated )
{
// Get the frame
frame = mlt_consumer_rt_frame( this );
// Check that we have a frame to work with
if ( frame != NULL )
{
+ // Increment frames despatched
+ frames ++;
+
// Default audio args
frame_properties = mlt_frame_properties( frame );
+ // Check for the terminated condition
+ terminated = terminate_on_pause && mlt_properties_get_double( frame_properties, "_speed" ) == 0.0;
+
// Get audio and append to the fifo
if ( audio_st )
{
samples = mlt_sample_calculator( fps, frequency, count );
mlt_frame_get_audio( frame, &pcm, &aud_fmt, &frequency, &channels, &samples );
+
+ // Create the fifo if we don't have one
+ if ( fifo == NULL )
+ {
+ fifo = sample_fifo_init( frequency, channels );
+ mlt_properties_set_data( properties, "sample_fifo", fifo, 0, ( mlt_destructor )sample_fifo_close, NULL );
+ }
+
+ // Append the samples
sample_fifo_append( fifo, pcm, samples * channels );
+ total_time += ( samples * 1000000 ) / frequency;
}
// Encode the image
mlt_deque_push_back( queue, frame );
else
mlt_frame_close( frame );
+ }
- // While we have stuff to process, process...
- while ( 1 )
+ // While we have stuff to process, process...
+ while ( 1 )
+ {
+ // Compute current audio and video time
+ if (audio_st)
+ audio_pts = (double)audio_st->pts.val * audio_st->time_base.num / audio_st->time_base.den;
+ else
+ audio_pts = 0.0;
+
+ if (video_st)
+ video_pts = (double)video_st->pts.val * video_st->time_base.num / video_st->time_base.den;
+ else
+ video_pts = 0.0;
+
+ // Write interleaved audio and video frames
+ if ( !video_st || ( video_st && audio_st && audio_pts < video_pts ) )
{
- // Compute current audio and video time
- if (audio_st)
- audio_pts = (double)audio_st->pts.val * oc->pts_num / oc->pts_den;
- else
- audio_pts = 0.0;
-
- if (video_st)
- video_pts = (double)video_st->pts.val * oc->pts_num / oc->pts_den;
- else
- video_pts = 0.0;
-
- // Write interleaved audio and video frames
- if ( !video_st || ( video_st && audio_st && audio_pts < video_pts ) )
+ if ( channels * audio_input_frame_size < sample_fifo_used( fifo ) )
{
- if ( channels * audio_input_frame_size < sample_fifo_used( fifo ) )
- {
- int out_size;
- AVCodecContext *c;
+ AVCodecContext *c;
+ AVPacket pkt;
+ av_init_packet( &pkt );
- c = &audio_st->codec;
+ c = &audio_st->codec;
- sample_fifo_fetch( fifo, buffer, channels * audio_input_frame_size );
+ sample_fifo_fetch( fifo, buffer, channels * audio_input_frame_size );
- out_size = avcodec_encode_audio( c, audio_outbuf, audio_outbuf_size, buffer );
+ pkt.size = avcodec_encode_audio( c, audio_outbuf, audio_outbuf_size, buffer );
+ // Write the compressed frame in the media file
+ pkt.pts= c->coded_frame->pts;
+ pkt.flags |= PKT_FLAG_KEY;
+ pkt.stream_index= audio_st->index;
+ pkt.data= audio_outbuf;
- // Write the compressed frame in the media file
- if (av_write_frame(oc, audio_st->index, audio_outbuf, out_size) != 0)
- fprintf(stderr, "Error while writing audio frame\n");
- }
- else
- {
- break;
- }
+ if ( av_write_frame( oc, &pkt ) != 0)
+ fprintf(stderr, "Error while writing audio frame\n");
}
- else if ( video_st )
+ else
{
- if ( mlt_deque_count( queue ) )
- {
- int out_size, ret;
- AVCodecContext *c;
-
- frame = mlt_deque_pop_front( queue );
- frame_properties = mlt_frame_properties( frame );
+ break;
+ }
+ }
+ else if ( video_st )
+ {
+ if ( mlt_deque_count( queue ) )
+ {
+ int out_size, ret;
+ AVCodecContext *c;
- if ( terminate_on_pause && mlt_properties_get_double( frame_properties, "_speed" ) == 0.0 )
- {
- mlt_properties_set_int( properties, "running", 0 );
- break;
- }
+ frame = mlt_deque_pop_front( queue );
+ frame_properties = mlt_frame_properties( frame );
- c = &video_st->codec;
-
- if ( mlt_properties_get_int( frame_properties, "rendered" ) )
- {
- int i = 0;
- int j = 0;
- uint8_t *p;
- uint8_t *q;
+ c = &video_st->codec;
+
+ if ( mlt_properties_get_int( frame_properties, "rendered" ) )
+ {
+ int i = 0;
+ int j = 0;
+ uint8_t *p;
+ uint8_t *q;
- mlt_frame_get_image( frame, &image, &img_fmt, &img_width, &img_height, 0 );
+ mlt_frame_get_image( frame, &image, &img_fmt, &img_width, &img_height, 0 );
- q = image;
+ q = image;
- for ( i = 0; i < height; i ++ )
+ for ( i = 0; i < height; i ++ )
+ {
+ p = input->data[ 0 ] + i * input->linesize[ 0 ];
+ j = width;
+ while( j -- )
{
- p = input->data[ 0 ] + i * input->linesize[ 0 ];
- j = width;
- while( j -- )
- {
- *p ++ = *q ++;
- *p ++ = *q ++;
- }
+ *p ++ = *q ++;
+ *p ++ = *q ++;
}
-
- img_convert( ( AVPicture * )output, PIX_FMT_YUV420P, ( AVPicture * )input, PIX_FMT_YUV422, width, height );
}
-
- if (oc->oformat->flags & AVFMT_RAWPICTURE)
- {
- // raw video case. The API will change slightly in the near future for that
- ret = av_write_frame(oc, video_st->index, (uint8_t *)output, sizeof(AVPicture));
- }
- else
- {
- // Encode the image
- out_size = avcodec_encode_video(c, video_outbuf, video_outbuf_size, output );
- // If zero size, it means the image was buffered
- if (out_size != 0)
- {
- // write the compressed frame in the media file
- // XXX: in case of B frames, the pts is not yet valid
- ret = av_write_frame( oc, video_st->index, video_outbuf, out_size );
- }
- }
- frame_count++;
- mlt_frame_close( frame );
+ img_convert( ( AVPicture * )output, PIX_FMT_YUV420P, ( AVPicture * )input, PIX_FMT_YUV422, width, height );
}
- else
+
+ if (oc->oformat->flags & AVFMT_RAWPICTURE)
{
- break;
- }
+ // raw video case. The API will change slightly in the near future for that
+ AVPacket pkt;
+ av_init_packet(&pkt);
+
+ pkt.flags |= PKT_FLAG_KEY;
+ pkt.stream_index= video_st->index;
+ pkt.data= (uint8_t *)output;
+ pkt.size= sizeof(AVPicture);
+
+ ret = av_write_frame(oc, &pkt);
+ }
+ else
+ {
+ // Set the quality
+ output->quality = video_st->quality;
+
+ // Encode the image
+ out_size = avcodec_encode_video(c, video_outbuf, video_outbuf_size, output );
+
+ // If zero size, it means the image was buffered
+ if (out_size != 0)
+ {
+ AVPacket pkt;
+ av_init_packet( &pkt );
+
+ pkt.pts= c->coded_frame->pts;
+ if(c->coded_frame->key_frame)
+ pkt.flags |= PKT_FLAG_KEY;
+ pkt.stream_index= video_st->index;
+ pkt.data= video_outbuf;
+ pkt.size= out_size;
+
+ // write the compressed frame in the media file
+ ret = av_write_frame(oc, &pkt);
+ }
+ }
+ frame_count++;
+ mlt_frame_close( frame );
}
+ else
+ {
+ break;
+ }
+ }
+ }
+
+ if ( real_time_output && frames % 25 == 0 )
+ {
+ long passed = time_difference( &ante );
+ if ( fifo != NULL )
+ {
+ long pending = ( ( ( long )sample_fifo_used( fifo ) * 1000 ) / frequency ) * 1000;
+ passed -= pending;
+ }
+ if ( passed < total_time )
+ {
+ long total = ( total_time - passed );
+ struct timespec t = { total / 1000000, ( total % 1000000 ) * 1000 };
+ nanosleep( &t, NULL );
}
}
}
- // close each codec
- if (video_st)
- close_video(oc, video_st);
- if (audio_st)
- close_audio(oc, audio_st);
+ // close each codec
+ if (video_st)
+ close_video(oc, video_st);
+ if (audio_st)
+ close_audio(oc, audio_st);
- // Write the trailer, if any
- av_write_trailer(oc);
-
- // Free the streams
- for(i = 0; i < oc->nb_streams; i++)
- av_freep(&oc->streams[i]);
+ // Write the trailer, if any
+ av_write_trailer(oc);
+
+ // Free the streams
+ for(i = 0; i < oc->nb_streams; i++)
+ av_freep(&oc->streams[i]);
// Close the output file
- if (!(fmt->flags & AVFMT_NOFILE))
- url_fclose(&oc->pb);
+ if (!(fmt->flags & AVFMT_NOFILE))
+ url_fclose(&oc->pb);
// Clean up input and output frames
- av_free( output->data[0] );
- av_free( output );
- av_free( input->data[0] );
- av_free( input );
- av_free( video_outbuf );
+ av_free( output->data[0] );
+ av_free( output );
+ av_free( input->data[0] );
+ av_free( input );
+ av_free( video_outbuf );
+ av_free( buffer );
+
+ // Free the stream
+ av_free(oc);
- // Free the stream
- av_free(oc);
+ // Just in case we terminated on pause
+ mlt_properties_set_int( properties, "running", 0 );
return NULL;
}
// Free the memory
free( this );
}
-