X-Git-Url: http://research.m1stereo.tv/gitweb?a=blobdiff_plain;f=src%2Fmodules%2Fcore%2Ftransition_composite.c;h=fdad2cfec6e1fa140aebb428621ec05108baf045;hb=bf3264b9e340ba5c11cbf59835a8af3db94e0cc2;hp=21909e8e026a7d945c930d699dc4e29114dd2308;hpb=e6c03148e1dc3b0e363c9f304b3ed93fbc1a91a9;p=melted diff --git a/src/modules/core/transition_composite.c b/src/modules/core/transition_composite.c index 21909e8..fdad2cf 100644 --- a/src/modules/core/transition_composite.c +++ b/src/modules/core/transition_composite.c @@ -3,19 +3,19 @@ * Copyright (C) 2003-2004 Ushodaya Enterprises Limited * Author: Dan Dennedy * - * This program is free software; you can redistribute it and/or modify - * it under the terms of the GNU General Public License as published by - * the Free Software Foundation; either version 2 of the License, or - * (at your option) any later version. + * This library is free software; you can redistribute it and/or + * modify it under the terms of the GNU Lesser General Public + * License as published by the Free Software Foundation; either + * version 2.1 of the License, or (at your option) any later version. * - * This program is distributed in the hope that it will be useful, + * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of - * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the - * GNU General Public License for more details. + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Lesser General Public License for more details. * - * You should have received a copy of the GNU General Public License - * along with this program; if not, write to the Free Software Foundation, - * Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. + * You should have received a copy of the GNU Lesser General Public + * License along with this library; if not, write to the Free Software + * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA */ #include "transition_composite.h" @@ -366,87 +366,81 @@ static void luma_read_yuv422( uint8_t *image, uint16_t **map, int width, int hei *p++ = ( image[ i ] - 16 ) * 299; // 299 = 65535 / 219 } +static inline int calculate_mix( uint16_t *luma, int j, int soft, int weight, int alpha ) +{ + return ( ( ( luma == NULL ) ? weight : smoothstep( luma[ j ], luma[ j ] + soft, weight + soft ) ) * alpha ) >> 8; +} + +static inline uint8_t sample_mix( uint8_t dest, uint8_t src, int mix ) +{ + return ( src * mix + dest * ( ( 1 << 16 ) - mix ) ) >> 16; +} /** Composite a source line over a destination line */ -static void composite_line_yuv( uint8_t *dest, uint8_t *src, int width, uint8_t *alpha_b, uint8_t *alpha_a, int weight, uint16_t *luma, int softness ) +static void composite_line_yuv( uint8_t *dest, uint8_t *src, int width, uint8_t *alpha_b, uint8_t *alpha_a, int weight, uint16_t *luma, int soft ) { register int j; - register int a; register int mix; - + for ( j = 0; j < width; j ++ ) { - a = *alpha_b ++; - mix = ( luma == NULL ) ? weight : smoothstep( luma[ j ], luma[ j ] + softness, weight + softness ); - mix = ( mix * a ) >> 8; - *dest = ( *src++ * mix + *dest * ( ( 1 << 16 ) - mix ) ) >> 16; + mix = calculate_mix( luma, j, soft, weight, *alpha_b ++ ); + *dest = sample_mix( *dest, *src++, mix ); dest++; - *dest = ( *src++ * mix + *dest * ( ( 1 << 16 ) - mix ) ) >> 16; + *dest = sample_mix( *dest, *src++, mix ); dest++; - *alpha_a = mix | *alpha_a; + *alpha_a = ( mix >> 8 ) | *alpha_a; alpha_a ++; } } -static void composite_line_yuv_or( uint8_t *dest, uint8_t *src, int width, uint8_t *alpha_b, uint8_t *alpha_a, int weight, uint16_t *luma, int softness ) +static void composite_line_yuv_or( uint8_t *dest, uint8_t *src, int width, uint8_t *alpha_b, uint8_t *alpha_a, int weight, uint16_t *luma, int soft ) { register int j; - register int a; register int mix; - + for ( j = 0; j < width; j ++ ) { - a = *alpha_b ++ | *alpha_a; - mix = ( luma == NULL ) ? weight : smoothstep( luma[ j ], luma[ j ] + softness, weight + softness ); - mix = ( mix * a ) >> 8; - *dest = ( *src++ * mix + *dest * ( ( 1 << 16 ) - mix ) ) >> 16; + mix = calculate_mix( luma, j, soft, weight, *alpha_b ++ | *alpha_a ); + *dest = sample_mix( *dest, *src++, mix ); dest++; - *dest = ( *src++ * mix + *dest * ( ( 1 << 16 ) - mix ) ) >> 16; + *dest = sample_mix( *dest, *src++, mix ); dest++; - *alpha_a = mix | *alpha_a; - alpha_a ++; + *alpha_a ++ = mix >> 8; } } -static void composite_line_yuv_and( uint8_t *dest, uint8_t *src, int width, uint8_t *alpha_b, uint8_t *alpha_a, int weight, uint16_t *luma, int softness ) +static void composite_line_yuv_and( uint8_t *dest, uint8_t *src, int width, uint8_t *alpha_b, uint8_t *alpha_a, int weight, uint16_t *luma, int soft ) { register int j; - register int a; register int mix; - + for ( j = 0; j < width; j ++ ) { - a = *alpha_b ++ & *alpha_a; - mix = ( luma == NULL ) ? weight : smoothstep( luma[ j ], luma[ j ] + softness, weight + softness ); - mix = ( mix * a ) >> 8; - *dest = ( *src++ * mix + *dest * ( ( 1 << 16 ) - mix ) ) >> 16; + mix = calculate_mix( luma, j, soft, weight, *alpha_b ++ & *alpha_a ); + *dest = sample_mix( *dest, *src++, mix ); dest++; - *dest = ( *src++ * mix + *dest * ( ( 1 << 16 ) - mix ) ) >> 16; + *dest = sample_mix( *dest, *src++, mix ); dest++; - *alpha_a = mix | *alpha_a; - alpha_a ++; + *alpha_a ++ = mix >> 8; } } -static void composite_line_yuv_xor( uint8_t *dest, uint8_t *src, int width, uint8_t *alpha_b, uint8_t *alpha_a, int weight, uint16_t *luma, int softness ) +static void composite_line_yuv_xor( uint8_t *dest, uint8_t *src, int width, uint8_t *alpha_b, uint8_t *alpha_a, int weight, uint16_t *luma, int soft ) { register int j; - register int a; register int mix; - + for ( j = 0; j < width; j ++ ) { - a = *alpha_b ++ ^ *alpha_a; - mix = ( luma == NULL ) ? weight : smoothstep( luma[ j ], luma[ j ] + softness, weight + softness ); - mix = ( mix * a ) >> 8; - *dest = ( *src++ * mix + *dest * ( ( 1 << 16 ) - mix ) ) >> 16; + mix = calculate_mix( luma, j, soft, weight, *alpha_b ++ ^ *alpha_a ); + *dest = sample_mix( *dest, *src++, mix ); dest++; - *dest = ( *src++ * mix + *dest * ( ( 1 << 16 ) - mix ) ) >> 16; + *dest = sample_mix( *dest, *src++, mix ); dest++; - *alpha_a = mix | *alpha_a; - alpha_a ++; + *alpha_a ++ = mix >> 8; } } @@ -458,7 +452,7 @@ static int composite_yuv( uint8_t *p_dest, int width_dest, int height_dest, uint int ret = 0; int i; int x_src = 0, y_src = 0; - int32_t weight = ( 1 << 16 ) * ( geometry.item.mix / 100 ); + int32_t weight = ( ( 1 << 16 ) - 1 ) * ( geometry.item.mix / 100 ); int step = ( field > -1 ) ? 2 : 1; int bpp = 2; int stride_src = width_src * bpp; @@ -467,7 +461,7 @@ static int composite_yuv( uint8_t *p_dest, int width_dest, int height_dest, uint // Adjust to consumer scale int x = rint( 0.5 + geometry.item.x * width_dest / geometry.nw ); int y = rint( 0.5 + geometry.item.y * height_dest / geometry.nh ); - int x_uneven = x & 1; + int uneven_x = ( x % 2 ); // optimization points - no work to do if ( width_src <= 0 || height_src <= 0 ) @@ -540,13 +534,10 @@ static int composite_yuv( uint8_t *p_dest, int width_dest, int height_dest, uint int alpha_b_stride = stride_src / bpp; int alpha_a_stride = stride_dest / bpp; - // Make sure than x and w are even - if ( x_uneven ) - { - p_src += 2; - width_src --; - alpha_a ++; - } + p_src += uneven_x * 2; + width_src -= 2 * uneven_x; + alpha_b += uneven_x; + uneven_x = 0; // now do the compositing only to cropped extents for ( i = 0; i < height_src; i += step ) @@ -724,44 +715,46 @@ static int get_b_frame_image( mlt_transition this, mlt_frame b_frame, uint8_t ** // Get the properties objects mlt_properties b_props = MLT_FRAME_PROPERTIES( b_frame ); mlt_properties properties = MLT_TRANSITION_PROPERTIES( this ); + uint8_t resize_alpha = mlt_properties_get_int( b_props, "resize_alpha" ); - if ( mlt_properties_get_int( properties, "distort" ) == 0 && mlt_properties_get_int( b_props, "distort" ) == 0 && geometry->item.distort == 0 ) + if ( mlt_properties_get_int( properties, "aligned" ) && mlt_properties_get_int( properties, "distort" ) == 0 && mlt_properties_get_int( b_props, "distort" ) == 0 && geometry->item.distort == 0 ) { // Adjust b_frame pixel aspect int normalised_width = geometry->item.w; int normalised_height = geometry->item.h; int real_width = get_value( b_props, "real_width", "width" ); int real_height = get_value( b_props, "real_height", "height" ); - double input_ar = mlt_frame_get_aspect_ratio( b_frame ); - double output_ar = mlt_properties_get_double( b_props, "consumer_aspect_ratio" ); - if ( input_ar == 0.0 ) input_ar = output_ar; - int scaled_width = input_ar / output_ar * real_width; + double input_ar = mlt_properties_get_double( b_props, "aspect_ratio" ); + double consumer_ar = mlt_properties_get_double( b_props, "consumer_aspect_ratio" ); + double background_ar = mlt_properties_get_double( b_props, "output_ratio" ); + double output_ar = background_ar != 0.0 ? background_ar : consumer_ar; + int scaled_width = rint( 0.5 + ( input_ar == 0.0 ? output_ar : input_ar ) / output_ar * real_width ); int scaled_height = real_height; - + // Now ensure that our images fit in the normalised frame if ( scaled_width > normalised_width ) { - scaled_height = scaled_height * normalised_width / scaled_width; + scaled_height = rint( 0.5 + scaled_height * normalised_width / scaled_width ); scaled_width = normalised_width; } if ( scaled_height > normalised_height ) { - scaled_width = scaled_width * normalised_height / scaled_height; + scaled_width = rint( 0.5 + scaled_width * normalised_height / scaled_height ); scaled_height = normalised_height; } // Honour the fill request - this will scale the image to fill width or height while maintaining a/r // ????: Shouln't this be the default behaviour? - if ( mlt_properties_get_int( properties, "fill" ) ) + if ( mlt_properties_get_int( properties, "fill" ) && scaled_width > 0 && scaled_height > 0 ) { - if ( scaled_height < normalised_height && scaled_width * normalised_height / scaled_height < normalised_width ) + if ( scaled_height < normalised_height && scaled_width * normalised_height / scaled_height <= normalised_width ) { - scaled_width = scaled_width * normalised_height / scaled_height; + scaled_width = rint( 0.5 + scaled_width * normalised_height / scaled_height ); scaled_height = normalised_height; } else if ( scaled_width < normalised_width && scaled_height * normalised_width / scaled_width < normalised_height ) { - scaled_height = scaled_height * normalised_width / scaled_width; + scaled_height = rint( 0.5 + scaled_height * normalised_width / scaled_width ); scaled_width = normalised_width; } } @@ -777,18 +770,26 @@ static int get_b_frame_image( mlt_transition this, mlt_frame b_frame, uint8_t ** } // We want to ensure that we bypass resize now... - mlt_properties_set_int( b_props, "distort", 1 ); + if ( resize_alpha == 0 ) + mlt_properties_set_int( b_props, "distort", mlt_properties_get_int( properties, "distort" ) ); + + // If we're not aligned, we want a non-transparent background + if ( mlt_properties_get_int( properties, "aligned" ) == 0 ) + mlt_properties_set_int( b_props, "resize_alpha", 255 ); - // Take into consideration alignment for optimisation + // Take into consideration alignment for optimisation (titles are a special case) if ( !mlt_properties_get_int( properties, "titles" ) ) alignment_calculate( geometry ); // Adjust to consumer scale - *width = geometry->sw * *width / geometry->nw; - *height = geometry->sh * *height / geometry->nh; + *width = rint( 0.5 + geometry->sw * *width / geometry->nw ); + *height = rint( 0.5 + geometry->sh * *height / geometry->nh ); ret = mlt_frame_get_image( b_frame, image, &format, width, height, 1 ); + // Set the frame back + mlt_properties_set_int( b_props, "resize_alpha", resize_alpha ); + return ret && image != NULL; } @@ -808,32 +809,43 @@ static mlt_geometry composite_calculate( mlt_transition this, struct geometry_s int normalised_width = mlt_properties_get_int( a_props, "normalised_width" ); int normalised_height = mlt_properties_get_int( a_props, "normalised_height" ); - // Now parse the geometries - if ( start == NULL ) - { - // Parse the transitions properties - start = transition_parse_keys( this, normalised_width, normalised_height ); + char *name = mlt_properties_get( properties, "_unique_id" ); + char key[ 256 ]; - // Assign to properties to ensure we get destroyed - mlt_properties_set_data( properties, "geometries", start, 0, ( mlt_destructor )mlt_geometry_close, NULL ); + sprintf( key, "%s.in", name ); + if ( mlt_properties_get( a_props, key ) ) + { + sscanf( mlt_properties_get( a_props, key ), "%f,%f,%f,%f,%f,%d,%d", &result->item.x, &result->item.y, &result->item.w, &result->item.h, &result->item.mix, &result->nw, &result->nh ); } else { - int length = mlt_transition_get_out( this ) - mlt_transition_get_in( this ) + 1; - double cycle = mlt_properties_get_double( properties, "cycle" ); - if ( cycle > 1 ) - length = cycle; - else if ( cycle > 0 ) - length *= cycle; - mlt_geometry_refresh( start, mlt_properties_get( properties, "geometry" ), length, normalised_width, normalised_height ); - } + // Now parse the geometries + if ( start == NULL ) + { + // Parse the transitions properties + start = transition_parse_keys( this, normalised_width, normalised_height ); - // Do the calculation - geometry_calculate( this, result, position ); + // Assign to properties to ensure we get destroyed + mlt_properties_set_data( properties, "geometries", start, 0, ( mlt_destructor )mlt_geometry_close, NULL ); + } + else + { + int length = mlt_transition_get_out( this ) - mlt_transition_get_in( this ) + 1; + double cycle = mlt_properties_get_double( properties, "cycle" ); + if ( cycle > 1 ) + length = cycle; + else if ( cycle > 0 ) + length *= cycle; + mlt_geometry_refresh( start, mlt_properties_get( properties, "geometry" ), length, normalised_width, normalised_height ); + } + + // Do the calculation + geometry_calculate( this, result, position ); - // Assign normalised info - result->nw = normalised_width; - result->nh = normalised_height; + // Assign normalised info + result->nw = normalised_width; + result->nh = normalised_height; + } // Now parse the alignment result->halign = alignment_parse( mlt_properties_get( properties, "halign" ) ); @@ -842,16 +854,6 @@ static mlt_geometry composite_calculate( mlt_transition this, struct geometry_s return start; } -static inline void inline_memcpy( uint8_t *dest, uint8_t *src, int length ) -{ - uint8_t *end = src + length; - while ( src < end ) - { - *dest ++ = *src ++; - *dest ++ = *src ++; - } -} - mlt_frame composite_copy_region( mlt_transition this, mlt_frame a_frame, mlt_position frame_position ) { // Create a frame to return @@ -866,6 +868,10 @@ mlt_frame composite_copy_region( mlt_transition this, mlt_frame a_frame, mlt_pos // Get the position int position = position_calculate( this, frame_position ); + // Get the unique id of the transition + char *name = mlt_properties_get( MLT_TRANSITION_PROPERTIES( this ), "_unique_id" ); + char key[ 256 ]; + // Destination image uint8_t *dest = NULL; @@ -873,6 +879,7 @@ mlt_frame composite_copy_region( mlt_transition this, mlt_frame a_frame, mlt_pos uint8_t *image = mlt_properties_get_data( a_props, "image", NULL ); int width = mlt_properties_get_int( a_props, "width" ); int height = mlt_properties_get_int( a_props, "height" ); + int format = mlt_properties_get_int( a_props, "format" ); // Pointers for copy operation uint8_t *p; @@ -889,10 +896,8 @@ mlt_frame composite_copy_region( mlt_transition this, mlt_frame a_frame, mlt_pos // Will need to know region to copy struct geometry_s result; - double delta = delta_calculate( this, a_frame, frame_position ); - // Calculate the region now - composite_calculate( this, &result, a_frame, position + delta / 2 ); + composite_calculate( this, &result, a_frame, position ); // Need to scale down to actual dimensions x = rint( 0.5 + result.item.x * width / result.nw ); @@ -900,19 +905,18 @@ mlt_frame composite_copy_region( mlt_transition this, mlt_frame a_frame, mlt_pos w = rint( 0.5 + result.item.w * width / result.nw ); h = rint( 0.5 + result.item.h * height / result.nh ); - // Make sure that x and w are even - if ( x & 1 ) + if ( x % 2 ) { x --; - w += 2; - if ( w & 1 ) - w --; - } - else if ( w & 1 ) - { w ++; } + // Store the key + sprintf( key, "%s.in=%d,%d,%d,%d,%f,%d,%d", name, x, y, w, h, result.item.mix, width, height ); + mlt_properties_parse( a_props, key ); + sprintf( key, "%s.out=%d,%d,%d,%d,%f,%d,%d", name, x, y, w, h, result.item.mix, width, height ); + mlt_properties_parse( a_props, key ); + ds = w * 2; ss = width * 2; @@ -923,6 +927,7 @@ mlt_frame composite_copy_region( mlt_transition this, mlt_frame a_frame, mlt_pos mlt_properties_set_data( b_props, "image", dest, w * h * 2, mlt_pool_release, NULL ); mlt_properties_set_int( b_props, "width", w ); mlt_properties_set_int( b_props, "height", h ); + mlt_properties_set_int( b_props, "format", format ); if ( y < 0 ) { @@ -948,7 +953,7 @@ mlt_frame composite_copy_region( mlt_transition this, mlt_frame a_frame, mlt_pos while ( h -- ) { - inline_memcpy( dest, p, w * 2 ); + memcpy( dest, p, w * 2 ); dest += ds; p += ss; } @@ -1012,6 +1017,10 @@ static int transition_get_image( mlt_frame a_frame, uint8_t **image, mlt_image_f int width_b = *width; int height_b = *height; + // Vars for alphas + uint8_t *alpha_a = NULL; + uint8_t *alpha_b = NULL; + // Composites always need scaling... defaulting to lowest char *rescale = mlt_properties_get( a_props, "rescale.interp" ); if ( rescale == NULL || !strcmp( rescale, "none" ) ) @@ -1020,14 +1029,16 @@ static int transition_get_image( mlt_frame a_frame, uint8_t **image, mlt_image_f mlt_properties_set( b_props, "rescale.interp", rescale ); // Do the calculation + // NB: Locks needed here since the properties are being modified + mlt_service_lock( MLT_TRANSITION_SERVICE( this ) ); composite_calculate( this, &result, a_frame, position ); + mlt_service_unlock( MLT_TRANSITION_SERVICE( this ) ); // Since we are the consumer of the b_frame, we must pass along these // consumer properties from the a_frame mlt_properties_set_double( b_props, "consumer_deinterlace", mlt_properties_get_double( a_props, "consumer_deinterlace" ) ); + mlt_properties_set( b_props, "consumer_deinterlace_method", mlt_properties_get( a_props, "consumer_deinterlace_method" ) ); mlt_properties_set_double( b_props, "consumer_aspect_ratio", mlt_properties_get_double( a_props, "consumer_aspect_ratio" ) ); - mlt_properties_set_int( b_props, "normalised_width", mlt_properties_get_double( a_props, "normalised_width" ) ); - mlt_properties_set_int( b_props, "normalised_height", mlt_properties_get_double( a_props, "normalised_height" ) ); // TODO: Dangerous/temporary optimisation - if nothing to do, then do nothing if ( mlt_properties_get_int( properties, "no_alpha" ) && @@ -1040,10 +1051,16 @@ static int transition_get_image( mlt_frame a_frame, uint8_t **image, mlt_image_f } if ( a_frame == b_frame ) + { + double aspect_ratio = mlt_frame_get_aspect_ratio( b_frame ); get_b_frame_image( this, b_frame, &image_b, &width_b, &height_b, &result ); + alpha_b = mlt_frame_get_alpha_mask( b_frame ); + mlt_properties_set_double( a_props, "aspect_ratio", aspect_ratio ); + } // Get the image from the a frame mlt_frame_get_image( a_frame, image, format, width, height, 1 ); + alpha_a = mlt_frame_get_alpha_mask( a_frame ); // Optimisation - no compositing required if ( result.item.mix == 0 || ( result.item.w == 0 && result.item.h == 0 ) ) @@ -1072,12 +1089,10 @@ static int transition_get_image( mlt_frame a_frame, uint8_t **image, mlt_image_f height_b = mlt_properties_get_int( a_props, "dest_height" ); } - if ( image_b != NULL || get_b_frame_image( this, b_frame, &image_b, &width_b, &height_b, &result ) == 0 ) + if ( *image != image_b && ( image_b != NULL || get_b_frame_image( this, b_frame, &image_b, &width_b, &height_b, &result ) == 0 ) ) { uint8_t *dest = *image; uint8_t *src = image_b; - uint8_t *alpha_b = mlt_frame_get_alpha_mask( b_frame ); - uint8_t *alpha_a = mlt_frame_get_alpha_mask( a_frame ); int progressive = mlt_properties_get_int( a_props, "consumer_deinterlace" ) || mlt_properties_get_int( properties, "progressive" ); @@ -1085,15 +1100,24 @@ static int transition_get_image( mlt_frame a_frame, uint8_t **image, mlt_image_f int32_t luma_softness = mlt_properties_get_double( properties, "softness" ) * ( 1 << 16 ); uint16_t *luma_bitmap = get_luma( properties, width_b, height_b ); + char *operator = mlt_properties_get( properties, "operator" ); + + alpha_b = alpha_b == NULL ? mlt_frame_get_alpha_mask( b_frame ) : alpha_b; + composite_line_fn line_fn = composite_line_yuv; - if ( mlt_properties_get_int( properties, "or" ) ) - line_fn = composite_line_yuv_or; - if ( mlt_properties_get_int( properties, "and" ) ) - line_fn = composite_line_yuv_and; - if ( mlt_properties_get_int( properties, "xor" ) ) - line_fn = composite_line_yuv_xor; + // Replacement and override + if ( operator != NULL ) + { + if ( !strcmp( operator, "or" ) ) + line_fn = composite_line_yuv_or; + if ( !strcmp( operator, "and" ) ) + line_fn = composite_line_yuv_and; + if ( !strcmp( operator, "xor" ) ) + line_fn = composite_line_yuv_xor; + } + // Allow the user to completely obliterate the alpha channels from both frames if ( mlt_properties_get( properties, "alpha_a" ) ) memset( alpha_a, mlt_properties_get_int( properties, "alpha_a" ), *width * *height ); @@ -1106,13 +1130,16 @@ static int transition_get_image( mlt_frame a_frame, uint8_t **image, mlt_image_f double field_position = position + field * delta; // Do the calculation if we need to + // NB: Locks needed here since the properties are being modified + mlt_service_lock( MLT_TRANSITION_SERVICE( this ) ); composite_calculate( this, &result, a_frame, field_position ); + mlt_service_unlock( MLT_TRANSITION_SERVICE( this ) ); if ( mlt_properties_get_int( properties, "titles" ) ) { - result.item.w = *width * ( result.item.w / result.nw ); + result.item.w = rint( 0.5 + *width * ( result.item.w / result.nw ) ); result.nw = result.item.w; - result.item.h = *height * ( result.item.h / result.nh ); + result.item.h = rint( 0.5 + *height * ( result.item.h / result.nh ) ); result.nh = *height; result.sw = width_b; result.sh = height_b; @@ -1178,12 +1205,11 @@ mlt_transition transition_composite_init( char *arg ) // Default factory mlt_properties_set( properties, "factory", "fezzik" ); + // Use alignment (and hence alpha of b frame) + mlt_properties_set_int( properties, "aligned", 1 ); + // Inform apps and framework that this is a video only transition mlt_properties_set_int( properties, "_transition_type", 1 ); - -#ifdef USE_MMX - //mlt_properties_set_int( properties, "_MMX", composite_have_mmx() ); -#endif } return this; }