--- /dev/null
+/*
+ * valerie_tokeniser.c -- String tokeniser
+ * Copyright (C) 2002-2003 Ushodaya Enterprises Limited
+ * Author: Charles Yates <charles.yates@pandora.be>
+ *
+ * This library is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Lesser General Public
+ * License as published by the Free Software Foundation; either
+ * version 2.1 of the License, or (at your option) any later version.
+ *
+ * This library is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ * Lesser General Public License for more details.
+ *
+ * You should have received a copy of the GNU Lesser General Public
+ * License along with this library; if not, write to the Free Software
+ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
+ */
+
+/* System header files */
+#include <stdlib.h>
+#include <string.h>
+
+/* Application header files */
+#include "valerie_tokeniser.h"
+
+/** Initialise a tokeniser.
+*/
+
+valerie_tokeniser valerie_tokeniser_init( )
+{
+ valerie_tokeniser tokeniser = malloc( sizeof( valerie_tokeniser_t ) );
+ if ( tokeniser != NULL )
+ memset( tokeniser, 0, sizeof( valerie_tokeniser_t ) );
+ return tokeniser;
+}
+
+/** Clear the tokeniser.
+*/
+
+static void valerie_tokeniser_clear( valerie_tokeniser tokeniser )
+{
+ int index = 0;
+ for ( index = 0; index < tokeniser->count; index ++ )
+ free( tokeniser->tokens[ index ] );
+ tokeniser->count = 0;
+ free( tokeniser->input );
+ tokeniser->input = NULL;
+}
+
+/** Append a string to the tokeniser.
+*/
+
+static int valerie_tokeniser_append( valerie_tokeniser tokeniser, char *token )
+{
+ int error = 0;
+
+ if ( tokeniser->count == tokeniser->size )
+ {
+ tokeniser->size += 20;
+ tokeniser->tokens = realloc( tokeniser->tokens, tokeniser->size * sizeof( char * ) );
+ }
+
+ if ( tokeniser->tokens != NULL )
+ {
+ tokeniser->tokens[ tokeniser->count ++ ] = strdup( token );
+ }
+ else
+ {
+ tokeniser->count = 0;
+ error = -1;
+ }
+ return error;
+}
+
+/** Parse a string by splitting on the delimiter provided.
+*/
+
+int valerie_tokeniser_parse_new( valerie_tokeniser tokeniser, char *string, const char *delimiter )
+{
+ int count = 0;
+ int length = strlen( string );
+ int delimiter_size = strlen( delimiter );
+ int index = 0;
+ char *token = strdup( string );
+
+ valerie_tokeniser_clear( tokeniser );
+ tokeniser->input = strdup( string );
+ strcpy( token, "" );
+
+ for ( index = 0; index < length; )
+ {
+ char *start = string + index;
+ char *end = strstr( start, delimiter );
+
+ if ( end == NULL )
+ {
+ strcat( token, start );
+ valerie_tokeniser_append( tokeniser, token );
+ index = length;
+ count ++;
+ }
+ else if ( start != end )
+ {
+ strncat( token, start, end - start );
+ index += end - start;
+ if ( token[ 0 ] != '\"' || ( token[ 0 ] == '\"' && token[ strlen( token ) - 1 ] == '\"' ) )
+ {
+ valerie_tokeniser_append( tokeniser, token );
+ strcpy( token, "" );
+ count ++;
+ }
+ else while ( strncmp( string + index, delimiter, delimiter_size ) == 0 )
+ {
+ strncat( token, delimiter, delimiter_size );
+ index += delimiter_size;
+ }
+ }
+ else
+ {
+ index += strlen( delimiter );
+ }
+ }
+
+ /* Special case - malformed string condition */
+ if ( !strcmp( token, "" ) )
+ {
+ count = 0 - ( count - 1 );
+ valerie_tokeniser_append( tokeniser, token );
+ }
+
+ free( token );
+ return count;
+}
+
+/** Get the original input.
+*/
+
+char *valerie_tokeniser_get_input( valerie_tokeniser tokeniser )
+{
+ return tokeniser->input;
+}
+
+/** Get the number of tokens.
+*/
+
+int valerie_tokeniser_count( valerie_tokeniser tokeniser )
+{
+ return tokeniser->count;
+}
+
+/** Get a token as a string.
+*/
+
+char *valerie_tokeniser_get_string( valerie_tokeniser tokeniser, int index )
+{
+ if ( index < tokeniser->count )
+ return tokeniser->tokens[ index ];
+ else
+ return NULL;
+}
+
+/** Close the tokeniser.
+*/
+
+void valerie_tokeniser_close( valerie_tokeniser tokeniser )
+{
+ valerie_tokeniser_clear( tokeniser );
+ free( tokeniser->tokens );
+ free( tokeniser );
+}