Server IP : 103.119.228.120 / Your IP : 3.17.155.142 Web Server : Apache System : Linux v8.techscape8.com 3.10.0-1160.119.1.el7.tuxcare.els2.x86_64 #1 SMP Mon Jul 15 12:09:18 UTC 2024 x86_64 User : nobody ( 99) PHP Version : 5.6.40 Disable Function : shell_exec,symlink,system,exec,proc_get_status,proc_nice,proc_terminate,define_syslog_variables,syslog,openlog,closelog,escapeshellcmd,passthru,ocinum cols,ini_alter,leak,listen,chgrp,apache_note,apache_setenv,debugger_on,debugger_off,ftp_exec,dl,dll,myshellexec,proc_open,socket_bind,proc_close,escapeshellarg,parse_ini_filepopen,fpassthru,exec,passthru,escapeshellarg,escapeshellcmd,proc_close,proc_open,ini_alter,popen,show_source,proc_nice,proc_terminate,proc_get_status,proc_close,pfsockopen,leak,apache_child_terminate,posix_kill,posix_mkfifo,posix_setpgid,posix_setsid,posix_setuid,dl,symlink,shell_exec,system,dl,passthru,escapeshellarg,escapeshellcmd,myshellexec,c99_buff_prepare,c99_sess_put,fpassthru,getdisfunc,fx29exec,fx29exec2,is_windows,disp_freespace,fx29sh_getupdate,fx29_buff_prepare,fx29_sess_put,fx29shexit,fx29fsearch,fx29ftpbrutecheck,fx29sh_tools,fx29sh_about,milw0rm,imagez,sh_name,myshellexec,checkproxyhost,dosyayicek,c99_buff_prepare,c99_sess_put,c99getsource,c99sh_getupdate,c99fsearch,c99shexit,view_perms,posix_getpwuid,posix_getgrgid,posix_kill,parse_perms,parsesort,view_perms_color,set_encoder_input,ls_setcheckboxall,ls_reverse_all,rsg_read,rsg_glob,selfURL,dispsecinfo,unix2DosTime,addFile,system,get_users,view_size,DirFiles,DirFilesWide,DirPrintHTMLHeaders,GetFilesTotal,GetTitles,GetTimeTotal,GetMatchesCount,GetFileMatchesCount,GetResultFiles,fs_copy_dir,fs_copy_obj,fs_move_dir,fs_move_obj,fs_rmdir,SearchText,getmicrotime MySQL : ON | cURL : ON | WGET : ON | Perl : ON | Python : ON | Sudo : ON | Pkexec : ON Directory : /usr/local/ssl/local/ssl/local/ssl/local/ssl/local/ssl/local/share/perl5/Regexp/ |
Upload File : |
# Regexp::Assemple.pm # # Copyright (c) 2004-2011 David Landgren # All rights reserved package Regexp::Assemble; use strict; use warnings; use constant DEBUG_ADD => 1; use constant DEBUG_TAIL => 2; use constant DEBUG_LEX => 4; use constant DEBUG_TIME => 8; use vars qw/$have_Storable $Current_Lexer $Default_Lexer $Single_Char $Always_Fail/; # The following patterns were generated with examples/naive. $Default_Lexer = qr/(?![[(\\]).(?:[*+?]\??|\{\d+(?:,\d*)?\}\??)?|\\(?:[bABCEGLQUXZ]|[lu].|(?:[^\w]|[aefnrtdDwWsS]|c.|0\d{2}|x(?:[\da-fA-F]{2}|{[\da-fA-F]{4}})|N\{\w+\}|[Pp](?:\{\w+\}|.))(?:[*+?]\??|\{\d+(?:,\d*)?\}\??)?)|\[.*?(?<!\\)\](?:[*+?]\??|\{\d+(?:,\d*)?\}\??)?|\(.*?(?<!\\)\)(?:[*+?]\??|\{\d+(?:,\d*)?\}\??)?/; # ]) restore equilibrium $Single_Char = qr/^(?:\\(?:[aefnrtdDwWsS]|c.|[^\w\/{|}-]|0\d{2}|x(?:[\da-fA-F]{2}|{[\da-fA-F]{4}}))|[^\$^])$/; # The pattern to return when nothing has been added (and thus not match anything) $Always_Fail = "^\\b\0"; our $VERSION = '0.38'; # ------------------------------------------------ sub new { my $class = shift; my %args = @_; my $anc; for $anc (qw(word line string)) { if (exists $args{"anchor_$anc"}) { my $val = delete $args{"anchor_$anc"}; for my $anchor ("anchor_${anc}_begin", "anchor_${anc}_end") { $args{$anchor} = $val unless exists $args{$anchor}; } } } # anchor_string_absolute sets anchor_string_begin and anchor_string_end_absolute if (exists $args{anchor_string_absolute}) { my $val = delete $args{anchor_string_absolute}; for my $anchor (qw(anchor_string_begin anchor_string_end_absolute)) { $args{$anchor} = $val unless exists $args{$anchor}; } } exists $args{$_} or $args{$_} = 0 for qw( anchor_word_begin anchor_word_end anchor_line_begin anchor_line_end anchor_string_begin anchor_string_end anchor_string_end_absolute debug dup_warn indent lookahead mutable track unroll_plus ); exists $args{$_} or $args{$_} = 1 for qw( fold_meta_pairs reduce chomp ); @args{qw(re str path)} = (undef, undef, []); $args{flags} ||= delete $args{modifiers} || ''; $args{lex} = $Current_Lexer if defined $Current_Lexer; my $self = bless \%args, $class; if ($self->_debug(DEBUG_TIME)) { $self->_init_time_func(); $self->{_begin_time} = $self->{_time_func}->(); } $self->{input_record_separator} = delete $self->{rs} if exists $self->{rs}; exists $self->{file} and $self->add_file($self->{file}); return $self; } sub _init_time_func { my $self = shift; return if exists $self->{_time_func}; # attempt to improve accuracy if (!defined($self->{_use_time_hires})) { eval {require Time::HiRes}; $self->{_use_time_hires} = $@; } $self->{_time_func} = length($self->{_use_time_hires}) > 0 ? sub { time } : \&Time::HiRes::time ; } sub clone { my $self = shift; my $clone; my @attr = grep {$_ ne 'path'} keys %$self; @{$clone}{@attr} = @{$self}{@attr}; $clone->{path} = _path_clone($self->_path); bless $clone, ref($self); } sub _fastlex { my $self = shift; my $record = shift; my $len = 0; my @path = (); my $case = ''; my $qm = ''; my $debug = $self->{debug} & DEBUG_LEX; my $unroll_plus = $self->{unroll_plus}; my $token; my $qualifier; $debug and print "# _lex <$record>\n"; my $modifier = q{(?:[*+?]\\??|\\{(?:\\d+(?:,\d*)?|,\d+)\\}\\??)?}; my $class_matcher = qr/\[(?:\[:[a-z]+:\]|\\?.)*?\]/; my $paren_matcher = qr/\(.*?(?<!\\)\)$modifier/; my $misc_matcher = qr/(?:(c)(.)|(0)(\d{2}))($modifier)/; my $regular_matcher = qr/([^\\[(])($modifier)/; my $qm_matcher = qr/(\\?.)/; my $matcher = $regular_matcher; { if ($record =~ /\G$matcher/gc) { # neither a \\ nor [ nor ( followed by a modifer if ($1 eq '\\E') { $debug and print "# E\n"; $case = $qm = ''; $matcher = $regular_matcher; redo; } elsif ($qm and ($1 eq '\\L' or $1 eq '\\U')) { $debug and print "# ignore \\L, \\U\n"; redo; } $token = $1; $qualifier = defined $2 ? $2 : ''; $debug and print "# token <$token> <$qualifier>\n"; if ($qm) { $token = quotemeta($token); $token =~ s/^\\([^\w$()*+.?@\[\\\]^|{}\/])$/$1/; } else { $token =~ s{\A([][{}*+?@\\/])\Z}{\\$1}; } if ($unroll_plus and $qualifier =~ s/\A\+(\?)?\Z/*/) { $1 and $qualifier .= $1; $debug and print " unroll <$token><$token><$qualifier>\n"; $case and $token = $case eq 'L' ? lc($token) : uc($token); push @path, $token, "$token$qualifier"; } else { $debug and print " clean <$token>\n"; push @path, $case eq 'L' ? lc($token).$qualifier : $case eq 'U' ? uc($token).$qualifier : $token.$qualifier ; } redo; } elsif ($record =~ /\G\\/gc) { $debug and print "# backslash\n"; # backslash if ($record =~ /\G([sdwSDW])($modifier)/gc) { ($token, $qualifier) = ($1, $2); $debug and print "# meta <$token> <$qualifier>\n"; push @path, ($unroll_plus and $qualifier =~ s/\A\+(\?)?\Z/*/) ? ("\\$token", "\\$token$qualifier" . (defined $1 ? $1 : '')) : "\\$token$qualifier"; } elsif ($record =~ /\Gx([\da-fA-F]{2})($modifier)/gc) { $debug and print "# x $1\n"; $token = quotemeta(chr(hex($1))); $qualifier = $2; $debug and print "# cooked <$token>\n"; $token =~ s/^\\([^\w$()*+.?\[\\\]^|{\/])$/$1/; # } balance $debug and print "# giving <$token>\n"; push @path, ($unroll_plus and $qualifier =~ s/\A\+(\?)?\Z/*/) ? ($token, "$token$qualifier" . (defined $1 ? $1 : '')) : "$token$qualifier"; } elsif ($record =~ /\GQ/gc) { $debug and print "# Q\n"; $qm = 1; $matcher = $qm_matcher; } elsif ($record =~ /\G([LU])/gc) { $debug and print "# case $1\n"; $case = $1; } elsif ($record =~ /\GE/gc) { $debug and print "# E\n"; $case = $qm = ''; $matcher = $regular_matcher; } elsif ($record =~ /\G([lu])(.)/gc) { $debug and print "# case $1 to <$2>\n"; push @path, $1 eq 'l' ? lc($2) : uc($2); } elsif (my @arg = grep {defined} $record =~ /\G$misc_matcher/gc) { if ($] < 5.007) { my $len = 0; $len += length($_) for @arg; $debug and print "# pos ", pos($record), " fixup add $len\n"; pos($record) = pos($record) + $len; } my $directive = shift @arg; if ($directive eq 'c') { $debug and print "# ctrl <@arg>\n"; push @path, "\\c" . uc(shift @arg); } else { # elsif ($directive eq '0') { $debug and print "# octal <@arg>\n"; my $ascii = oct(shift @arg); push @path, ($ascii < 32) ? "\\c" . chr($ascii+64) : chr($ascii) ; } $path[-1] .= join( '', @arg ); # if @arg; redo; } elsif ($record =~ /\G(.)/gc) { $token = $1; $token =~ s{[AZabefnrtz\[\]{}()\\\$*+.?@|/^]}{\\$token}; $debug and print "# meta <$token>\n"; push @path, $token; } else { $debug and print "# ignore char at ", pos($record), " of <$record>\n"; } redo; } elsif ($record =~ /\G($class_matcher)($modifier)/gc) { # [class] followed by a modifer my $class = $1; my $qualifier = defined $2 ? $2 : ''; $debug and print "# class begin <$class> <$qualifier>\n"; if ($class =~ /\A\[\\?(.)]\Z/) { $class = quotemeta $1; $class =~ s{\A\\([!@%])\Z}{$1}; $debug and print "# class unwrap $class\n"; } $debug and print "# class end <$class> <$qualifier>\n"; push @path, ($unroll_plus and $qualifier =~ s/\A\+(\?)?\Z/*/) ? ($class, "$class$qualifier" . (defined $1 ? $1 : '')) : "$class$qualifier"; redo; } elsif ($record =~ /\G($paren_matcher)/gc) { $debug and print "# paren <$1>\n"; # (paren) followed by a modifer push @path, $1; redo; } } return \@path; } sub _lex { my $self = shift; my $record = shift; my $len = 0; my @path = (); my $case = ''; my $qm = ''; my $re = defined $self->{lex} ? $self->{lex} : defined $Current_Lexer ? $Current_Lexer : $Default_Lexer; my $debug = $self->{debug} & DEBUG_LEX; $debug and print "# _lex <$record>\n"; my ($token, $next_token, $diff, $token_len); while( $record =~ /($re)/g ) { $token = $1; $token_len = length($token); $debug and print "# lexed <$token> len=$token_len\n"; if( pos($record) - $len > $token_len ) { $next_token = $token; $token = substr( $record, $len, $diff = pos($record) - $len - $token_len ); $debug and print "# recover <", substr( $record, $len, $diff ), "> as <$token>, save <$next_token>\n"; $len += $diff; } $len += $token_len; TOKEN: { if( substr( $token, 0, 1 ) eq '\\' ) { if( $token =~ /^\\([ELQU])$/ ) { if( $1 eq 'E' ) { $qm and $re = defined $self->{lex} ? $self->{lex} : defined $Current_Lexer ? $Current_Lexer : $Default_Lexer; $case = $qm = ''; } elsif( $1 eq 'Q' ) { $qm = $1; # switch to a more precise lexer to quotemeta individual characters $re = qr/\\?./; } else { $case = $1; } $debug and print "# state change qm=<$qm> case=<$case>\n"; goto NEXT_TOKEN; } elsif( $token =~ /^\\([lu])(.)$/ ) { $debug and print "# apply case=<$1> to <$2>\n"; push @path, $1 eq 'l' ? lc($2) : uc($2); goto NEXT_TOKEN; } elsif( $token =~ /^\\x([\da-fA-F]{2})$/ ) { $token = quotemeta(chr(hex($1))); $debug and print "# cooked <$token>\n"; $token =~ s/^\\([^\w$()*+.?@\[\\\]^|{\/])$/$1/; # } balance $debug and print "# giving <$token>\n"; } else { $token =~ s/^\\([^\w$()*+.?@\[\\\]^|{\/])$/$1/; # } balance $debug and print "# backslashed <$token>\n"; } } else { $case and $token = $case eq 'U' ? uc($token) : lc($token); $qm and $token = quotemeta($token); $token = '\\/' if $token eq '/'; } # undo quotemeta's brute-force escapades $qm and $token =~ s/^\\([^\w$()*+.?@\[\\\]^|{}\/])$/$1/; $debug and print "# <$token> case=<$case> qm=<$qm>\n"; push @path, $token; NEXT_TOKEN: if( defined $next_token ) { $debug and print "# redo <$next_token>\n"; $token = $next_token; $next_token = undef; redo TOKEN; } } } if( $len < length($record) ) { # NB: the remainder only arises in the case of degenerate lexer, # and if \Q is operative, the lexer will have been switched to # /\\?./, which means there can never be a remainder, so we # don't have to bother about quotemeta. In other words: # $qm will never be true in this block. my $remain = substr($record,$len); $case and $remain = $case eq 'U' ? uc($remain) : lc($remain); $debug and print "# add remaining <$remain> case=<$case> qm=<$qm>\n"; push @path, $remain; } $debug and print "# _lex out <@path>\n"; return \@path; } sub add { my $self = shift; my $record; my $debug = $self->{debug} & DEBUG_LEX; while( defined( $record = shift @_ )) { CORE::chomp($record) if $self->{chomp}; next if $self->{pre_filter} and not $self->{pre_filter}->($record); $debug and print "# add <$record>\n"; $self->{stats_raw} += length $record; my $list = $record =~ /[+*?(\\\[{]/ # }]) restore equilibrium ? $self->{lex} ? $self->_lex($record) : $self->_fastlex($record) : [split //, $record] ; next if $self->{filter} and not $self->{filter}->(@$list); $self->_insertr( $list ); } return $self; } sub add_file { my $self = shift; my $rs; my @file; if (ref($_[0]) eq 'HASH') { my $arg = shift; $rs = $arg->{rs} || $arg->{input_record_separator} || $self->{input_record_separator} || $/; @file = ref($arg->{file}) eq 'ARRAY' ? @{$arg->{file}} : $arg->{file}; } else { $rs = $self->{input_record_separator} || $/; @file = @_; } local $/ = $rs; my $file; for $file (@file) { open my $fh, '<', $file or do { require Carp; Carp::croak("cannot open $file for input: $!"); }; while (defined (my $rec = <$fh>)) { $self->add($rec); } close $fh; } return $self; } sub insert { my $self = shift; return if $self->{filter} and not $self->{filter}->(@_); $self->_insertr( [@_] ); return $self; } sub _insertr { my $self = shift; my $dup = $self->{stats_dup} || 0; $self->{path} = $self->_insert_path( $self->_path, $self->_debug(DEBUG_ADD), $_[0] ); if( not defined $self->{stats_dup} or $dup == $self->{stats_dup} ) { ++$self->{stats_add}; $self->{stats_cooked} += defined($_) ? length($_) : 0 for @{$_[0]}; } elsif( $self->{dup_warn} ) { if( ref $self->{dup_warn} eq 'CODE' ) { $self->{dup_warn}->($self, $_[0]); } else { my $pattern = join( '', @{$_[0]} ); require Carp; Carp::carp("duplicate pattern added: /$pattern/"); } } $self->{str} = $self->{re} = undef; } sub lexstr { return shift->_lex(shift); } sub pre_filter { my $self = shift; my $pre_filter = shift; if( defined $pre_filter and ref($pre_filter) ne 'CODE' ) { require Carp; Carp::croak("pre_filter method not passed a coderef"); } $self->{pre_filter} = $pre_filter; return $self; } sub filter { my $self = shift; my $filter = shift; if( defined $filter and ref($filter) ne 'CODE' ) { require Carp; Carp::croak("filter method not passed a coderef"); } $self->{filter} = $filter; return $self; } sub as_string { my $self = shift; if( not defined $self->{str} ) { if( $self->{track} ) { $self->{m} = undef; $self->{mcount} = 0; $self->{mlist} = []; $self->{str} = _re_path_track($self, $self->_path, '', ''); } else { $self->_reduce unless ($self->{mutable} or not $self->{reduce}); my $arg = {@_}; $arg->{indent} = $self->{indent} if not exists $arg->{indent} and $self->{indent} > 0; if( exists $arg->{indent} and $arg->{indent} > 0 ) { $arg->{depth} = 0; $self->{str} = _re_path_pretty($self, $self->_path, $arg); } elsif( $self->{lookahead} ) { $self->{str} = _re_path_lookahead($self, $self->_path); } else { $self->{str} = _re_path($self, $self->_path); } } if (not length $self->{str}) { # explicitly fail to match anything if no pattern was generated $self->{str} = $Always_Fail; } else { my $begin = $self->{anchor_word_begin} ? '\\b' : $self->{anchor_line_begin} ? '^' : $self->{anchor_string_begin} ? '\A' : '' ; my $end = $self->{anchor_word_end} ? '\\b' : $self->{anchor_line_end} ? '$' : $self->{anchor_string_end} ? '\Z' : $self->{anchor_string_end_absolute} ? '\z' : '' ; $self->{str} = "$begin$self->{str}$end"; } $self->{path} = [] unless $self->{mutable}; } return $self->{str}; } sub re { my $self = shift; $self->_build_re($self->as_string(@_)) unless defined $self->{re}; return $self->{re}; } use overload '""' => sub { my $self = shift; return $self->{re} if $self->{re}; $self->_build_re($self->as_string()); return $self->{re}; }; sub _build_re { my $self = shift; my $str = shift; if( $self->{track} ) { use re 'eval'; $self->{re} = length $self->{flags} ? qr/(?$self->{flags}:$str)/ : qr/$str/ ; } else { # how could I not repeat myself? $self->{re} = length $self->{flags} ? qr/(?$self->{flags}:$str)/ : qr/$str/ ; } } sub match { my $self = shift; my $target = shift; $self->_build_re($self->as_string(@_)) unless defined $self->{re}; $self->{m} = undef; $self->{mvar} = []; if( not $target =~ /$self->{re}/ ) { $self->{mbegin} = []; $self->{mend} = []; return undef; } $self->{m} = $^R if $] >= 5.009005; $self->{mbegin} = _path_copy([@-]); $self->{mend} = _path_copy([@+]); my $n = 0; for( my $n = 0; $n < @-; ++$n ) { push @{$self->{mvar}}, substr($target, $-[$n], $+[$n] - $-[$n]) if defined $-[$n] and defined $+[$n]; } if( $self->{track} ) { return defined $self->{m} ? $self->{mlist}[$self->{m}] : 1; } else { return 1; } } sub source { my $self = shift; return unless $self->{track}; defined($_[0]) and return $self->{mlist}[$_[0]]; return unless defined $self->{m}; return $self->{mlist}[$self->{m}]; } sub mbegin { my $self = shift; return exists $self->{mbegin} ? $self->{mbegin} : []; } sub mend { my $self = shift; return exists $self->{mend} ? $self->{mend} : []; } sub mvar { my $self = shift; return undef unless exists $self->{mvar}; return defined($_[0]) ? $self->{mvar}[$_[0]] : $self->{mvar}; } sub capture { my $self = shift; if( $self->{mvar} ) { my @capture = @{$self->{mvar}}; shift @capture; return @capture; } return (); } sub matched { my $self = shift; return defined $self->{m} ? $self->{mlist}[$self->{m}] : undef; } sub stats_add { my $self = shift; return $self->{stats_add} || 0; } sub stats_dup { my $self = shift; return $self->{stats_dup} || 0; } sub stats_raw { my $self = shift; return $self->{stats_raw} || 0; } sub stats_cooked { my $self = shift; return $self->{stats_cooked} || 0; } sub stats_length { my $self = shift; return (defined $self->{str} and $self->{str} ne $Always_Fail) ? length $self->{str} : 0; } sub dup_warn { my $self = shift; $self->{dup_warn} = defined($_[0]) ? $_[0] : 1; return $self; } sub anchor_word_begin { my $self = shift; $self->{anchor_word_begin} = defined($_[0]) ? $_[0] : 1; return $self; } sub anchor_word_end { my $self = shift; $self->{anchor_word_end} = defined($_[0]) ? $_[0] : 1; return $self; } sub anchor_word { my $self = shift; my $state = shift; $self->anchor_word_begin($state)->anchor_word_end($state); return $self; } sub anchor_line_begin { my $self = shift; $self->{anchor_line_begin} = defined($_[0]) ? $_[0] : 1; return $self; } sub anchor_line_end { my $self = shift; $self->{anchor_line_end} = defined($_[0]) ? $_[0] : 1; return $self; } sub anchor_line { my $self = shift; my $state = shift; $self->anchor_line_begin($state)->anchor_line_end($state); return $self; } sub anchor_string_begin { my $self = shift; $self->{anchor_string_begin} = defined($_[0]) ? $_[0] : 1; return $self; } sub anchor_string_end { my $self = shift; $self->{anchor_string_end} = defined($_[0]) ? $_[0] : 1; return $self; } sub anchor_string_end_absolute { my $self = shift; $self->{anchor_string_end_absolute} = defined($_[0]) ? $_[0] : 1; return $self; } sub anchor_string { my $self = shift; my $state = defined($_[0]) ? $_[0] : 1; $self->anchor_string_begin($state)->anchor_string_end($state); return $self; } sub anchor_string_absolute { my $self = shift; my $state = defined($_[0]) ? $_[0] : 1; $self->anchor_string_begin($state)->anchor_string_end_absolute($state); return $self; } sub debug { my $self = shift; $self->{debug} = defined($_[0]) ? $_[0] : 0; if ($self->_debug(DEBUG_TIME)) { # hmm, debugging time was switched on after instantiation $self->_init_time_func; $self->{_begin_time} = $self->{_time_func}->(); } return $self; } sub dump { return _dump($_[0]->_path); } sub chomp { my $self = shift; $self->{chomp} = defined($_[0]) ? $_[0] : 1; return $self; } sub fold_meta_pairs { my $self = shift; $self->{fold_meta_pairs} = defined($_[0]) ? $_[0] : 1; return $self; } sub indent { my $self = shift; $self->{indent} = defined($_[0]) ? $_[0] : 0; return $self; } sub lookahead { my $self = shift; $self->{lookahead} = defined($_[0]) ? $_[0] : 1; return $self; } sub flags { my $self = shift; $self->{flags} = defined($_[0]) ? $_[0] : ''; return $self; } sub modifiers { my $self = shift; return $self->flags(@_); } sub track { my $self = shift; $self->{track} = defined($_[0]) ? $_[0] : 1; return $self; } sub unroll_plus { my $self = shift; $self->{unroll_plus} = defined($_[0]) ? $_[0] : 1; return $self; } sub lex { my $self = shift; $self->{lex} = qr($_[0]); return $self; } sub reduce { my $self = shift; $self->{reduce} = defined($_[0]) ? $_[0] : 1; return $self; } sub mutable { my $self = shift; $self->{mutable} = defined($_[0]) ? $_[0] : 1; return $self; } sub reset { # reinitialise the internal state of the object my $self = shift; $self->{path} = []; $self->{re} = undef; $self->{str} = undef; return $self; } sub Default_Lexer { if( $_[0] ) { if( my $refname = ref($_[0]) ) { require Carp; Carp::croak("Cannot pass a $refname to Default_Lexer"); } $Current_Lexer = $_[0]; } return defined $Current_Lexer ? $Current_Lexer : $Default_Lexer; } # --- no user serviceable parts below --- # -- debug helpers sub _debug { my $self = shift; return $self->{debug} & shift() ? 1 : 0; } # -- helpers sub _path { # access the path return $_[0]->{path}; } # -- the heart of the matter $have_Storable = do { eval { require Storable; import Storable 'dclone'; }; $@ ? 0 : 1; }; sub _path_clone { $have_Storable ? dclone($_[0]) : _path_copy($_[0]); } sub _path_copy { my $path = shift; my $new = []; for( my $p = 0; $p < @$path; ++$p ) { if( ref($path->[$p]) eq 'HASH' ) { push @$new, _node_copy($path->[$p]); } elsif( ref($path->[$p]) eq 'ARRAY' ) { push @$new, _path_copy($path->[$p]); } else { push @$new, $path->[$p]; } } return $new; } sub _node_copy { my $node = shift; my $new = {}; while( my( $k, $v ) = each %$node ) { $new->{$k} = defined($v) ? _path_copy($v) : undef ; } return $new; } sub _insert_path { my $self = shift; my $list = shift; my $debug = shift; my @in = @{shift()}; # create a new copy if( @$list == 0 ) { # special case the first time if( @in == 0 or (@in == 1 and (not defined $in[0] or $in[0] eq ''))) { return [{'' => undef}]; } else { return \@in; } } $debug and print "# _insert_path @{[_dump(\@in)]} into @{[_dump($list)]}\n"; my $path = $list; my $offset = 0; my $token; if( not @in ) { if( ref($list->[0]) ne 'HASH' ) { return [ { '' => undef, $list->[0] => $list } ]; } else { $list->[0]{''} = undef; return $list; } } while( defined( $token = shift @in )) { if( ref($token) eq 'HASH' ) { $debug and print "# p0=", _dump($path), "\n"; $path = $self->_insert_node( $path, $offset, $token, $debug, @in ); $debug and print "# p1=", _dump($path), "\n"; last; } if( ref($path->[$offset]) eq 'HASH' ) { $debug and print "# at (off=$offset len=@{[scalar @$path]}) ", _dump($path->[$offset]), "\n"; my $node = $path->[$offset]; if( exists( $node->{$token} )) { if ($offset < $#$path) { my $new = { $token => [$token, @in], _re_path($self, [$node]) => [@{$path}[$offset..$#$path]], }; splice @$path, $offset, @$path-$offset, $new; last; } else { $debug and print "# descend key=$token @{[_dump($node->{$token})]}\n"; $path = $node->{$token}; $offset = 0; redo; } } else { $debug and print "# add path ($token:@{[_dump(\@in)]}) into @{[_dump($path)]} at off=$offset to end=@{[scalar $#$path]}\n"; if( $offset == $#$path ) { $node->{$token} = [ $token, @in ]; } else { my $new = { _node_key($token) => [ $token, @in ], _node_key($node) => [@{$path}[$offset..$#{$path}]], }; splice( @$path, $offset, @$path - $offset, $new ); $debug and print "# fused node=@{[_dump($new)]} path=@{[_dump($path)]}\n"; } last; } } if( $debug ) { my $msg = ''; my $n; for( $n = 0; $n < @$path; ++$n ) { $msg .= ' ' if $n; my $atom = ref($path->[$n]) eq 'HASH' ? '{'.join( ' ', keys(%{$path->[$n]})).'}' : $path->[$n] ; $msg .= $n == $offset ? "<$atom>" : $atom; } print "# at path ($msg)\n"; } if( $offset >= @$path ) { push @$path, { $token => [ $token, @in ], '' => undef }; $debug and print "# added remaining @{[_dump($path)]}\n"; last; } elsif( $token ne $path->[$offset] ) { $debug and print "# token $token not present\n"; splice @$path, $offset, @$path-$offset, { length $token ? ( _node_key($token) => [$token, @in]) : ( '' => undef ) , $path->[$offset] => [@{$path}[$offset..$#{$path}]], }; $debug and print "# path=@{[_dump($path)]}\n"; last; } elsif( not @in ) { $debug and print "# last token to add\n"; if( defined( $path->[$offset+1] )) { ++$offset; if( ref($path->[$offset]) eq 'HASH' ) { $debug and print "# add sentinel to node\n"; $path->[$offset]{''} = undef; } else { $debug and print "# convert <$path->[$offset]> to node for sentinel\n"; splice @$path, $offset, @$path-$offset, { '' => undef, $path->[$offset] => [ @{$path}[$offset..$#{$path}] ], }; } } else { # already seen this pattern ++$self->{stats_dup}; } last; } # if we get here then @_ still contains a token ++$offset; } $list; } sub _insert_node { my $self = shift; my $path = shift; my $offset = shift; my $token = shift; my $debug = shift; my $path_end = [@{$path}[$offset..$#{$path}]]; # NB: $path->[$offset] and $[path_end->[0] are equivalent my $token_key = _re_path($self, [$token]); $debug and print "# insert node(@{[_dump($token)]}:@{[_dump(\@_)]}) (key=$token_key)", " at path=@{[_dump($path_end)]}\n"; if( ref($path_end->[0]) eq 'HASH' ) { if( exists($path_end->[0]{$token_key}) ) { if( @$path_end > 1 ) { my $path_key = _re_path($self, [$path_end->[0]]); my $new = { $path_key => [ @$path_end ], $token_key => [ $token, @_ ], }; $debug and print "# +bifurcate new=@{[_dump($new)]}\n"; splice( @$path, $offset, @$path_end, $new ); } else { my $old_path = $path_end->[0]{$token_key}; my $new_path = []; while( @$old_path and _node_eq( $old_path->[0], $token )) { $debug and print "# identical nodes in sub_path ", ref($token) ? _dump($token) : $token, "\n"; push @$new_path, shift(@$old_path); $token = shift @_; } if( @$new_path ) { my $new; my $token_key = $token; if( @_ ) { $new = { _re_path($self, $old_path) => $old_path, $token_key => [$token, @_], }; $debug and print "# insert_node(bifurc) n=@{[_dump([$new])]}\n"; } else { $debug and print "# insert $token into old path @{[_dump($old_path)]}\n"; if( @$old_path ) { $new = ($self->_insert_path( $old_path, $debug, [$token] ))->[0]; } else { $new = { '' => undef, $token => [$token] }; } } push @$new_path, $new; } $path_end->[0]{$token_key} = $new_path; $debug and print "# +_insert_node result=@{[_dump($path_end)]}\n"; splice( @$path, $offset, @$path_end, @$path_end ); } } elsif( not _node_eq( $path_end->[0], $token )) { if( @$path_end > 1 ) { my $path_key = _re_path($self, [$path_end->[0]]); my $new = { $path_key => [ @$path_end ], $token_key => [ $token, @_ ], }; $debug and print "# path->node1 at $path_key/$token_key @{[_dump($new)]}\n"; splice( @$path, $offset, @$path_end, $new ); } else { $debug and print "# next in path is node, trivial insert at $token_key\n"; $path_end->[0]{$token_key} = [$token, @_]; splice( @$path, $offset, @$path_end, @$path_end ); } } else { while( @$path_end and _node_eq( $path_end->[0], $token )) { $debug and print "# identical nodes @{[_dump([$token])]}\n"; shift @$path_end; $token = shift @_; ++$offset; } if( @$path_end ) { $debug and print "# insert at $offset $token:@{[_dump(\@_)]} into @{[_dump($path_end)]}\n"; $path_end = $self->_insert_path( $path_end, $debug, [$token, @_] ); $debug and print "# got off=$offset s=@{[scalar @_]} path_add=@{[_dump($path_end)]}\n"; splice( @$path, $offset, @$path - $offset, @$path_end ); $debug and print "# got final=@{[_dump($path)]}\n"; } else { $token_key = _node_key($token); my $new = { '' => undef, $token_key => [ $token, @_ ], }; $debug and print "# convert opt @{[_dump($new)]}\n"; push @$path, $new; } } } else { if( @$path_end ) { my $new = { $path_end->[0] => [ @$path_end ], $token_key => [ $token, @_ ], }; $debug and print "# atom->node @{[_dump($new)]}\n"; splice( @$path, $offset, @$path_end, $new ); $debug and print "# out=@{[_dump($path)]}\n"; } else { $debug and print "# add opt @{[_dump([$token,@_])]} via $token_key\n"; push @$path, { '' => undef, $token_key => [ $token, @_ ], }; } } $path; } sub _reduce { my $self = shift; my $context = { debug => $self->_debug(DEBUG_TAIL), depth => 0 }; if ($self->_debug(DEBUG_TIME)) { $self->_init_time_func; my $now = $self->{_time_func}->(); if (exists $self->{_begin_time}) { printf "# load=%0.6f\n", $now - $self->{_begin_time}; } else { printf "# load-epoch=%0.6f\n", $now; } $self->{_begin_time} = $self->{_time_func}->(); } my ($head, $tail) = _reduce_path( $self->_path, $context ); $context->{debug} and print "# final head=", _dump($head), ' tail=', _dump($tail), "\n"; if( !@$head ) { $self->{path} = $tail; } else { $self->{path} = [ @{_unrev_path( $tail, $context )}, @{_unrev_path( $head, $context )}, ]; } if ($self->_debug(DEBUG_TIME)) { my $now = $self->{_time_func}->(); if (exists $self->{_begin_time}) { printf "# reduce=%0.6f\n", $now - $self->{_begin_time}; } else { printf "# reduce-epoch=%0.6f\n", $now; } $self->{_begin_time} = $self->{_time_func}->(); } $context->{debug} and print "# final path=", _dump($self->{path}), "\n"; return $self; } sub _remove_optional { if( exists $_[0]->{''} ) { delete $_[0]->{''}; return 1; } return 0; } sub _reduce_path { my ($path, $ctx) = @_; my $indent = ' ' x $ctx->{depth}; my $debug = $ctx->{debug}; $debug and print "#$indent _reduce_path $ctx->{depth} ", _dump($path), "\n"; my $new; my $head = []; my $tail = []; while( defined( my $p = pop @$path )) { if( ref($p) eq 'HASH' ) { my ($node_head, $node_tail) = _reduce_node($p, _descend($ctx) ); $debug and print "#$indent| head=", _dump($node_head), " tail=", _dump($node_tail), "\n"; push @$head, @$node_head if scalar @$node_head; push @$tail, ref($node_tail) eq 'HASH' ? $node_tail : @$node_tail; } else { if( @$head ) { $debug and print "#$indent| push $p leaves @{[_dump($path)]}\n"; push @$tail, $p; } else { $debug and print "#$indent| unshift $p\n"; unshift @$tail, $p; } } } $debug and print "#$indent| tail nr=@{[scalar @$tail]} t0=", ref($tail->[0]), (ref($tail->[0]) eq 'HASH' ? " n=" . scalar(keys %{$tail->[0]}) : '' ), "\n"; if( @$tail > 1 and ref($tail->[0]) eq 'HASH' and keys %{$tail->[0]} == 2 ) { my $opt; my $fixed; while( my ($key, $path) = each %{$tail->[0]} ) { $debug and print "#$indent| scan k=$key p=@{[_dump($path)]}\n"; next unless $path; if (@$path == 1 and ref($path->[0]) eq 'HASH') { $opt = $path->[0]; } else { $fixed = $path; } } if( exists $tail->[0]{''} ) { my $path = [@{$tail}[1..$#{$tail}]]; $tail = $tail->[0]; ($head, $tail, $path) = _slide_tail( $head, $tail, $path, _descend($ctx) ); $tail = [$tail, @$path]; } } $debug and print "#$indent _reduce_path $ctx->{depth} out head=", _dump($head), ' tail=', _dump($tail), "\n"; return ($head, $tail); } sub _reduce_node { my ($node, $ctx) = @_; my $indent = ' ' x $ctx->{depth}; my $debug = $ctx->{debug}; my $optional = _remove_optional($node); $debug and print "#$indent _reduce_node $ctx->{depth} in @{[_dump($node)]} opt=$optional\n"; if( $optional and scalar keys %$node == 1 ) { my $path = (values %$node)[0]; if( not grep { ref($_) eq 'HASH' } @$path ) { # if we have removed an optional, and there is only one path # left then there is nothing left to compare. Because of the # optional it cannot participate in any further reductions. # (unless we test for equality among sub-trees). my $result = { '' => undef, $path->[0] => $path }; $debug and print "#$indent| fast fail @{[_dump($result)]}\n"; return [], $result; } } my( $fail, $reduce ) = _scan_node( $node, _descend($ctx) ); $debug and print "#$indent|_scan_node done opt=$optional reduce=@{[_dump($reduce)]} fail=@{[_dump($fail)]}\n"; # We now perform tail reduction on each of the nodes in the reduce # hash. If we have only one key, we know we will have a successful # reduction (since everything that was inserted into the node based # on the value of the last token of each path all mapped to the same # value). if( @$fail == 0 and keys %$reduce == 1 and not $optional) { # every path shares a common path my $path = (values %$reduce)[0]; my ($common, $tail) = _do_reduce( $path, _descend($ctx) ); $debug and print "#$indent|_reduce_node $ctx->{depth} common=@{[_dump($common)]} tail=", _dump($tail), "\n"; return( $common, $tail ); } # this node resulted in a list of paths, game over $ctx->{indent} = $indent; return _reduce_fail( $reduce, $fail, $optional, _descend($ctx) ); } sub _reduce_fail { my( $reduce, $fail, $optional, $ctx ) = @_; my( $debug, $depth, $indent ) = @{$ctx}{qw(debug depth indent)}; my %result; $result{''} = undef if $optional; my $p; for $p (keys %$reduce) { my $path = $reduce->{$p}; if( scalar @$path == 1 ) { $path = $path->[0]; $debug and print "#$indent| -simple opt=$optional unrev @{[_dump($path)]}\n"; $path = _unrev_path($path, _descend($ctx) ); $result{_node_key($path->[0])} = $path; } else { $debug and print "#$indent| _do_reduce(@{[_dump($path)]})\n"; my ($common, $tail) = _do_reduce( $path, _descend($ctx) ); $path = [ ( ref($tail) eq 'HASH' ? _unrev_node($tail, _descend($ctx) ) : _unrev_path($tail, _descend($ctx) ) ), @{_unrev_path($common, _descend($ctx) )} ]; $debug and print "#$indent| +reduced @{[_dump($path)]}\n"; $result{_node_key($path->[0])} = $path; } } my $f; for $f( @$fail ) { $debug and print "#$indent| +fail @{[_dump($f)]}\n"; $result{$f->[0]} = $f; } $debug and print "#$indent _reduce_fail $depth fail=@{[_dump(\%result)]}\n"; return ( [], \%result ); } sub _scan_node { my( $node, $ctx ) = @_; my $indent = ' ' x $ctx->{depth}; my $debug = $ctx->{debug}; # For all the paths in the node, reverse them. If the first token # of the path is a scalar, push it onto an array in a hash keyed by # the value of the scalar. # # If it is a node, call _reduce_node on this node beforehand. If we # get back a common head, all of the paths in the subnode shared a # common tail. We then store the common part and the remaining node # of paths (which is where the paths diverged from the end and install # this into the same hash. At this point both the common and the tail # are in reverse order, just as simple scalar paths are. # # On the other hand, if there were no common path returned then all # the paths of the sub-node diverge at the end character. In this # case the tail cannot participate in any further reductions and will # appear in forward order. # # certainly the hurgliest function in the whole file :( # $debug = 1 if $depth >= 8; my @fail; my %reduce; my $n; for $n( map { substr($_, index($_, '#')+1) } sort map { join( '|' => scalar(grep {ref($_) eq 'HASH'} @{$node->{$_}}), _node_offset($node->{$_}), scalar @{$node->{$_}}, ) . "#$_" } keys %$node ) { my( $end, @path ) = reverse @{$node->{$n}}; if( ref($end) ne 'HASH' ) { $debug and print "# $indent|_scan_node push reduce ($end:@{[_dump(\@path)]})\n"; push @{$reduce{$end}}, [ $end, @path ]; } else { $debug and print "# $indent|_scan_node head=", _dump(\@path), ' tail=', _dump($end), "\n"; my $new_path; # deal with sing, singing => s(?:ing)?ing if( keys %$end == 2 and exists $end->{''} ) { my ($key, $opt_path) = each %$end; ($key, $opt_path) = each %$end if $key eq ''; $opt_path = [reverse @{$opt_path}]; $debug and print "# $indent| check=", _dump($opt_path), "\n"; my $end = { '' => undef, $opt_path->[0] => [@$opt_path] }; my $head = []; my $path = [@path]; ($head, my $slide, $path) = _slide_tail( $head, $end, $path, $ctx ); if( @$head ) { $new_path = [ @$head, $slide, @$path ]; } } if( $new_path ) { $debug and print "# $indent|_scan_node slid=", _dump($new_path), "\n"; push @{$reduce{$new_path->[0]}}, $new_path; } else { my( $common, $tail ) = _reduce_node( $end, _descend($ctx) ); if( not @$common ) { $debug and print "# $indent| +failed $n\n"; push @fail, [reverse(@path), $tail]; } else { my $path = [@path]; $debug and print "# $indent|_scan_node ++recovered common=@{[_dump($common)]} tail=", _dump($tail), " path=@{[_dump($path)]}\n"; if( ref($tail) eq 'HASH' and keys %$tail == 2 ) { if( exists $tail->{''} ) { ($common, $tail, $path) = _slide_tail( $common, $tail, $path, $ctx ); } } push @{$reduce{$common->[0]}}, [ @$common, (ref($tail) eq 'HASH' ? $tail : @$tail ), @$path ]; } } } } $debug and print "# $indent|_scan_node counts: reduce=@{[scalar keys %reduce]} fail=@{[scalar @fail]}\n"; return( \@fail, \%reduce ); } sub _do_reduce { my ($path, $ctx) = @_; my $indent = ' ' x $ctx->{depth}; my $debug = $ctx->{debug}; my $ra = Regexp::Assemble->new(chomp=>0); $ra->debug($debug); $debug and print "# $indent| do @{[_dump($path)]}\n"; $ra->_insertr( $_ ) for # When nodes come into the picture, we have to be careful # about how we insert the paths into the assembly. # Paths with nodes first, then closest node to front # then shortest path. Merely because if we can control # order in which paths containing nodes get inserted, # then we can make a couple of assumptions that simplify # the code in _insert_node. sort { scalar(grep {ref($_) eq 'HASH'} @$a) <=> scalar(grep {ref($_) eq 'HASH'} @$b) || _node_offset($b) <=> _node_offset($a) || scalar @$a <=> scalar @$b } @$path ; $path = $ra->_path; my $common = []; push @$common, shift @$path while( ref($path->[0]) ne 'HASH' ); my $tail = scalar( @$path ) > 1 ? [@$path] : $path->[0]; $debug and print "# $indent| _do_reduce common=@{[_dump($common)]} tail=@{[_dump($tail)]}\n"; return ($common, $tail); } sub _node_offset { # return the offset that the first node is found, or -ve # optimised for speed my $nr = @{$_[0]}; my $atom = -1; ref($_[0]->[$atom]) eq 'HASH' and return $atom while ++$atom < $nr; return -1; } sub _slide_tail { my $head = shift; my $tail = shift; my $path = shift; my $ctx = shift; my $indent = ' ' x $ctx->{depth}; my $debug = $ctx->{debug}; $debug and print "# $indent| slide in h=", _dump($head), ' t=', _dump($tail), ' p=', _dump($path), "\n"; my $slide_path = (each %$tail)[-1]; $slide_path = (each %$tail)[-1] unless defined $slide_path; $debug and print "# $indent| slide potential ", _dump($slide_path), " over ", _dump($path), "\n"; while( defined $path->[0] and $path->[0] eq $slide_path->[0] ) { $debug and print "# $indent| slide=tail=$slide_path->[0]\n"; my $slide = shift @$path; shift @$slide_path; push @$slide_path, $slide; push @$head, $slide; } $debug and print "# $indent| slide path ", _dump($slide_path), "\n"; my $slide_node = { '' => undef, _node_key($slide_path->[0]) => $slide_path, }; $debug and print "# $indent| slide out h=", _dump($head), ' s=', _dump($slide_node), ' p=', _dump($path), "\n"; return ($head, $slide_node, $path); } sub _unrev_path { my ($path, $ctx) = @_; my $indent = ' ' x $ctx->{depth}; my $debug = $ctx->{debug}; my $new; if( not grep { ref($_) } @$path ) { $debug and print "# ${indent}_unrev path fast ", _dump($path); $new = [reverse @$path]; $debug and print "# -> ", _dump($new), "\n"; return $new; } $debug and print "# ${indent}unrev path in ", _dump($path), "\n"; while( defined( my $p = pop @$path )) { push @$new, ref($p) eq 'HASH' ? _unrev_node($p, _descend($ctx) ) : ref($p) eq 'ARRAY' ? _unrev_path($p, _descend($ctx) ) : $p ; } $debug and print "# ${indent}unrev path out ", _dump($new), "\n"; return $new; } sub _unrev_node { my ($node, $ctx ) = @_; my $indent = ' ' x $ctx->{depth}; my $debug = $ctx->{debug}; my $optional = _remove_optional($node); $debug and print "# ${indent}unrev node in ", _dump($node), " opt=$optional\n"; my $new; $new->{''} = undef if $optional; my $n; for $n( keys %$node ) { my $path = _unrev_path($node->{$n}, _descend($ctx) ); $new->{_node_key($path->[0])} = $path; } $debug and print "# ${indent}unrev node out ", _dump($new), "\n"; return $new; } sub _node_key { my $node = shift; return _node_key($node->[0]) if ref($node) eq 'ARRAY'; return $node unless ref($node) eq 'HASH'; my $key = ''; my $k; for $k( keys %$node ) { next if $k eq ''; $key = $k if $key eq '' or $key gt $k; } return $key; } sub _descend { # Take a context object, and increase the depth by one. # By creating a fresh hash each time, we don't have to # bother adding make-work code to decrease the depth # when we return from what we called. my $ctx = shift; return {%$ctx, depth => $ctx->{depth}+1}; } ##################################################################### sub _make_class { my $self = shift; my %set = map { ($_,1) } @_; delete $set{'\\d'} if exists $set{'\\w'}; delete $set{'\\D'} if exists $set{'\\W'}; return '.' if exists $set{'.'} or ($self->{fold_meta_pairs} and ( (exists $set{'\\d'} and exists $set{'\\D'}) or (exists $set{'\\s'} and exists $set{'\\S'}) or (exists $set{'\\w'} and exists $set{'\\W'}) )) ; for my $meta( q/\\d/, q/\\D/, q/\\s/, q/\\S/, q/\\w/, q/\\W/ ) { if( exists $set{$meta} ) { my $re = qr/$meta/; my @delete; $_ =~ /^$re$/ and push @delete, $_ for keys %set; delete @set{@delete} if @delete; } } return (keys %set)[0] if keys %set == 1; for my $meta( '.', '+', '*', '?', '(', ')', '^', '@', '$', '[', '/', ) { exists $set{"\\$meta"} and $set{$meta} = delete $set{"\\$meta"}; } my $dash = exists $set{'-'} ? do { delete($set{'-'}), '-' } : ''; my $caret = exists $set{'^'} ? do { delete($set{'^'}), '^' } : ''; my $class = join( '' => sort keys %set ); $class =~ s/0123456789/\\d/ and $class eq '\\d' and return $class; return "[$dash$class$caret]"; } sub _re_sort { return length $b <=> length $a || $a cmp $b } sub _combine { my $self = shift; my $type = shift; # print "c in = @{[_dump(\@_)]}\n"; # my $combine = return '(' . $type . do { my( @short, @long ); push @{ /^$Single_Char$/ ? \@short : \@long}, $_ for @_; if( @short == 1 ) { @long = sort _re_sort @long, @short; } elsif( @short > 1 ) { # yucky but true my @combine = (_make_class($self, @short), sort _re_sort @long); @long = @combine; } else { @long = sort _re_sort @long; } join( '|', @long ); } . ')'; # print "combine <$combine>\n"; # $combine; } sub _combine_new { my $self = shift; my( @short, @long ); push @{ /^$Single_Char$/ ? \@short : \@long}, $_ for @_; if( @short == 1 and @long == 0 ) { return $short[0]; } elsif( @short > 1 and @short == @_ ) { return _make_class($self, @short); } else { return '(?:' . join( '|' => @short > 1 ? ( _make_class($self, @short), sort _re_sort @long) : ( (sort _re_sort( @long )), @short ) ) . ')'; } } sub _re_path { my $self = shift; # in shorter assemblies, _re_path() is the second hottest # routine. after insert(), so make it fast. if ($self->{unroll_plus}) { # but we can't easily make this blockless my @arr = @{$_[0]}; my $str = ''; my $skip = 0; for my $i (0..$#arr) { if (ref($arr[$i]) eq 'ARRAY') { $str .= _re_path($self, $arr[$i]); } elsif (ref($arr[$i]) eq 'HASH') { $str .= exists $arr[$i]->{''} ? _combine_new( $self, map { _re_path( $self, $arr[$i]->{$_} ) } grep { $_ ne '' } keys %{$arr[$i]} ) . '?' : _combine_new($self, map { _re_path( $self, $arr[$i]->{$_} ) } keys %{$arr[$i]}) ; } elsif ($i < $#arr and $arr[$i+1] =~ /\A$arr[$i]\*(\??)\Z/) { $str .= "$arr[$i]+" . (defined $1 ? $1 : ''); ++$skip; } elsif ($skip) { $skip = 0; } else { $str .= $arr[$i]; } } return $str; } return join( '', @_ ) unless grep { length ref $_ } @_; my $p; return join '', map { ref($_) eq '' ? $_ : ref($_) eq 'HASH' ? do { # In the case of a node, see whether there's a '' which # indicates that the whole thing is optional and thus # requires a trailing ? # Unroll the two different paths to avoid the needless # grep when it isn't necessary. $p = $_; exists $_->{''} ? _combine_new( $self, map { _re_path( $self, $p->{$_} ) } grep { $_ ne '' } keys %$_ ) . '?' : _combine_new($self, map { _re_path( $self, $p->{$_} ) } keys %$_ ) } : _re_path($self, $_) # ref($_) eq 'ARRAY' } @{$_[0]} } sub _lookahead { my $in = shift; my %head; my $path; for $path( keys %$in ) { next unless defined $in->{$path}; # print "look $path: ", ref($in->{$path}[0]), ".\n"; if( ref($in->{$path}[0]) eq 'HASH' ) { my $next = 0; while( ref($in->{$path}[$next]) eq 'HASH' and @{$in->{$path}} > $next + 1 ) { if( exists $in->{$path}[$next]{''} ) { ++$head{$in->{$path}[$next+1]}; } ++$next; } my $inner = _lookahead( $in->{$path}[0] ); @head{ keys %$inner } = (values %$inner); } elsif( ref($in->{$path}[0]) eq 'ARRAY' ) { my $subpath = $in->{$path}[0]; for( my $sp = 0; $sp < @$subpath; ++$sp ) { if( ref($subpath->[$sp]) eq 'HASH' ) { my $follow = _lookahead( $subpath->[$sp] ); @head{ keys %$follow } = (values %$follow); last unless exists $subpath->[$sp]{''}; } else { ++$head{$subpath->[$sp]}; last; } } } else { ++$head{ $in->{$path}[0] }; } } # print "_lookahead ", _dump($in), '==>', _dump([keys %head]), "\n"; return \%head; } sub _re_path_lookahead { my $self = shift; my $in = shift; # print "_re_path_la in ", _dump($in), "\n"; my $out = ''; for( my $p = 0; $p < @$in; ++$p ) { if( ref($in->[$p]) eq '' ) { $out .= $in->[$p]; next; } elsif( ref($in->[$p]) eq 'ARRAY' ) { $out .= _re_path_lookahead($self, $in->[$p]); next; } # print "$p ", _dump($in->[$p]), "\n"; my $path = [ map { _re_path_lookahead($self, $in->[$p]{$_} ) } grep { $_ ne '' } keys %{$in->[$p]} ]; my $ahead = _lookahead($in->[$p]); my $more = 0; if( exists $in->[$p]{''} and $p + 1 < @$in ) { my $next = 1; while( $p + $next < @$in ) { if( ref( $in->[$p+$next] ) eq 'HASH' ) { my $follow = _lookahead( $in->[$p+$next] ); @{$ahead}{ keys %$follow } = (values %$follow); } else { ++$ahead->{$in->[$p+$next]}; last; } ++$next; } $more = 1; } my $nr_one = grep { /^$Single_Char$/ } @$path; my $nr = @$path; if( $nr_one > 1 and $nr_one == $nr ) { $out .= _make_class($self, @$path); $out .= '?' if exists $in->[$p]{''}; } else { my $zwla = keys(%$ahead) > 1 ? _combine($self, '?=', grep { s/\+$//; $_ } keys %$ahead ) : ''; my $patt = $nr > 1 ? _combine($self, '?:', @$path ) : $path->[0]; # print "have nr=$nr n1=$nr_one n=", _dump($in->[$p]), ' a=', _dump([keys %$ahead]), " zwla=$zwla patt=$patt @{[_dump($path)]}\n"; if( exists $in->[$p]{''} ) { $out .= $more ? "$zwla(?:$patt)?" : "(?:$zwla$patt)?"; } else { $out .= "$zwla$patt"; } } } return $out; } sub _re_path_track { my $self = shift; my $in = shift; my $normal = shift; my $augmented = shift; my $o; my $simple = ''; my $augment = ''; for( my $n = 0; $n < @$in; ++$n ) { if( ref($in->[$n]) eq '' ) { $o = $in->[$n]; $simple .= $o; $augment .= $o; if( ( $n < @$in - 1 and ref($in->[$n+1]) eq 'HASH' and exists $in->[$n+1]{''} ) or $n == @$in - 1 ) { push @{$self->{mlist}}, $normal . $simple ; $augment .= $] < 5.009005 ? "(?{\$self->{m}=$self->{mcount}})" : "(?{$self->{mcount}})" ; ++$self->{mcount}; } } else { my $path = [ map { $self->_re_path_track( $in->[$n]{$_}, $normal.$simple , $augmented.$augment ) } grep { $_ ne '' } keys %{$in->[$n]} ]; $o = '(?:' . join( '|' => sort _re_sort @$path ) . ')'; $o .= '?' if exists $in->[$n]{''}; $simple .= $o; $augment .= $o; } } return $augment; } sub _re_path_pretty { my $self = shift; my $in = shift; my $arg = shift; my $pre = ' ' x (($arg->{depth}+0) * $arg->{indent}); my $indent = ' ' x (($arg->{depth}+1) * $arg->{indent}); my $out = ''; $arg->{depth}++; my $prev_was_paren = 0; for( my $p = 0; $p < @$in; ++$p ) { if( ref($in->[$p]) eq '' ) { $out .= "\n$pre" if $prev_was_paren; $out .= $in->[$p]; $prev_was_paren = 0; } elsif( ref($in->[$p]) eq 'ARRAY' ) { $out .= _re_path($self, $in->[$p]); } else { my $path = [ map { _re_path_pretty($self, $in->[$p]{$_}, $arg ) } grep { $_ ne '' } keys %{$in->[$p]} ]; my $nr = @$path; my( @short, @long ); push @{/^$Single_Char$/ ? \@short : \@long}, $_ for @$path; if( @short == $nr ) { $out .= $nr == 1 ? $path->[0] : _make_class($self, @short); $out .= '?' if exists $in->[$p]{''}; } else { $out .= "\n" if length $out; $out .= $pre if $p; $out .= "(?:\n$indent"; if( @short < 2 ) { my $r = 0; $out .= join( "\n$indent|" => map { $r++ and $_ =~ s/^\(\?:/\n$indent(?:/; $_ } sort _re_sort @$path ); } else { $out .= join( "\n$indent|" => ( (sort _re_sort @long), _make_class($self, @short) )); } $out .= "\n$pre)"; if( exists $in->[$p]{''} ) { $out .= "\n$pre?"; $prev_was_paren = 0; } else { $prev_was_paren = 1; } } } } $arg->{depth}--; return $out; } sub _node_eq { return 0 if not defined $_[0] or not defined $_[1]; return 0 if ref $_[0] ne ref $_[1]; # Now that we have determined that the reference of each # argument are the same, we only have to test the first # one, which gives us a nice micro-optimisation. if( ref($_[0]) eq 'HASH' ) { keys %{$_[0]} == keys %{$_[1]} and # does this short-circuit to avoid _re_path() cost more than it saves? join( '|' => sort keys %{$_[0]}) eq join( '|' => sort keys %{$_[1]}) and _re_path(undef, [$_[0]] ) eq _re_path(undef, [$_[1]] ); } elsif( ref($_[0]) eq 'ARRAY' ) { scalar @{$_[0]} == scalar @{$_[1]} and _re_path(undef, $_[0]) eq _re_path(undef, $_[1]); } else { $_[0] eq $_[1]; } } sub _pretty_dump { return sprintf "\\x%02x", ord(shift); } sub _dump { my $path = shift; return _dump_node($path) if ref($path) eq 'HASH'; my $dump = '['; my $d; my $nr = 0; for $d( @$path ) { $dump .= ' ' if $nr++; if( ref($d) eq 'HASH' ) { $dump .= _dump_node($d); } elsif( ref($d) eq 'ARRAY' ) { $dump .= _dump($d); } elsif( defined $d ) { # D::C indicates the second test is redundant # $dump .= ( $d =~ /\s/ or not length $d ) $dump .= ( $d =~ /\s/ ? qq{'$d'} : $d =~ /^[\x00-\x1f]$/ ? _pretty_dump($d) : $d ); } else { $dump .= '*'; } } return $dump . ']'; } sub _dump_node { my $node = shift; my $dump = '{'; my $nr = 0; my $n; for $n (sort keys %$node) { $dump .= ' ' if $nr++; # Devel::Cover shows this to test to be redundant # $dump .= ( $n eq '' and not defined $node->{$n} ) $dump .= $n eq '' ? '*' : ($n =~ /^[\x00-\x1f]$/ ? _pretty_dump($n) : $n) . "=>" . _dump($node->{$n}) ; } return $dump . '}'; } =pod =head1 NAME Regexp::Assemble - Assemble multiple Regular Expressions into a single RE =head1 SYNOPSIS use Regexp::Assemble; my $ra = Regexp::Assemble->new; $ra->add( 'ab+c' ); $ra->add( 'ab+-' ); $ra->add( 'a\w\d+' ); $ra->add( 'a\d+' ); print $ra->re; # prints a(?:\w?\d+|b+[-c]) =head1 DESCRIPTION Regexp::Assemble takes an arbitrary number of regular expressions and assembles them into a single regular expression (or RE) that matches all that the individual REs match. As a result, instead of having a large list of expressions to loop over, a target string only needs to be tested against one expression. This is interesting when you have several thousand patterns to deal with. Serious effort is made to produce the smallest pattern possible. It is also possible to track the original patterns, so that you can determine which, among the source patterns that form the assembled pattern, was the one that caused the match to occur. You should realise that large numbers of alternations are processed in perl's regular expression engine in O(n) time, not O(1). If you are still having performance problems, you should look at using a trie. Note that Perl's own regular expression engine will implement trie optimisations in perl 5.10 (they are already available in perl 5.9.3 if you want to try them out). C<Regexp::Assemble> will do the right thing when it knows it's running on a trie'd perl. (At least in some version after this one). Some more examples of usage appear in the accompanying README. If that file is not easy to access locally, you can find it on a web repository such as L<http://search.cpan.org/dist/Regexp-Assemble/README> or L<http://cpan.uwinnipeg.ca/htdocs/Regexp-Assemble/README.html>. See also L</LIMITATIONS>. =head1 Methods =head2 add(LIST) Takes a string, breaks it apart into a set of tokens (respecting meta characters) and inserts the resulting list into the C<R::A> object. It uses a naive regular expression to lex the string that may be fooled complex expressions (specifically, it will fail to lex nested parenthetical expressions such as C<ab(cd(ef)?gh)ij> correctly). If this is the case, the end of the string will not be tokenised correctly and returned as one long string. On the one hand, this may indicate that the patterns you are trying to feed the C<R::A> object are too complex. Simpler patterns might allow the algorithm to work more effectively and perform more reductions in the resulting pattern. On the other hand, you can supply your own pattern to perform the lexing if you need. The test suite contains an example of a lexer pattern that will match one level of nested parentheses. Note that there is an internal optimisation that will bypass a much of the lexing process. If a string contains no C<\> (backslash), C<[> (open square bracket), C<(> (open paren), C<?> (question mark), C<+> (plus), C<*> (star) or C<{> (open curly), a character split will be performed directly. A list of strings may be supplied, thus you can pass it a file handle of a file opened for reading: $re->add( '\d+-\d+-\d+-\d+\.example\.com' ); $re->add( <IN> ); If the file is very large, it may be more efficient to use a C<while> loop, to read the file line-by-line: $re->add($_) while <IN>; The C<add> method will chomp the lines automatically. If you do not want this to occur (you want to keep the record separator), then disable C<chomp>ing. $re->chomp(0); $re->add($_) while <IN>; This method is chainable. =head2 add_file(FILENAME [...]) Takes a list of file names. Each file is opened and read line by line. Each line is added to the assembly. $r->add_file( 'file.1', 'file.2' ); If a file cannot be opened, the method will croak. If you cannot afford to let this happen then you should wrap the call in a C<eval> block. Chomping happens automatically unless you the C<chomp(0)> method to disable it. By default, input lines are read according to the value of the C<input_record_separator> attribute (if defined), and will otherwise fall back to the current setting of the system C<$/> variable. The record separator may also be specified on each call to C<add_file>. Internally, the routine C<local>ises the value of C<$/> to whatever is required, for the duration of the call. An alternate calling mechanism using a hash reference is available. The recognised keys are: =over 4 =item file Reference to a list of file names, or the name of a single file. $r->add_file({file => ['file.1', 'file.2', 'file.3']}); $r->add_file({file => 'file.n'}); =item input_record_separator If present, indicates what constitutes a line $r->add_file({file => 'data.txt', input_record_separator => ':' }); =item rs An alias for input_record_separator (mnemonic: same as the English variable names). =back $r->add_file( { file => [ 'pattern.txt', 'more.txt' ], input_record_separator => "\r\n", }); =head2 clone() Clones the contents of a Regexp::Assemble object and creates a new object (in other words it performs a deep copy). If the Storable module is installed, its dclone method will be used, otherwise the cloning will be performed using a pure perl approach. You can use this method to take a snapshot of the patterns that have been added so far to an object, and generate an assembly from the clone. Additional patterns may to be added to the original object afterwards. my $re = $main->clone->re(); $main->add( 'another-pattern-\\d+' ); =head2 insert(LIST) Takes a list of tokens representing a regular expression and stores them in the object. Note: you should not pass it a bare regular expression, such as C<ab+c?d*e>. You must pass it as a list of tokens, I<e.g.> C<('a', 'b+', 'c?', 'd*', 'e')>. This method is chainable, I<e.g.>: my $ra = Regexp::Assemble->new ->insert( qw[ a b+ c? d* e ] ) ->insert( qw[ a c+ d+ e* f ] ); Lexing complex patterns with metacharacters and so on can consume a significant proportion of the overall time to build an assembly. If you have the information available in a tokenised form, calling C<insert> directly can be a big win. =head2 lexstr Use the C<lexstr> method if you are curious to see how a pattern gets tokenised. It takes a scalar on input, representing a pattern, and returns a reference to an array, containing the tokenised pattern. You can recover the original pattern by performing a C<join>: my @token = $re->lexstr($pattern); my $new_pattern = join( '', @token ); If the original pattern contains unnecessary backslashes, or C<\x4b> escapes, or quotemeta escapes (C<\Q>...C<\E>) the resulting pattern may not be identical. Call C<lexstr> does not add the pattern to the object, it is merely for exploratory purposes. It will, however, update various statistical counters. =head2 pre_filter(CODE) Allows you to install a callback to check that the pattern being loaded contains valid input. It receives the pattern as a whole to be added, before it been tokenised by the lexer. It may to return 0 or C<undef> to indicate that the pattern should not be added, any true value indicates that the contents are fine. A filter to strip out trailing comments (marked by #): $re->pre_filter( sub { $_[0] =~ s/\s*#.*$//; 1 } ); A filter to ignore blank lines: $re->pre_filter( sub { length(shift) } ); If you want to remove the filter, pass C<undef> as a parameter. $ra->pre_filter(undef); This method is chainable. =head2 filter(CODE) Allows you to install a callback to check that the pattern being loaded contains valid input. It receives a list on input, after it has been tokenised by the lexer. It may to return 0 or undef to indicate that the pattern should not be added, any true value indicates that the contents are fine. If you know that all patterns you expect to assemble contain a restricted set of of tokens (e.g. no spaces), you could do the following: $ra->filter(sub { not grep { / / } @_ }); or sub only_spaces_and_digits { not grep { ![\d ] } @_ } $ra->filter( \&only_spaces_and_digits ); These two examples will silently ignore faulty patterns, If you want the user to be made aware of the problem you should raise an error (via C<warn> or C<die>), log an error message, whatever is best. If you want to remove a filter, pass C<undef> as a parameter. $ra->filter(undef); This method is chainable. =head2 as_string Assemble the expression and return it as a string. You may want to do this if you are writing the pattern to a file. The following arguments can be passed to control the aspect of the resulting pattern: B<indent>, the number of spaces used to indent nested grouping of a pattern. Use this to produce a pretty-printed pattern (for some definition of "pretty"). The resulting output is rather verbose. The reason is to ensure that the metacharacters C<(?:> and C<)> always occur on otherwise empty lines. This allows you grep the result for an even more synthetic view of the pattern: egrep -v '^ *[()]' <regexp.file> The result of the above is quite readable. Remember to backslash the spaces appearing in your own patterns if you wish to use an indented pattern in an C<m/.../x> construct. Indenting is ignored if tracking is enabled. The B<indent> argument takes precedence over the C<indent> method/attribute of the object. Calling this method will drain the internal data structure. Large numbers of patterns can eat a significant amount of memory, and this lets perl recover the memory used for other purposes. If you want to reduce the pattern I<and> continue to add new patterns, clone the object and reduce the clone, leaving the original object intact. =head2 re Assembles the pattern and return it as a compiled RE, using the C<qr//> operator. As with C<as_string>, calling this method will reset the internal data structures to free the memory used in assembling the RE. The B<indent> attribute, documented in the C<as_string> method, can be used here (it will be ignored if tracking is enabled). With method chaining, it is possible to produce a RE without having a temporary C<Regexp::Assemble> object lying around, I<e.g.>: my $re = Regexp::Assemble->new ->add( q[ab+cd+e] ) ->add( q[ac\\d+e] ) ->add( q[c\\d+e] ) ->re; The C<$re> variable now contains a Regexp object that can be used directly: while( <> ) { /$re/ and print "Something in [$_] matched\n"; ) The C<re> method is called when the object is used in string context (hence, within an C<m//> operator), so by and large you do not even need to save the RE in a separate variable. The following will work as expected: my $re = Regexp::Assemble->new->add( qw[ fee fie foe fum ] ); while( <IN> ) { if( /($re)/ ) { print "Here be giants: $1\n"; } } This approach does not work with tracked patterns. The C<match> and C<matched> methods must be used instead, see below. =head2 match(SCALAR) The following information applies to Perl 5.8 and below. See the section that follows for information on Perl 5.10. If pattern tracking is in use, you must C<use re 'eval'> in order to make things work correctly. At a minimum, this will make your code look like this: my $did_match = do { use re 'eval'; $target =~ /$ra/ } if( $did_match ) { print "matched ", $ra->matched, "\n"; } (The main reason is that the C<$^R> variable is currently broken and an ugly workaround that runs some Perl code during the match is required, in order to simulate what C<$^R> should be doing. See Perl bug #32840 for more information if you are curious. The README also contains more information). This bug has been fixed in 5.10. The important thing to note is that with C<use re 'eval'>, THERE ARE SECURITY IMPLICATIONS WHICH YOU IGNORE AT YOUR PERIL. The problem is this: if you do not have strict control over the patterns being fed to C<Regexp::Assemble> when tracking is enabled, and someone slips you a pattern such as C</^(?{system 'rm -rf /'})/> and you attempt to match a string against the resulting pattern, you will know Fear and Loathing. What is more, the C<$^R> workaround means that that tracking does not work if you perform a bare C</$re/> pattern match as shown above. You have to instead call the C<match> method, in order to supply the necessary context to take care of the tracking housekeeping details. if( defined( my $match = $ra->match($_)) ) { print " $_ matched by $match\n"; } In the case of a successful match, the original matched pattern is returned directly. The matched pattern will also be available through the C<matched> method. (Except that the above is not true for 5.6.0: the C<match> method returns true or undef, and the C<matched> method always returns undef). If you are capturing parts of the pattern I<e.g.> C<foo(bar)rat> you will want to get at the captures. See the C<mbegin>, C<mend>, C<mvar> and C<capture> methods. If you are not using captures then you may safely ignore this section. In 5.10, since the bug concerning C<$^R> has been resolved, there is no need to use C<re 'eval'> and the assembled pattern does not require any Perl code to be executed during the match. =head2 new() Creates a new C<Regexp::Assemble> object. The following optional key/value parameters may be employed. All keys have a corresponding method that can be used to change the behaviour later on. As a general rule, especially if you're just starting out, you don't have to bother with any of these. B<anchor_*>, a family of optional attributes that allow anchors (C<^>, C<\b>, C<\Z>...) to be added to the resulting pattern. B<flags>, sets the C<imsx> flags to add to the assembled regular expression. Warning: no error checking is done, you should ensure that the flags you pass are understood by the version of Perl you are using. B<modifiers> exists as an alias, for users familiar with L<Regexp::List>. B<chomp>, controls whether the pattern should be chomped before being lexed. Handy if you are reading patterns from a file. By default, C<chomp>ing is performed (this behaviour changed as of version 0.24, prior versions did not chomp automatically). See also the C<file> attribute and the C<add_file> method. B<file>, slurp the contents of the specified file and add them to the assembly. Multiple files may be processed by using a list. my $r = Regexp::Assemble->new(file => 're.list'); my $r = Regexp::Assemble->new(file => ['re.1', 're.2']); If you really don't want chomping to occur, you will have to set the C<chomp> attribute to 0 (zero). You may also want to look at the C<input_record_separator> attribute, as well. B<input_record_separator>, controls what constitutes a record separator when using the C<file> attribute or the C<add_file> method. May be abbreviated to B<rs>. See the C<$/> variable in L<perlvar>. B<lookahead>, controls whether the pattern should contain zero-width lookahead assertions (For instance: (?=[abc])(?:bob|alice|charles). This is not activated by default, because in many circumstances the cost of processing the assertion itself outweighs the benefit of its faculty for short-circuiting a match that will fail. This is sensitive to the probability of a match succeeding, so if you're worried about performance you'll have to benchmark a sample population of targets to see which way the benefits lie. B<track>, controls whether you want know which of the initial patterns was the one that matched. See the C<matched> method for more details. Note for version 5.8 of Perl and below, in this mode of operation YOU SHOULD BE AWARE OF THE SECURITY IMPLICATIONS that this entails. Perl 5.10 does not suffer from any such restriction. B<indent>, the number of spaces used to indent nested grouping of a pattern. Use this to produce a pretty-printed pattern. See the C<as_string> method for a more detailed explanation. B<pre_filter>, allows you to add a callback to enable sanity checks on the pattern being loaded. This callback is triggered before the pattern is split apart by the lexer. In other words, it operates on the entire pattern. If you are loading patterns from a file, this would be an appropriate place to remove comments. B<filter>, allows you to add a callback to enable sanity checks on the pattern being loaded. This callback is triggered after the pattern has been split apart by the lexer. B<unroll_plus>, controls whether to unroll, for example, C<x+> into C<x>, C<x*>, which may allow additional reductions in the resulting assembled pattern. B<reduce>, controls whether tail reduction occurs or not. If set, patterns like C<a(?:bc+d|ec+d)> will be reduced to C<a[be]c+d>. That is, the end of the pattern in each part of the b... and d... alternations is identical, and hence is hoisted out of the alternation and placed after it. On by default. Turn it off if you're really pressed for short assembly times. B<lex>, specifies the pattern used to lex the input lines into tokens. You could replace the default pattern by a more sophisticated version that matches arbitrarily nested parentheses, for example. B<debug>, controls whether copious amounts of output is produced during the loading stage or the reducing stage of assembly. my $ra = Regexp::Assemble->new; my $rb = Regexp::Assemble->new( chomp => 1, debug => 3 ); B<mutable>, controls whether new patterns can be added to the object after the assembled pattern is generated. DEPRECATED. This method/attribute will be removed in a future release. It doesn't really serve any purpose, and may be more effectively replaced by cloning an existing C<Regexp::Assemble> object and spinning out a pattern from that instead. =head2 source() When using tracked mode, after a successful match is made, returns the original source pattern that caused the match. In Perl 5.10, the C<$^R> variable can be used to as an index to fetch the correct pattern from the object. If no successful match has been performed, or the object is not in tracked mode, this method returns C<undef>. my $r = Regexp::Assemble->new->track(1)->add(qw(foo? bar{2} [Rr]at)); for my $w (qw(this food is rather barren)) { if ($w =~ /$r/) { print "$w matched by ", $r->source($^R), $/; } else { print "$w no match\n"; } } =head2 mbegin() This method returns a copy of C<@-> at the moment of the last match. You should ordinarily not need to bother with this, C<mvar> should be able to supply all your needs. =head2 mend() This method returns a copy of C<@+> at the moment of the last match. =head2 mvar(NUMBER) The C<mvar> method returns the captures of the last match. C<mvar(1)> corresponds to $1, C<mvar(2)> to $2, and so on. C<mvar(0)> happens to return the target string matched, as a byproduct of walking down the C<@-> and C<@+> arrays after the match. If called without a parameter, C<mvar> will return a reference to an array containing all captures. =head2 capture The C<capture> method returns the the captures of the last match as an array. Unlink C<mvar>, this method does not include the matched string. It is equivalent to getting an array back that contains C<$1, $2, $3, ...>. If no captures were found in the match, an empty array is returned, rather than C<undef>. You are therefore guaranteed to be able to use C<< for my $c ($re->capture) { ... >> without have to check whether anything was captured. =head2 matched() If pattern tracking has been set, via the C<track> attribute, or through the C<track> method, this method will return the original pattern of the last successful match. Returns undef match has yet been performed, or tracking has not been enabled. See below in the NOTES section for additional subtleties of which you should be aware of when tracking patterns. Note that this method is not available in 5.6.0, due to limitations in the implementation of C<(?{...})> at the time. =head2 Statistics/Reporting routines =head2 stats_add Returns the number of patterns added to the assembly (whether by C<add> or C<insert>). Duplicate patterns are not included in this total. =head2 stats_dup Returns the number of duplicate patterns added to the assembly. If non-zero, this may be a sign that something is wrong with your data (or at the least, some needless redundancy). This may occur when you have two patterns (for instance, C<a\-b> and C<a-b>) which map to the same result. =head2 stats_raw() Returns the raw number of bytes in the patterns added to the assembly. This includes both original and duplicate patterns. For instance, adding the two patterns C<ab> and C<ab> will count as 4 bytes. =head2 stats_cooked() Return the true number of bytes added to the assembly. This will not include duplicate patterns. Furthermore, it may differ from the raw bytes due to quotemeta treatment. For instance, C<abc\,def> will count as 7 (not 8) bytes, because C<\,> will be stored as C<,>. Also, C<\Qa.b\E> is 7 bytes long, however, after the quotemeta directives are processed, C<a\.b> will be stored, for a total of 4 bytes. =head2 stats_length() Returns the length of the resulting assembled expression. Until C<as_string> or C<re> have been called, the length will be 0 (since the assembly will have not yet been performed). The length includes only the pattern, not the additional (C<(?-xism...>) fluff added by the compilation. =head2 dup_warn(NUMBER|CODEREF) Turns warnings about duplicate patterns on or off. By default, no warnings are emitted. If the method is called with no parameters, or a true parameter, the object will carp about patterns it has already seen. To turn off the warnings, use 0 as a parameter. $r->dup_warn(); The method may also be passed a code block. In this case the code will be executed and it will receive a reference to the object in question, and the lexed pattern. $r->dup_warn( sub { my $self = shift; print $self->stats_add, " patterns added at line $.\n", join( '', @_ ), " added previously\n"; } ) =head2 Anchor routines Suppose you wish to assemble a series of patterns that all begin with C<^> and end with C<$> (anchor pattern to the beginning and end of line). Rather than add the anchors to each and every pattern (and possibly forget to do so when a new entry is added), you may specify the anchors in the object, and they will appear in the resulting pattern, and you no longer need to (or should) put them in your source patterns. For example, the two following snippets will produce identical patterns: $r->add(qw(^this ^that ^them))->as_string; $r->add(qw(this that them))->anchor_line_begin->as_string; # both techniques will produce ^th(?:at|em|is) All anchors are possible word (C<\b>) boundaries, line boundaries (C<^> and C<$>) and string boundaries (C<\A> and C<\Z> (or C<\z> if you absolutely need it)). The shortcut C<anchor_I<mumble>> implies both C<anchor_I<mumble>_begin> C<anchor_I<mumble>_end> is also available. If different anchors are specified the most specific anchor wins. For instance, if both C<anchor_word_begin> and C<anchor_line_begin> are specified, C<anchor_word_begin> takes precedence. All the anchor methods are chainable. =head2 anchor_word_begin The resulting pattern will be prefixed with a C<\b> word boundary assertion when the value is true. Set to 0 to disable. $r->add('pre')->anchor_word_begin->as_string; # produces '\bpre' =head2 anchor_word_end The resulting pattern will be suffixed with a C<\b> word boundary assertion when the value is true. Set to 0 to disable. $r->add(qw(ing tion)) ->anchor_word_end ->as_string; # produces '(?:tion|ing)\b' =head2 anchor_word The resulting pattern will be have C<\b> word boundary assertions at the beginning and end of the pattern when the value is true. Set to 0 to disable. $r->add(qw(cat carrot) ->anchor_word(1) ->as_string; # produces '\bca(?:rro)t\b' =head2 anchor_line_begin The resulting pattern will be prefixed with a C<^> line boundary assertion when the value is true. Set to 0 to disable. $r->anchor_line_begin; # or $r->anchor_line_begin(1); =head2 anchor_line_end The resulting pattern will be suffixed with a C<$> line boundary assertion when the value is true. Set to 0 to disable. # turn it off $r->anchor_line_end(0); =head2 anchor_line The resulting pattern will be have the C<^> and C<$> line boundary assertions at the beginning and end of the pattern, respectively, when the value is true. Set to 0 to disable. $r->add(qw(cat carrot) ->anchor_line ->as_string; # produces '^ca(?:rro)t$' =head2 anchor_string_begin The resulting pattern will be prefixed with a C<\A> string boundary assertion when the value is true. Set to 0 to disable. $r->anchor_string_begin(1); =head2 anchor_string_end The resulting pattern will be suffixed with a C<\Z> string boundary assertion when the value is true. Set to 0 to disable. # disable the string boundary end anchor $r->anchor_string_end(0); =head2 anchor_string_end_absolute The resulting pattern will be suffixed with a C<\z> string boundary assertion when the value is true. Set to 0 to disable. # disable the string boundary absolute end anchor $r->anchor_string_end_absolute(0); If you don't understand the difference between C<\Z> and C<\z>, the former will probably do what you want. =head2 anchor_string The resulting pattern will be have the C<\A> and C<\Z> string boundary assertions at the beginning and end of the pattern, respectively, when the value is true. Set to 0 to disable. $r->add(qw(cat carrot) ->anchor_string ->as_string; # produces '\Aca(?:rro)t\Z' =head2 anchor_string_absolute The resulting pattern will be have the C<\A> and C<\z> string boundary assertions at the beginning and end of the pattern, respectively, when the value is true. Set to 0 to disable. $r->add(qw(cat carrot) ->anchor_string_absolute ->as_string; # produces '\Aca(?:rro)t\z' =head2 debug(NUMBER) Turns debugging on or off. Statements are printed to the currently selected file handle (STDOUT by default). If you are already using this handle, you will have to arrange to select an output handle to a file of your own choosing, before call the C<add>, C<as_string> or C<re>) functions, otherwise it will scribble all over your carefully formatted output. =over 4 =item * 0 Off. Turns off all debugging output. =item * 1 Add. Trace the addition of patterns. =item * 2 Reduce. Trace the process of reduction and assembly. =item * 4 Lex. Trace the lexing of the input patterns into its constituent tokens. =item * 8 Time. Print to STDOUT the time taken to load all the patterns. This is nothing more than the difference between the time the object was instantiated and the time reduction was initiated. # load=<num> Any lengthy computation performed in the client code will be reflected in this value. Another line will be printed after reduction is complete. # reduce=<num> The above output lines will be changed to C<load-epoch> and C<reduce-epoch> if the internal state of the object is corrupted and the initial timestamp is lost. The code attempts to load L<Time::HiRes> in order to report fractional seconds. If this is not successful, the elapsed time is displayed in whole seconds. =back Values can be added (or or'ed together) to trace everything $r->debug(7)->add( '\\d+abc' ); Calling C<debug> with no arguments turns debugging off. =head2 dump() Produces a synthetic view of the internal data structure. How to interpret the results is left as an exercise to the reader. print $r->dump; =head2 chomp(0|1) Turns chomping on or off. IMPORTANT: As of version 0.24, chomping is now on by default as it makes C<add_file> Just Work. The only time you may run into trouble is with C<add("\\$/")>. So don't do that, or else explicitly turn off chomping. To avoid incorporating (spurious) record separators (such as "\n" on Unix) when reading from a file, C<add()> C<chomp>s its input. If you don't want this to happen, call C<chomp> with a false value. $re->chomp(0); # really want the record separators $re->add(<DATA>); =head2 fold_meta_pairs(NUMBER) Determines whether C<\s>, C<\S> and C<\w>, C<\W> and C<\d>, C<\D> are folded into a C<.> (dot). Folding happens by default (for reasons of backwards compatibility, even though it is wrong when the C</s> expression modifier is active). Call this method with a false value to prevent this behaviour (which is only a problem when dealing with C<\n> if the C</s> expression modifier is also set). $re->add( '\\w', '\\W' ); my $clone = $re->clone; $clone->fold_meta_pairs(0); print $clone->as_string; # prints '.' print $re->as_string; # print '[\W\w]' =head2 indent(NUMBER) Sets the level of indent for pretty-printing nested groups within a pattern. See the C<as_string> method for more details. When called without a parameter, no indenting is performed. $re->indent( 4 ); print $re->as_string; =head2 lookahead(0|1) Turns on zero-width lookahead assertions. This is usually beneficial when you expect that the pattern will usually fail. If you expect that the pattern will usually match you will probably be worse off. =head2 flags(STRING) Sets the flags that govern how the pattern behaves (for versions of Perl up to 5.9 or so, these are C<imsx>). By default no flags are enabled. =head2 modifiers(STRING) An alias of the C<flags> method, for users familiar with C<Regexp::List>. =head2 track(0|1) Turns tracking on or off. When this attribute is enabled, additional housekeeping information is inserted into the assembled expression using C<({...}> embedded code constructs. This provides the necessary information to determine which, of the original patterns added, was the one that caused the match. $re->track( 1 ); if( $target =~ /$re/ ) { print "$target matched by ", $re->matched, "\n"; } Note that when this functionality is enabled, no reduction is performed and no character classes are generated. In other words, C<brag|tag> is not reduced down to C<(?:br|t)ag> and C<dig|dim> is not reduced to C<di[gm]>. =head2 unroll_plus(0|1) Turns the unrolling of plus metacharacters on or off. When a pattern is broken up, C<a+> becomes C<a>, C<a*> (and C<b+?> becomes C<b>, C<b*?>. This may allow the freed C<a> to assemble with other patterns. Not enabled by default. =head2 lex(SCALAR) Change the pattern used to break a string apart into tokens. You can examine the C<eg/naive> script as a starting point. =head2 reduce(0|1) Turns pattern reduction on or off. A reduced pattern may be considerably shorter than an unreduced pattern. Consider C</sl(?:ip|op|ap)/> I<versus> C</sl[aio]p/>. An unreduced pattern will be very similar to those produced by C<Regexp::Optimizer>. Reduction is on by default. Turning it off speeds assembly (but assembly is pretty fast -- it's the breaking up of the initial patterns in the lexing stage that can consume a non-negligible amount of time). =head2 mutable(0|1) This method has been marked as DEPRECATED. It will be removed in a future release. See the C<clone> method for a technique to replace its functionality. =head2 reset() Empties out the patterns that have been C<add>ed or C<insert>-ed into the object. Does not modify the state of controller attributes such as C<debug>, C<lex>, C<reduce> and the like. =head2 Default_Lexer B<Warning:> the C<Default_Lexer> function is a class method, not an object method. It is a fatal error to call it as an object method. The C<Default_Lexer> method lets you replace the default pattern used for all subsequently created C<Regexp::Assemble> objects. It will not have any effect on existing objects. (It is also possible to override the lexer pattern used on a per-object basis). The parameter should be an ordinary scalar, not a compiled pattern. If the pattern fails to match all parts of the string, the missing parts will be returned as single chunks. Therefore the following pattern is legal (albeit rather cork-brained): Regexp::Assemble::Default_Lexer( '\\d' ); The above pattern will split up input strings digit by digit, and all non-digit characters as single chunks. =head1 DIAGNOSTICS "Cannot pass a C<refname> to Default_Lexer" You tried to replace the default lexer pattern with an object instead of a scalar. Solution: You probably tried to call C<< $obj->Default_Lexer >>. Call the qualified class method instead C<Regexp::Assemble::Default_Lexer>. "filter method not passed a coderef" "pre_filter method not passed a coderef" A reference to a subroutine (anonymous or otherwise) was expected. Solution: read the documentation for the C<filter> method. "duplicate pattern added: /.../" The C<dup_warn> attribute is active, and a duplicate pattern was added (well duh!). Solution: clean your data. "cannot open [file] for input: [reason]" The C<add_file> method was unable to open the specified file for whatever reason. Solution: make sure the file exists and the script has the required privileges to read it. =head1 NOTES This module has been tested successfully with a range of versions of perl, from 5.005_03 to 5.9.3. Use of 5.6.0 is not recommended. The expressions produced by this module can be used with the PCRE library. Remember to "double up" your backslashes if the patterns are hard-coded as constants in your program. That is, you should literally C<add('a\\d+b')> rather than C<add('a\d+b')>. It usually will work either way, but it's good practice to do so. Where possible, supply the simplest tokens possible. Don't add C<X(?-\d+){2})Y> when C<X-\d+-\d+Y> will do. The reason is that if you also add C<X\d+Z> the resulting assembly changes dramatically: C<X(?:(?:-\d+){2}Y|-\d+Z)> I<versus> C<X-\d+(?:-\d+Y|Z)>. Since R::A doesn't perform enough analysis, it won't "unroll" the C<{2}> quantifier, and will fail to notice the divergence after the first C<-d\d+>. Furthermore, when the string 'X-123000P' is matched against the first assembly, the regexp engine will have to backtrack over each alternation (the one that ends in Y B<and> the one that ends in Z) before determining that there is no match. No such backtracking occurs in the second pattern: as soon as the engine encounters the 'P' in the target string, neither of the alternations at that point (C<-\d+Y> or C<Z>) could succeed and so the match fails. C<Regexp::Assemble> does, however, know how to build character classes. Given C<a-b>, C<axb> and C<a\db>, it will assemble these into C<a[-\dx]b>. When C<-> (dash) appears as a candidate for a character class it will be the first character in the class. When C<^> (circumflex) appears as a candidate for a character class it will be the last character in the class. It also knows about meta-characters than can "absorb" regular characters. For instance, given C<X\d> and C<X5>, it knows that C<5> can be represented by C<\d> and so the assembly is just C<X\d>. The "absorbent" meta-characters it deals with are C<.>, C<\d>, C<\s> and C<\W> and their complements. It will replace C<\d>/C<\D>, C<\s>/C<\S> and C<\w>/C<\W> by C<.> (dot), and it will drop C<\d> if C<\w> is also present (as will C<\D> in the presence of C<\W>). C<Regexp::Assemble> deals correctly with C<quotemeta>'s propensity to backslash many characters that have no need to be. Backslashes on non-metacharacters will be removed. Similarly, in character classes, a number of characters lose their magic and so no longer need to be backslashed within a character class. Two common examples are C<.> (dot) and C<$>. Such characters will lose their backslash. At the same time, it will also process C<\Q...\E> sequences. When such a sequence is encountered, the inner section is extracted and C<quotemeta> is applied to the section. The resulting quoted text is then used in place of the original unquoted text, and the C<\Q> and C<\E> metacharacters are thrown away. Similar processing occurs with the C<\U...\E> and C<\L...\E> sequences. This may have surprising effects when using a dispatch table. In this case, you will need to know exactly what the module makes of your input. Use the C<lexstr> method to find out what's going on: $pattern = join( '', @{$re->lexstr($pattern)} ); If all the digits 0..9 appear in a character class, C<Regexp::Assemble> will replace them by C<\d>. I'd do it for letters as well, but thinking about accented characters and other glyphs hurts my head. In an alternation, the longest paths are chosen first (for example, C<horse|bird|dog>). When two paths have the same length, the path with the most subpaths will appear first. This aims to put the "busiest" paths to the front of the alternation. For example, the list C<bad>, C<bit>, C<few>, C<fig> and C<fun> will produce the pattern C<(?:f(?:ew|ig|un)|b(?:ad|it))>. See F<eg/tld> for a real-world example of how alternations are sorted. Once you have looked at that, everything should be crystal clear. When tracking is in use, no reduction is performed. nor are character classes formed. The reason is that it is too difficult to determine the original pattern afterwards. Consider the two patterns C<pale> and C<palm>. These should be reduced to C<pal[em]>. The final character matches one of two possibilities. To resolve whether it matched an C<'e'> or C<'m'> would require keeping track of the fact that the pattern finished up in a character class, which would the require a whole lot more work to figure out which character of the class matched. Without character classes it becomes much easier. Instead, C<pal(?:e|m)> is produced, which lets us find out more simply where we ended up. Similarly, C<dogfood> and C<seafood> should form C<(?:dog|sea)food>. When the pattern is being assembled, the tracking decision needs to be made at the end of the grouping, but the tail of the pattern has not yet been visited. Deferring things to make this work correctly is a vast hassle. In this case, the pattern becomes merely C<(?:dogfood|seafood>. Tracked patterns will therefore be bulkier than simple patterns. There is an open bug on this issue: L<http://rt.perl.org/rt3/Ticket/Display.html?id=32840> If this bug is ever resolved, tracking would become much easier to deal with (none of the C<match> hassle would be required - you could just match like a regular RE and it would Just Work). =head1 SEE ALSO =over 4 =item L<perlre> General information about Perl's regular expressions. =item L<re> Specific information about C<use re 'eval'>. =item Regex::PreSuf C<Regex::PreSuf> takes a string and chops it itself into tokens of length 1. Since it can't deal with tokens of more than one character, it can't deal with meta-characters and thus no regular expressions. Which is the main reason why I wrote this module. =item Regexp::Optimizer C<Regexp::Optimizer> produces regular expressions that are similar to those produced by R::A with reductions switched off. It's biggest drawback is that it is exponentially slower than Regexp::Assemble on very large sets of patterns. =item Regexp::Parser Fine grained analysis of regular expressions. =item Regexp::Trie Funnily enough, this was my working name for C<Regexp::Assemble> during its development. I changed the name because I thought it was too obscure. Anyway, C<Regexp::Trie> does much the same as C<Regexp::Optimizer> and C<Regexp::Assemble> except that it runs much faster (according to the author). It does not recognise meta characters (that is, 'a+b' is interpreted as 'a\+b'). =item Text::Trie C<Text::Trie> is well worth investigating. Tries can outperform very bushy (read: many alternations) patterns. =item Tree::Trie C<Tree::Trie> is another module that builds tries. The algorithm that C<Regexp::Assemble> uses appears to be quite similar to the algorithm described therein, except that C<R::A> solves its end-marker problem without having to rewrite the leaves. =back =head1 See Also For alternatives to this module, consider one of: =over 4 =item o L<Data::Munge> =item o L<OnSearch::Regex> =item o L<Regex::PreSuf> =back =head1 LIMITATIONS Some mildly complex cases are not handled well. See examples/failure.01.pl and L<https://rt.cpan.org/Public/Bug/Display.html?id=104897>. See also L<https://rt.cpan.org/Public/Bug/Display.html?id=106480> for a discussion of some of the issues arising with the use of a huge number of alterations. Thanx to Slaven Rezic for the details of trie 'v' non-trie operations within Perl which influence regexp handling of alternations. <Regexp::Assemble> does not attempt to find common substrings. For instance, it will not collapse C</cabababc/> down to C</c(?:ab){3}c/>. If there's a module out there that performs this sort of string analysis I'd like to know about it. But keep in mind that the algorithms that do this are very expensive: quadratic or worse. C<Regexp::Assemble> does not interpret meta-character modifiers. For instance, if the following two patterns are given: C<X\d> and C<X\d+>, it will not determine that C<\d> can be matched by C<\d+>. Instead, it will produce C<X(?:\d|\d+)>. Along a similar line of reasoning, it will not determine that C<Z> and C<Z\d+> is equivalent to C<Z\d*> (It will produce C<Z(?:\d+)?> instead). You cannot remove a pattern that has been added to an object. You'll just have to start over again. Adding a pattern is difficult enough, I'd need a solid argument to convince me to add a C<remove> method. If you need to do this you should read the documentation for the C<clone> method. C<Regexp::Assemble> does not (yet)? employ the C<(?E<gt>...)> construct. The module does not produce POSIX-style regular expressions. This would be quite easy to add, if there was a demand for it. =head1 BUGS Patterns that generate look-ahead assertions sometimes produce incorrect patterns in certain obscure corner cases. If you suspect that this is occurring in your pattern, disable lookaheads. Tracking doesn't really work at all with 5.6.0. It works better in subsequent 5.6 releases. For maximum reliability, the use of a 5.8 release is strongly recommended. Tracking barely works with 5.005_04. Of note, using C<\d>-style meta-characters invariably causes panics. Tracking really comes into its own in Perl 5.10. If you feed C<Regexp::Assemble> patterns with nested parentheses, there is a chance that the resulting pattern will be uncompilable due to mismatched parentheses (not enough closing parentheses). This is normal, so long as the default lexer pattern is used. If you want to find out which pattern among a list of 3000 patterns are to blame (speaking from experience here), the F<eg/debugging> script offers a strategy for pinpointing the pattern at fault. While you may not be able to use the script directly, the general approach is easy to implement. The algorithm used to assemble the regular expressions makes extensive use of mutually-recursive functions (that is, A calls B, B calls A, ...) For deeply similar expressions, it may be possible to provoke "Deep recursion" warnings. The module has been tested extensively, and has an extensive test suite (that achieves close to 100% statement coverage), but you never know... A bug may manifest itself in two ways: creating a pattern that cannot be compiled, such as C<a\(bc)>, or a pattern that compiles correctly but that either matches things it shouldn't, or doesn't match things it should. It is assumed that Such problems will occur when the reduction algorithm encounters some sort of edge case. A temporary work-around is to disable reductions: my $pattern = $assembler->reduce(0)->re; A discussion about implementation details and where bugs might lurk appears in the README file. If this file is not available locally, you should be able to find a copy on the Web at your nearest CPAN mirror. Seriously, though, a number of people have been using this module to create expressions anywhere from 140Kb to 600Kb in size, and it seems to be working according to spec. Thus, I don't think there are any serious bugs remaining. If you are feeling brave, extensive debugging traces are available to figure out where assembly goes wrong. Please report all bugs at L<http://rt.cpan.org/NoAuth/Bugs.html?Dist=Regexp-Assemble> Make sure you include the output from the following two commands: perl -MRegexp::Assemble -le 'print $Regexp::Assemble::VERSION' perl -V There is a mailing list for the discussion of C<Regexp::Assemble>. Subscription details are available at L<http://listes.mongueurs.net/mailman/listinfo/regexp-assemble>. =head1 ACKNOWLEDGEMENTS This module grew out of work I did building access maps for Postfix, a modern SMTP mail transfer agent. See L<http://www.postfix.org/> for more information. I used Perl to build large regular expressions for blocking dynamic/residential IP addresses to cut down on spam and viruses. Once I had the code running for this, it was easy to start adding stuff to block really blatant spam subject lines, bogus HELO strings, spammer mailer-ids and more... I presented the work at the French Perl Workshop in 2004, and the thing most people asked was whether the underlying mechanism for assembling the REs was available as a module. At that time it was nothing more that a twisty maze of scripts, all different. The interest shown indicated that a module was called for. I'd like to thank the people who showed interest. Hey, it's going to make I<my> messy scripts smaller, in any case. Thomas Drugeon was a valuable sounding board for trying out early ideas. Jean Forget and Philippe Blayo looked over an early version. H.Merijn Brandt stopped over in Paris one evening, and discussed things over a few beers. Nicholas Clark pointed out that while what this module does (?:c|sh)ould be done in perl's core, as per the 2004 TODO, he encouraged me to continue with the development of this module. In any event, this module allows one to gauge the difficulty of undertaking the endeavour in C. I'd rather gouge my eyes out with a blunt pencil. Paul Johnson settled the question as to whether this module should live in the Regex:: namespace, or Regexp:: namespace. If you're not convinced, try running the following one-liner: perl -le 'print ref qr//' Philippe Bruhat found a couple of corner cases where this module could produce incorrect results. Such feedback is invaluable, and only improves the module's quality. =head1 Machine-Readable Change Log The file Changes was converted into Changelog.ini by L<Module::Metadata::Changes>. =head1 AUTHOR David Landgren Copyright (C) 2004-2011. All rights reserved. http://www.landgren.net/perl/ If you use this module, I'd love to hear about what you're using it for. If you want to be informed of updates, send me a note. Ron Savage is co-maint of the module, starting with V 0.36. =head1 Repository L<https://github.com/ronsavage/Regexp-Assemble.git> =head1 TODO 1. Tree equivalencies. Currently, /contend/ /content/ /resend/ /resent/ produces (?:conten[dt]|resend[dt]) but it is possible to produce (?:cont|res)en[dt] if one can spot the common tail nodes (and walk back the equivalent paths). Or be by me my => /[bm][ey]/ in the simplest case. To do this requires a certain amount of restructuring of the code. Currently, the algorithm uses a two-phase approach. In the first phase, the trie is traversed and reductions are performed. In the second phase, the reduced trie is traversed and the pattern is emitted. What has to occur is that the reduction and emission have to occur together. As a node is completed, it is replaced by its string representation. This then allows child nodes to be compared for equality with a simple 'eq'. Since there is only a single traversal, the overall generation time might drop, even though the context baggage required to delve through the tree will be more expensive to carry along (a hash rather than a couple of scalars). Actually, a simpler approach is to take on a secret sentinel atom at the end of every pattern, which gives the reduction algorithm sufficient traction to create a perfect trie. I'm rewriting the reduction code using this technique. 2. Investigate how (?>foo) works. Can it be applied? 5. How can a tracked pattern be serialised? (Add freeze and thaw methods). 6. Store callbacks per tracked pattern. 12. utf-8... hmmmm... 14. Adding qr//'ed patterns. For example, consider $r->add ( qr/^abc/i ) ->add( qr/^abd/ ) ->add( qr/^ab e/x ); this should admit abc abC aBc aBC abd abe as matches 16. Allow a fast, unsafe tracking mode, that can be used if a(?bc)? can't happen. (Possibly carp if it does appear during traversal)? 17. given a-\d+-\d+-\d+-\d+-b, produce a(?:-\d+){4}-b. Something along the lines of (.{4))(\1+) would let the regexp engine itself be brought to bear on the matter, which is a rather appealing idea. Consider while(/(?!\+)(\S{2,}?)(\1+)/g) { ... $1, $2 ... } as a starting point. 19. The reduction code has become unbelievably baroque. Adding code to handle (sting,singing,sing) => s(?:(?:ing)?|t)ing was far too difficult. Adding more stuff just breaks existing behaviour. And fixing the ^abcd$ ... bug broke stuff all over again. Now that the corner cases are more clearly identified, a full rewrite of the reduction code is needed. And would admit the possibility of implementing items 1 and 17. 20. Handle debug unrev with a separate bit 23. Japhy's http://www.perlmonks.org/index.pl?node_id=90876 list2range regexp 24. Lookahead assertions contain serious bugs (as shown by assembling powersets. Need to save more context during reduction, which in turn will simplify the preparation of the lookahead classes. See also 19. 26. _lex() swamps the overall run-time. It stems from the decision to use a single regexp to pull apart any pattern. A suite of simpler regexp to pick of parens, char classes, quantifiers and bare tokens may be faster. (This has been implemented as _fastlex(), but it's only marginally faster. Perhaps split-by- char and lex a la C? 27. We don't, as yet, unroll_plus a paren e.g. (abc)+? 28. We don't reroll unrolled a a* to a+ in indented or tracked output 29. Use (*MARK n) in blead for tracked patterns, and use (*FAIL) for the unmatchable pattern. =head1 LICENSE This library is free software; you can redistribute it and/or modify it under the same terms as Perl itself. =cut # Return a +ve value to tell Perl the module is ready to go. 'The Lusty Decadent Delights of Imperial Pompeii';