package Bio::DB::SeqFeature::Store::GFF3Loader; # $Id: GFF3Loader.pm 16168 2009-09-25 21:07:32Z cjfields $ =head1 NAME Bio::DB::SeqFeature::Store::GFF3Loader -- GFF3 file loader for Bio::DB::SeqFeature::Store =head1 SYNOPSIS use Bio::DB::SeqFeature::Store; use Bio::DB::SeqFeature::Store::GFF3Loader; # Open the sequence database my $db = Bio::DB::SeqFeature::Store->new( -adaptor => 'DBI::mysql', -dsn => 'dbi:mysql:test', -write => 1 ); my $loader = Bio::DB::SeqFeature::Store::GFF3Loader->new(-store => $db, -verbose => 1, -fast => 1); $loader->load('./my_genome.gff3'); =head1 DESCRIPTION The Bio::DB::SeqFeature::Store::GFF3Loader object parsers GFF3-format sequence annotation files and loads Bio::DB::SeqFeature::Store databases. For certain combinations of SeqFeature classes and SeqFeature::Store databases it features a "fast load" mode which will greatly accelerate the loading of GFF3 databases by a factor of 5-10. The GFF3 file format has been extended very slightly to accomodate Bio::DB::SeqFeature::Store. First, the loader recognizes is a new directive: # #index-subfeatures [0|1] Note that you can place a space between the two #'s in order to prevent GFF3 validators from complaining. If this is true, then subfeatures are indexed (the default) so that they can be retrieved with a query. See L for an explanation of this. If false, then subfeatures can only be accessed through their parent feature. Second, the loader recognizes a new attribute tag called index, which if present, controls indexing of the current feature. Example: ctg123 . TF_binding_site 1000 1012 . + . ID=tfbs00001;index=1 You can use this to turn indexing on and off, overriding the default for a particular feature. Note that the loader keeps a record -- in memory -- of each feature that it has processed. If you find the loader running out of memory on particularly large GFF3 files, please split the input file into smaller pieces and do the load in steps. =cut # load utility - incrementally load the store based on GFF3 file # # two modes: # slow mode -- features can occur in any order in the GFF3 file # fast mode -- all features with same ID must be contiguous in GFF3 file use strict; use Carp 'croak'; use Bio::DB::GFF::Util::Rearrange; use Bio::DB::SeqFeature::Store::LoadHelper; use base 'Bio::DB::SeqFeature::Store::Loader'; my %Special_attributes =( Gap => 1, Target => 1, Parent => 1, Name => 1, Alias => 1, ID => 1, index => 1, Index => 1, ); my %Strandedness = ( '+' => 1, '-' => -1, '.' => 0, '' => 0, 0 => 0, 1 => 1, -1 => -1, +1 => 1, undef => 0, ); =head2 new Title : new Usage : $loader = Bio::DB::SeqFeature::Store::GFF3Loader->new(@options) Function: create a new parser Returns : a Bio::DB::SeqFeature::Store::GFF3Loader gff3 parser and loader Args : several - see below Status : public This method creates a new GFF3 loader and establishes its connection with a Bio::DB::SeqFeature::Store database. Arguments are -name=E$value pairs as described in this table: Name Value ---- ----- -store A writeable Bio::DB::SeqFeature::Store database handle. -seqfeature_class The name of the type of Bio::SeqFeatureI object to create and store in the database (Bio::DB::SeqFeature by default) -sf_class A shorter alias for -seqfeature_class -verbose Send progress information to standard error. -fast If true, activate fast loading (see below) -chunk_size Set the storage chunk size for nucleotide/protein sequences (default 2000 bytes) -tmp Indicate a temporary directory to use when loading non-normalized features. -ignore_seqregion Ignore ##sequence-region directives. The default is to create a feature corresponding to the directive. When you call new(), a connection to a Bio::DB::SeqFeature::Store database should already have been established and the database initialized (if appropriate). Some combinations of Bio::SeqFeatures and Bio::DB::SeqFeature::Store databases support a fast loading mode. Currently the only reliable implementation of fast loading is the combination of DBI::mysql with Bio::DB::SeqFeature. The other important restriction on fast loading is the requirement that a feature that contains subfeatures must occur in the GFF3 file before any of its subfeatures. Otherwise the subfeatures that occurred before the parent feature will not be attached to the parent correctly. This restriction does not apply to normal (slow) loading. If you use an unnormalized feature class, such as Bio::SeqFeature::Generic, then the loader needs to create a temporary database in which to cache features until all their parts and subparts have been seen. This temporary databases uses the "berkeleydb" adaptor. The -tmp option specifies the directory in which that database will be created. If not present, it defaults to the system default tmp directory specified by File::Spec-Etmpdir(). The -chunk_size option allows you to tune the representation of DNA/Protein sequence in the Store database. By default, sequences are split into 2000 base/residue chunks and then reassembled as needed. This avoids the problem of pulling a whole chromosome into memory in order to fetch a short subsequence from somewhere in the middle. Depending on your usage patterns, you may wish to tune this parameter using a chunk size that is larger or smaller than the default. =cut sub new { my $class = shift; my $self = $class->SUPER::new(@_); my ($ignore_seqregion) = rearrange(['IGNORE_SEQREGION'],@_); $self->ignore_seqregion($ignore_seqregion); $self; } =head2 ignore_seqregion $ignore_it = $loader->ignore_seqregion([$new_flag]) Get or set the ignore_seqregion flag, which if true, will cause GFF3 ##sequence-region directives to be ignored. The default behavior is to create a feature corresponding to the region. =cut sub ignore_seqregion { my $self = shift; my $d = $self->{ignore_seqregion}; $self->{ignore_seqregion} = shift if @_; $d; } =head2 load Title : load Usage : $count = $loader->load(@ARGV) Function: load the indicated files or filehandles Returns : number of feature lines loaded Args : list of files or filehandles Status : public Once the loader is created, invoke its load() method with a list of GFF3 or FASTA file paths or previously-opened filehandles in order to load them into the database. Compressed files ending with .gz, .Z and .bz2 are automatically recognized and uncompressed on the fly. Paths beginning with http: or ftp: are treated as URLs and opened using the LWP GET program (which must be on your path). FASTA files are recognized by their initial "E" character. Do not feed the loader a file that is neither GFF3 nor FASTA; I don't know what will happen, but it will probably not be what you expect. =cut # sub load { } inherited =head2 accessors The following read-only accessors return values passed or created during new(): store() the long-term Bio::DB::SeqFeature::Store object tmp_store() the temporary Bio::DB::SeqFeature::Store object used during loading sfclass() the Bio::SeqFeatureI class fast() whether fast loading is active seq_chunk_size() the sequence chunk size verbose() verbose progress messages =cut # sub store inherited # sub tmp_store inherited # sub sfclass inherited # sub fast inherited # sub seq_chunk_size inherited # sub verbose inherited =head2 Internal Methods The following methods are used internally and may be overidden by subclasses. =over 4 =item default_seqfeature_class $class = $loader->default_seqfeature_class Return the default SeqFeatureI class (Bio::DB::SeqFeature). =cut # sub default_seqfeature_class { } inherited =item subfeatures_normalized $flag = $loader->subfeatures_normalized([$new_flag]) Get or set a flag that indicates that the subfeatures are normalized. This is deduced from the SeqFeature class information. =cut # sub subfeatures_normalized { } inherited =item subfeatures_in_table $flag = $loader->subfeatures_in_table([$new_flag]) Get or set a flag that indicates that feature/subfeature relationships are stored in a table. This is deduced from the SeqFeature class and Store information. =cut # sub subfeatures_in_table { } inherited =item load_fh $count = $loader->load_fh($filehandle) Load the GFF3 data at the other end of the filehandle and return true if successful. Internally, load_fh() invokes: start_load(); do_load($filehandle); finish_load(); =cut # sub load_fh { } inherited =item start_load, finish_load These methods are called at the start and end of a filehandle load. =cut sub create_load_data { #overridden my $self = shift; $self->SUPER::create_load_data; $self->{load_data}{TemporaryID} = "GFFLoad0000000"; $self->{load_data}{IndexSubfeatures} = $self->index_subfeatures(); $self->{load_data}{mode} = 'gff'; $self->{load_data}{Helper} = Bio::DB::SeqFeature::Store::LoadHelper->new($self->{tmpdir}); } sub finish_load { #overridden my $self = shift; $self->store_current_feature(); # during fast loading, we will have a feature left at the very end $self->start_or_finish_sequence(); # finish any half-loaded sequences $self->msg("Building object tree..."); my $start = $self->time(); $self->build_object_tree; $self->msg(sprintf "%5.2fs\n",$self->time()-$start); if ($self->fast) { $self->msg("Loading bulk data into database..."); $start = $self->time(); $self->store->finish_bulk_update; $self->msg(sprintf "%5.2fs\n",$self->time()-$start); } eval {$self->store->commit}; # don't delete load data so that caller can ask for the loaded IDs # $self->delete_load_data; } =item do_load $count = $loader->do_load($fh) This is called by load_fh() to load the GFF3 file's filehandle and return the number of lines loaded. =cut # sub do_load { } inherited =item load_line $loader->load_line($data); Load a line of a GFF3 file. You must bracket this with calls to start_load() and finish_load()! $loader->start_load(); $loader->load_line($_) while ; $loader->finish_load(); =cut sub load_line { #overridden my $self = shift; my $line = shift; chomp($line); my $load_data = $self->{load_data}; $load_data->{line}++; return unless $line =~ /^\S/; # blank line $load_data->{mode} = 'gff' if /\t/; # if it has a tab in it, switch to gff mode if ($line =~ /^\#\s?\#\s*(.+)/) { ## meta instruction $load_data->{mode} = 'gff'; $self->handle_meta($1); } elsif ($line =~ /^\#/) { $load_data->{mode} = 'gff'; # just to be safe return; # comment } elsif ($line =~ /^>\s*(\S+)/) { # FASTA lines are coming $load_data->{mode} = 'fasta'; $self->start_or_finish_sequence($1); } elsif ($load_data->{mode} eq 'fasta') { $self->load_sequence($line); } elsif ($load_data->{mode} eq 'gff') { $self->handle_feature($line); if (++$load_data->{count} % 1000 == 0) { my $now = $self->time(); my $nl = -t STDOUT && !$ENV{EMACS} ? "\r" : "\n"; local $^W = 0; # kill uninit variable warning $self->msg(sprintf("%d features loaded in %5.2fs (%5.2fs/1000 features)...%s$nl", $load_data->{count},$now - $load_data->{start_time}, $now - $load_data->{millenium_time}, ' ' x 80 )); $load_data->{millenium_time} = $now; } } else { $self->throw("I don't know what to do with this line:\n$line"); } } =item handle_meta $loader->handle_meta($meta_directive) This method is called to handle meta-directives such as ##sequence-region. The method will receive the directive with the initial ## stripped off. =cut sub handle_meta { my $self = shift; my $instruction = shift; if ( $instruction =~ /^#$/ ) { $self->store_current_feature() ; # during fast loading, we will have a feature left at the very end $self->start_or_finish_sequence(); # finish any half-loaded sequences if ( $self->store->can('handle_resolution_meta') ) { $self->store->handle_resolution_meta($instruction); } return; } if ($instruction =~ /sequence-region\s+(.+)\s+(-?\d+)\s+(-?\d+)/i && !$self->ignore_seqregion()) { my($ref,$start,$end,$strand) = $self->_remap($1,$2,$3,+1); my $feature = $self->sfclass->new(-name => $ref, -seq_id => $ref, -start => $start, -end => $end, -strand => $strand, -primary_tag => 'region'); $self->store->store($feature); return; } if ($instruction =~/index-subfeatures\s+(\S+)/i) { $self->{load_data}{IndexSubfeatures} = $1; $self->store->index_subfeatures($1); return; } if ( $self->store->can('handle_unrecognized_meta') ) { $self->store->handle_unrecognized_meta($instruction); return; } } =item handle_feature $loader->handle_feature($gff3_line) This method is called to process a single GFF3 line. It manipulates information stored a data structure called $self-E{load_data}. =cut sub handle_feature { #overridden my $self = shift; my $gff_line = shift; my $ld = $self->{load_data}; my $allow_whitespace = $self->allow_whitespace; $gff_line =~ s/\s+/\t/g if $allow_whitespace; my @columns = map {$_ eq '.' ? undef : $_ } split /\t/,$gff_line; $self->invalid_gff($gff_line) if @columns < 4; $self->invalid_gff($gff_line) if @columns > 9 && $allow_whitespace; { local $^W = 0; if (@columns > 9) { #oops, split too much due to whitespace $columns[8] = join(' ',@columns[8..$#columns]); } } my ($refname,$source,$method,$start,$end,$score,$strand,$phase,$attributes) = @columns; $self->invalid_gff($gff_line) unless defined $refname; $self->invalid_gff($gff_line) unless $start eq '.' || $start =~ /^[\d.-]+$/; $self->invalid_gff($gff_line) unless $end eq '.' || $end =~ /^[\d.-]+$/; $self->invalid_gff($gff_line) unless defined $method; $strand = $Strandedness{$strand||0}; my ($reserved,$unreserved) = $attributes ? $self->parse_attributes($attributes) : (); my $name = ($reserved->{Name} && $reserved->{Name}[0]); my $has_loadid = defined $reserved->{ID}[0]; my $feature_id = defined $reserved->{ID}[0] ? $reserved->{ID}[0] : $ld->{TemporaryID}++; my @parent_ids = @{$reserved->{Parent}} if defined $reserved->{Parent}; my $index_it = $ld->{IndexSubfeatures}; if (exists $reserved->{Index} || exists $reserved->{index}) { $index_it = $reserved->{Index}[0] || $reserved->{index}[0]; } # Everything in the unreserved hash becomes an attribute, so we copy # some attributes over $unreserved->{Note} = $reserved->{Note} if exists $reserved->{Note}; $unreserved->{Alias} = $reserved->{Alias} if exists $reserved->{Alias}; $unreserved->{Target} = $reserved->{Target} if exists $reserved->{Target}; $unreserved->{Gap} = $reserved->{Gap} if exists $reserved->{Gap}; $unreserved->{load_id}= $reserved->{ID} if exists $reserved->{ID}; # mec@stowers-institute.org, wondering why not all attributes are # carried forward, adds ID tag in particular service of # round-tripping ID, which, though present in database as load_id # attribute, was getting lost as itself # $unreserved->{ID}= $reserved->{ID} if exists $reserved->{ID}; # TEMPORARY HACKS TO SIMPLIFY DEBUGGING $feature_id = '' unless defined $feature_id; $name = '' unless defined $name; # prevent uninit variable warnings # push @{$unreserved->{Alias}},$feature_id if $has_loadid && $feature_id ne $name; $unreserved->{parent_id} = \@parent_ids if @parent_ids; # POSSIBLY A PERMANENT HACK -- TARGETS BECOME ALIASES # THIS IS TO ALLOW FOR TARGET-BASED LOOKUPS if (exists $reserved->{Target}) { my %aliases = map {$_=>1} @{$unreserved->{Alias}}; for my $t (@{$reserved->{Target}}) { (my $tc = $t) =~ s/\s+.*$//; # get rid of coordinates $name ||= $tc; push @{$unreserved->{Alias}},$tc unless $name eq $tc || $aliases{$tc}; } } ($refname,$start,$end,$strand) = $self->_remap($refname,$start,$end,$strand) or return; my @args = (-display_name => $name, -seq_id => $refname, -start => $start, -end => $end, -strand => $strand || 0, -score => $score, -phase => $phase, -primary_tag => $method || 'feature', -source => $source, -tag => $unreserved, -attributes => $unreserved, ); # Here's where we handle feature lines that have the same ID (multiple locations, not # parent/child relationships) my $old_feat; # Current feature is the same as the previous feature, which hasn't yet been loaded if (defined $ld->{CurrentID} && $ld->{CurrentID} eq $feature_id) { $old_feat = $ld->{CurrentFeature}; } # Current feature is the same as a feature that was loaded earlier elsif (defined(my $id = $self->{load_data}{Helper}->local2global($feature_id))) { $old_feat = $self->fetch($feature_id) or $self->warn(<{TemporaryID}++; # AND they have a Parent attribute, this causes an undesirable } # additional layer of aggregation. Changing the ID fixes this. elsif ( $old_feat->seq_id ne $refname || $old_feat->start != $start || $old_feat->end != $end # make sure endpoints are distinct ) { $self->add_segment($old_feat,$self->sfclass->new(@args)); return; } } # we get here if this is a new feature # first of all, store the current feature if it is there $self->store_current_feature() if defined $ld->{CurrentID}; # now create the new feature # (index top-level features only if policy asks us to) my $feature = $self->sfclass->new(@args); $feature->object_store($self->store) if $feature->can('object_store'); # for lazy table features $ld->{CurrentFeature} = $feature; $ld->{CurrentID} = $feature_id; my $top_level = !@parent_ids; my $has_id = defined $reserved->{ID}[0]; $index_it ||= $top_level; my $helper = $ld->{Helper}; $helper->indexit($feature_id=>1) if $index_it; $helper->toplevel($feature_id=>1) if !$self->{fast} && $top_level; # need to track top level features # remember parentage for my $parent (@parent_ids) { $helper->add_children($parent=>$feature_id); } } sub invalid_gff { my $self = shift; my $line = shift; $self->throw("invalid GFF line at line $self->{load_data}{line}.\n".$line); } =item allow_whitespace $allow_it = $loader->allow_whitespace([$newvalue]); Get or set the allow_whitespace flag. If true, then GFF3 files are allowed to be delimited with whitespace in addition to tabs. =cut sub allow_whitespace { my $self = shift; my $d = $self->{allow_whitespace}; $self->{allow_whitespace} = shift if @_; $d; } =item store_current_feature $loader->store_current_feature() This method is called to store the currently active feature in the database. It uses a data structure stored in $self-E{load_data}. =cut # sub store_current_feature { } inherited =item build_object_tree $loader->build_object_tree() This method gathers together features and subfeatures and builds the graph that connects them. =cut ### # put objects together # sub build_object_tree { my $self = shift; $self->subfeatures_in_table ? $self->build_object_tree_in_tables : $self->build_object_tree_in_features; } =item build_object_tree_in_tables $loader->build_object_tree_in_tables() This method gathers together features and subfeatures and builds the graph that connects them, assuming that parent/child relationships will be stored in a database table. =cut sub build_object_tree_in_tables { my $self = shift; my $store = $self->store; my $helper = $self->{load_data}{Helper}; while (my ($load_id,$children) = $helper->each_family()) { my $parent_id = $helper->local2global($load_id); die $self->throw("$load_id doesn't have a primary id") unless defined $parent_id; my @children = map {$helper->local2global($_)} @$children; # this updates the table that keeps track of parent/child relationships, # but does not update the parent object -- so (start,end) had better be right!!! $store->add_SeqFeature($parent_id,@children); } } =item build_object_tree_in_features $loader->build_object_tree_in_features() This method gathers together features and subfeatures and builds the graph that connects them, assuming that parent/child relationships are stored in the seqfeature objects themselves. =cut sub build_object_tree_in_features { my $self = shift; my $store = $self->store; my $tmp = $self->tmp_store; my $ld = $self->{load_data}; my $normalized = $self->subfeatures_normalized; my $helper = $ld->{Helper}; while (my $load_id = $helper->each_toplevel) { my $feature = $self->fetch($load_id) or $self->throw("$load_id (id=" .$helper->local2global($load_id) ." should have a database entry, but doesn't"); $self->attach_children($store,$ld,$load_id,$feature); # Indexed objects are updated, not created anew $feature->primary_id(undef) unless $helper->indexit($load_id); $store->store($feature); } } =item attach_children $loader->attach_children($store,$load_data,$load_id,$feature) This recursively adds children to features and their subfeatures. It is called when subfeatures are directly contained within other features, rather than stored in a relational table. =cut sub attach_children { my $self = shift; my ($store,$ld,$load_id,$feature) = @_; my $children = $ld->{Helper}->children() or return; for my $child_id (@$children) { my $child = $self->fetch($child_id) or $self->throw("$child_id should have a database entry, but doesn't"); $self->attach_children($store,$ld,$child_id,$child); # recursive call $feature->add_SeqFeature($child); } } =item fetch my $feature = $loader->fetch($load_id) Given a load ID (from the ID= attribute) this method returns the feature from the temporary database or the permanent one, depending on where it is stored. =cut sub fetch { my $self = shift; my $load_id = shift; my $helper = $self->{load_data}{Helper}; my $id = $helper->local2global($load_id); return ($self->subfeatures_normalized || $helper->indexit($load_id) ? $self->store->fetch($id) : $self->tmp_store->fetch($id) ); } =item add_segment $loader->add_segment($parent,$child) This method is used to add a split location to the parent. =cut sub add_segment { my $self = shift; my ($parent,$child) = @_; if ($parent->can('add_segment')) { # probably a lazy table feature my $segment_count = $parent->can('denormalized_segment_count') ? $parent->denormalized_segment_count : $parent->can('denormalized_segments ') ? $parent->denormalized_segments : $parent->can('segments') ? $parent->segments : 0; unless ($segment_count) { # convert into a segmented object my $segment; if ($parent->can('clone')) { $segment = $parent->clone; } else { my %clone = %$parent; $segment = bless \%clone,ref $parent; } delete $segment->{segments}; eval {$segment->object_store(undef) }; $segment->primary_id(undef); # this updates the object and expands its start and end positions without writing # the segments into the database as individual objects $parent->add_segment($segment); } $parent->add_segment($child); 1; # for debugging } # a conventional Bio::SeqFeature::Generic object - create a split location else { my $current_location = $parent->location; if ($current_location->can('add_sub_Location')) { $current_location->add_sub_Location($child->location); } else { eval "require Bio::Location::Split" unless Bio::Location::Split->can('add_sub_Location'); my $new_location = Bio::Location::Split->new(); $new_location->add_sub_Location($current_location); $new_location->add_sub_Location($child->location); $parent->location($new_location); } } } =item parse_attributes ($reserved,$unreserved) = $loader->parse_attributes($attribute_line) This method parses the information contained in the $attribute_line into two hashrefs, one containing the values of reserved attribute tags (e.g. ID) and the other containing the values of unreserved ones. =cut sub parse_attributes { my $self = shift; my $att = shift; unless ($att =~ /=/) { # ouch! must be a GFF line require Bio::DB::SeqFeature::Store::GFF2Loader unless Bio::DB::SeqFeature::Store::GFF2Loader->can('parse_attributes'); return $self->Bio::DB::SeqFeature::Store::GFF2Loader::parse_attributes($att); } my @pairs = map { my ($name,$value) = split '='; [$self->unescape($name) => $value]; } split ';',$att; my (%reserved,%unreserved); foreach (@pairs) { my $tag = $_->[0]; unless (defined $_->[1]) { warn "$tag does not have a value at GFF3 file line $.\n"; next; } my @values = split ',',$_->[1]; map {$_ = $self->unescape($_);} @values; if ($Special_attributes{$tag}) { # reserved attribute push @{$reserved{$tag}},@values; } else { push @{$unreserved{$tag}},@values } } return (\%reserved,\%unreserved); } =item start_or_finish_sequence $loader->start_or_finish_sequence('Chr9') This method is called at the beginning and end of a fasta section. =cut # sub start_or_finish_sequence { } inherited =item load_sequence $loader->load_sequence('gatttcccaaa') This method is called to load some amount of sequence after start_or_finish_sequence() is first called. =cut # sub load_sequence { } inherited =item open_fh my $io_file = $loader->open_fh($filehandle_or_path) This method opens up the indicated file or pipe, using some intelligence to recognized compressed files and URLs and doing the right thing. =cut # sub open_fh { } inherited # sub msg { } inherited =item time my $time = $loader->time This method returns the current time in seconds, using Time::HiRes if available. =cut # sub time { } inherited =item unescape my $unescaped = GFF3Loader::unescape($escaped) This is an internal utility. It is the same as CGI::Util::unescape, but doesn't change pluses into spaces and ignores unicode escapes. =cut # sub unescape { } inherited sub _remap { my $self = shift; my ($ref,$start,$end,$strand) = @_; my $mapper = $self->coordinate_mapper; return ($ref,$start,$end,$strand) unless $mapper; my ($newref,$coords) = $mapper->($ref,[$start,$end]); return unless defined $coords->[0]; if ($coords->[0] > $coords->[1]) { @{$coords} = reverse(@{$coords}); $strand *= -1; } return ($newref,@{$coords},$strand); } sub _indexit { # override my $self = shift; return $self->{load_data}{Helper}->indexit(@_); } sub _local2global { # override my $self = shift; return $self->{load_data}{Helper}->local2global(@_); } =item local_ids my $ids = $self->local_ids; my $id_cnt = @$ids; After performing a load, this returns an array ref containing all the load file IDs that were contained within the file just loaded. =cut sub local_ids { # override my $self = shift; return $self->{load_data}{Helper}->local_ids(@_); } =item loaded_ids my $ids = $loader->loaded_ids; my $id_cnt = @$ids; After performing a load, this returns an array ref containing all the feature primary ids that were created during the load. =cut sub loaded_ids { # override my $self = shift; return $self->{load_data}{Helper}->loaded_ids(@_); } 1; __END__ =back =head1 BUGS This is an early version, so there are certainly some bugs. Please use the BioPerl bug tracking system to report bugs. =head1 SEE ALSO L, L, L, L, L, L =head1 AUTHOR Lincoln Stein Elstein@cshl.orgE. Copyright (c) 2006 Cold Spring Harbor Laboratory. This library is free software; you can redistribute it and/or modify it under the same terms as Perl itself. =cut