1BP_BULK_LOAD_GFF(1) User Contributed Perl Documentation BP_BULK_LOAD_GFF(1)
2
3
4
6 bulk_load_gff.pl - Bulk-load a Bio::DB::GFF database from GFF files.
7
9 % bulk_load_gff.pl -d testdb dna1.fa dna2.fa features1.gff features2.gff ...
10
12 This script loads a Bio::DB::GFF database with the features contained
13 in a list of GFF files and/or FASTA sequence files. You must use the
14 exact variant of GFF described in Bio::DB::GFF. Various command-line
15 options allow you to control which database to load and whether to
16 allow an existing database to be overwritten.
17
18 This script differs from bp_load_gff.pl in that it is hard-coded to use
19 MySQL and cannot perform incremental loads. See bp_load_gff.pl for an
20 incremental loader that works with all databases supported by
21 Bio::DB::GFF, and bp_fast_load_gff.pl for a MySQL loader that supports
22 fast incremental loads.
23
24 NOTES
25 If the filename is given as "-" then the input is taken from standard
26 input. Compressed files (.gz, .Z, .bz2) are automatically uncompressed.
27
28 FASTA format files are distinguished from GFF files by their filename
29 extensions. Files ending in .fa, .fasta, .fast, .seq, .dna and their
30 uppercase variants are treated as FASTA files. Everything else is
31 treated as a GFF file. If you wish to load -fasta files from STDIN,
32 then use the -f command-line swith with an argument of '-', as in
33
34 gunzip my_data.fa.gz | bp_fast_load_gff.pl -d test -f -
35
36 The nature of the bulk load requires that the database be on the local
37 machine and that the indicated user have the "file" privilege to load
38 the tables and have enough room in /usr/tmp (or whatever is specified
39 by the \$TMPDIR environment variable), to hold the tables transiently.
40
41 Local data may now be uploaded to a remote server via the --local
42 option with the database host specified in the dsn, e.g.
43 dbi:mysql:test:db_host
44
45 The adaptor used is dbi::mysqlopt. There is currently no way to change
46 this.
47
48 About maxfeature: the default value is 100,000,000 bases. If you have
49 features that are close to or greater that 100Mb in length, then the
50 value of maxfeature should be increased to 1,000,000,000. This value
51 must be a power of 10.
52
53 Note that Windows users must use the --create option.
54
55 If the list of GFF or fasta files exceeds the kernel limit for the
56 maximum number of command-line arguments, use the --long_list
57 /path/to/files option.
58
60 Command-line options can be abbreviated to single-letter options. e.g.
61 -d instead of --database.
62
63 --database <dsn> Database name (default dbi:mysql:test)
64 --adaptor Adaptor name (default mysql)
65 --create Reinitialize/create data tables without asking
66 --user Username to log in as
67 --fasta File or directory containing fasta files to load
68 --long_list Directory containing a very large number of
69 GFF and/or FASTA files
70 --password Password to use for authentication
71 (Does not work with Postgres, password must be
72 supplied interactively or be left empty for
73 ident authentication)
74 --maxbin Set the value of the maximum bin size
75 --local Flag to indicate that the data source is local
76 --maxfeature Set the value of the maximum feature size (power of 10)
77 --group A list of one or more tag names (comma or space separated)
78 to be used for grouping in the 9th column.
79 --gff3_munge Activate GFF3 name munging (see Bio::DB::GFF)
80 --Temporary Location of a writable scratch directory
81
83 Bio::DB::GFF, fast_load_gff.pl, load_gff.pl
84
86 Lincoln Stein, lstein@cshl.org
87
88 Copyright (c) 2002 Cold Spring Harbor Laboratory
89
90 This library is free software; you can redistribute it and/or modify it
91 under the same terms as Perl itself. See DISCLAIMER.txt for
92 disclaimers of warranty.
93
94
95
96perl v5.12.0 2010-04-29 BP_BULK_LOAD_GFF(1)