-
Notifications
You must be signed in to change notification settings - Fork 5
Expand file tree
/
Copy pathinterproscan.properties
More file actions
365 lines (288 loc) · 11.8 KB
/
interproscan.properties
File metadata and controls
365 lines (288 loc) · 11.8 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
# This is the InterProScan configuration file
##
## Temporary files and directory
##
# The text [UNIQUE], if present, will be replaced by a value unique to your running instance
# Temporary files used by the analyses will be placed in directories here:
temporary.file.directory.suffix=[UNIQUE]
temporary.file.directory=temp/${temporary.file.directory.suffix}
##
## H2 database
##
# The H2 database is copied by the standalone version of interproscan
i5.h2.database.original.location=work/template/interpro.zip
# LOCK_TIMEOUT: Sets the lock timeout (in milliseconds) for the current session
i5.database.connection.url=jdbc:h2:mem:interpro;LOCK_TIMEOUT=10000000
##
## binary paths
##
# Configure the version of perl to use when running member databases perl binaries
perl.command=perl
# Binary file locations
#rpsblast
binary.rpsblast.path=bin/blast/ncbi-blast-2.3.0+/rpsblast
#rpsbproc
binary.rpsbproc.path=bin/blast/ncbi-blast-2.3.0+/rpsbproc
#hmmer 3
binary.hmmer3.path=bin/hmmer/hmmer3/3.1b1
binary.hmmer3.hmmscan.path=bin/hmmer/hmmer3/3.1b1/hmmscan
binary.hmmer3.hmmsearch.path=bin/hmmer/hmmer3/3.1b1/hmmsearch
#hmmer 2
binary.hmmer2.hmmsearch.path=bin/hmmer/hmmer2/2.3.2/hmmsearch
binary.hmmer2.hmmpfam.path=bin/hmmer/hmmer2/2.3.2/hmmpfam
binary.fingerprintscan.path=bin/prints/fingerPRINTScan
binary.coils.path=bin/ncoils/2.2.1/ncoils
domainfinder3.path=bin/gene3d/DomainFinder3
binary.prodom.2006.1.prodomblast3i.pl.path=bin/prodom/2006.1/ProDomBlast3i.pl
# Note: Correct prosite binary distribution for your platform can be downloaded: ftp://ftp.expasy.org/databases/prosite/ps_scan/
binary.prosite.psscan.pl.path=bin/prosite/ps_scan.pl
binary.prosite.pfscan.path=bin/prosite/pfscan
binary.prosite.pfsearch.path=bin/prosite/pfsearch
binary.prosite.python.path=
binary.prosite.pfsearch.wrapperpath=bin/prosite/pfsearch_wrapper.py
#panther
binary.panther.path=bin/panther/pantherScore/2.1/pantherScore.pl
binary.panther.perl.lib.dir=bin/panther/pantherScore/2.0/lib
binary.panther.blast.path=bin/blast/2.2.24/bin/blastall
binary.superfamily.1.75.ass3.pl.path=bin/superfamily/1.75/ass3_single_threaded.pl
binary.pirsf.pl.path=bin/pirsf/3.01/pirsf.pl
binary.blastall.2.2.6.path=bin/blast/2.2.6/blastall
binary.blast.2.2.19.path=bin/blast/2.2.19
binary.getorf.path=bin/nucleotide/getorf
#SFLD
sfld.postprocess.command=bin/sfld/sfld_postprocess
#signalp
# Note: SignalP binary not distributed with InterProScan 5, please install separately e.g. in bin/signalp/4.1/signalp
signalp_euk.signature.library.release=4.1
signalp_gram_positive.signature.library.release=4.1
signalp_gram_negative.signature.library.release=4.1
binary.signalp.path=bin/signalp/4.1/signalp
signalp.perl.library.dir=bin/signalp/4.1/lib
#TMHMM 2.0
tmhmm.signature.library.release=2.0c
# Note: TMHMM binary not distributed with InterProScan 5, please install separately e.g. in bin/tmhmm/2.0c/decodeanhmm
binary.tmhmm.path=bin/tmhmm/2.0c/decodeanhmm
#PHOBIUS
# Note: Phobius binary not distributed with InterProScan 5, please install separately e.g. in bin/phobius/1.01/phobius.pl
phobius.signature.library.release=1.01
binary.phobius.pl.path=bin/phobius/1.01/phobius.pl
#
##
## Member database model / data file locations (alphabetically sorted)
##
#CDD
cdd.signature.library.release=3.14
cdd.signature.list.path=data/cdd/cddid.tbl
cdd.library.path=data/cdd/3.14/db/Cdd_NCBI
cdd.data.path=data/cdd/3.14/data
cdd.rpsbproc.data.config.path=data/cdd/3.14/rpsbproc.ini
# Gene3D
gene3d.signature.library.release=3.5.0
gene3d.hmm.path=data/gene3d/3.5.0/gene3d_classified.hmm
gene3d.model2sf_map.path=data/gene3d/3.5.0/model_to_family_map.csv
#gene3d.signature.library.release=4.1.0
#gene3d.hmm.path=data/gene3d/4.1.0/cath-gene3d-hmm3.lib
#gene3d.model2sf_map.path=data/gene3d/4.1.0/cath-gene3d-names.txt
# HAMAP
hamap.profile.models.path=data/hamap/201605.11/hamap.prf
hamap.profile.models.dir=data/hamap/201605.11/profiles
hamap.hmm.lib.path=data/hamap/201605.11/hamap.hmm.lib
hamap.signature.library.release=201605.11
#MobiDB
binary.mobidb.path=bin/mobidb/mobidb-lite.py
binary.mobidb.binx.path=bin/mobidb/binx
mobidb.signature.library.release=1.0
#PANTHER
panther.temporary.file.directory=
panther.signature.library.release=11.1
panther.models.dir=data/panther/11.1/
panther.hmm.path=data/panther/11.1/panther.hmm
# PFam
pfam-a.signature.library.release=30.0
pfam-a.hmm.path=data/pfam/30.0/pfam_a.hmm
pfam-a.seed.path=data/pfam/30.0/pfam_a.seed
pfam-clans.path=data/pfam/30.0/pfam_clans
##
#PIRSF 3.01
pirsf.sfhmm.path.3.01=data/pirsf/3.01/sf_hmm_all
pirsf.dat.path.3.01=data/pirsf/3.01/pirsf.dat
#PRINTS 42.0
prints.kdat.path.42.0=data/prints/42.0/prints42_0.kdat
prints.pval.path.42.0=data/prints/42.0/prints.pval
prints.hierarchy.path.42.0=data/prints/42.0/FingerPRINTShierarchy.db
#ProDom 2006.1
prodom.ipr.path.2006.1=data/prodom/2006.1/prodom.ipr
#Prosite
prosite.patterns.signature.library.release=20.119
prosite.profiles.signature.library.release=20.119
prosite.models.path=data/prosite/20.119/prosite.dat
prosite.evaluator.models.path=data/prosite/20.119/evaluator.dat
#
#SFLD
sfld.signature.library.release=2
sfld.hmm.path=data/sfld/2/sfld.hmm
sfld.sites.annotation.file.path=data/sfld/2/sfld_sites.annot
#
# SignalP 4.0
signalp.4.0.perl.library.dir=bin/signalp/4.0/lib
#smart 7.1
smart.signature.library.release=7.1
smart.hmm.path=data/smart/7.1/smart.HMMs
smart.hmm.bin.path=data/smart/7.1/smart.HMMs.bin
smart.overlapping.path=
smart.threshold.path=
#SuperFamily 1.75
superfamily.hmm.path.3.0=data/superfamily/1.75/hmmlib_1.75
superfamily.self.hits.path.1.75=data/superfamily/1.75/self_hits.tab
superfamily.cla.path.1.75=data/superfamily/1.75/dir.cla.scop.txt_1.75
superfamily.model.tab.path.1.75=data/superfamily/1.75/model.tab
superfamily.pdbj95d.path.1.75=data/superfamily/1.75/pdbj95d
#tigrfam 15.0
tigrfam.signature.library.release=15.0
tigrfam.hmm.path=data/tigrfam/15.0/TIGRFAMs_HMM.LIB
#TMHMM 2.0
# Note: TMHMM model files not distributed with InterProScan 5, please install separately e.g. in data/tmhmm/2.0/TMHMM2.0.model
tmhmm.model.path=data/tmhmm/2.0c/TMHMM2.0c.model
##
## cpu options for parallel processing
##
#hmmer cpu options for the different jobs
hmmer3.hmmsearch.cpu.switch.gene3d=--cpu 1
hmmer3.hmmsearch.cpu.switch.panther=--cpu 1
hmmer3.hmmsearch.cpu.switch.pfama=--cpu 1
hmmer3.hmmsearch.cpu.switch.sfld=--cpu 1
hmmer3.hmmsearch.cpu.switch.superfamily=--cpu 1
hmmer3.hmmsearch.cpu.switch.tigrfam=--cpu 1
hmmer3.hmmsearch.cpu.switch.hmmfilter=--cpu 1
hmmer2.hmmpfam.cpu.switch.smart=--cpu 1
hmmer2.hmmpfam.cpu.switch.pirsf=--cpu 1
#blastall cpu options
blastall.cpu.switch.pirsf=-a 1
#panther binary cpu options (for blastall and hmmsearch)
panther.binary.cpu.switch=-c 1
#pirsf binary cpu options (for hmmscan)
pirsf.pl.binary.cpu.switch=-cpu 1
##
## max number of proteins per analysis batch
##
# These values control the maximum number of proteins put through
# an analysis in one go - different algorithms have different optimum values.
# Note that if you suffer from out of memory errors, reducing these values
# will almost certainly help, but may reduce the speed of analysis.
analysis.max.sequence.count.CDD=2000
analysis.max.sequence.count.GENE3D=4000
analysis.max.sequence.count.SFLD=32000
analysis.max.sequence.count.TMHMM=16000
analysis.max.sequence.count.PANTHER=800
analysis.max.sequence.count.SMART=800
analysis.max.sequence.count.TIGRFAM=4000
analysis.max.sequence.count.PRINTS=1000
analysis.max.sequence.count.PROSITE_PROFILES=150
analysis.max.sequence.count.PROSITE_PATTERNS=8000
analysis.max.sequence.count.PIRSF=4000
analysis.max.sequence.count.PRODOM=4000
analysis.max.sequence.count.SSF=2000
analysis.max.sequence.count.HAMAP=32000
analysis.max.sequence.count.PFAM=4000
analysis.max.sequence.count.COILS=32000
analysis.max.sequence.count.PHOBIUS=16000
analysis.max.sequence.count.SIGNALP=16000
analysis.max.sequence.count.MOBIDB_LITE=2000
##
## General settings
##
# If multiple hosts are sharing the same file system, a delay may be required to
# avoid stale NFS handles
# nfs.delay.milliseconds=0
# Instructs I5 to completely clean up after itself - leave set to true.
delete.temporary.directory.on.completion=true
##
## Broker TCP Connection
##
# A list of TCP ports that should not be used for messaging. (Apart from this, only ports > 1024 and < 65535 will be used.)
tcp.port.exclusion.list=3879,3878,3881,3882
##
## precalculated match lookup service
##
# By default, if the sequence already has matches available from the EBI, this service will look them
# up for you. Note - at present it will always return all the available matches, ignoring any -appl options
# set on the command line.
precalculated.match.lookup.service.url=http://www.ebi.ac.uk/interpro/match-lookup
#proxy set up
precalculated.match.lookup.service.proxy.host=
precalculated.match.lookup.service.proxy.port=3128
precalculated.match.protein.lookup.batch.size=100
precalculated.match.protein.insert.batch.size=500
precalculated.match.protein.insert.batch.size.nolookup=4000
#Exclude sites from output (residue level annotations)
exclude.sites.from.output=false
##
## getorf configuration for nucleic acid sequences
##
# the following are roughly the times getorf takes to find sequences of open reading frames (ORFs) in n nucleotide sequences
#number of sequences -> approx. time it takes in our tests
# 600000 -> 10 minutes
# 3600000 -> 1 hour
# 7200000 -> 2 hours
# 43200000 -> 12 hours
# JOB: jobLoadNucleicAcidSequence
getorf.minsize=50
##
## Output format
##
# TRUE by default, which means all generated graphical output documents (only SVG at the moment) will be archived (using the Linux command tar).
# This simple switch allows you to switch the archive mode off (simply set it to FALSE).
archiveSVGOutput=true
##
## Master/Stand alone embedded workers
##
# Set the number of embedded workers to the number of processors that you would like to employ
# on the machine you are using to run InterProScan.
#number of embedded workers a master process can have
number.of.embedded.workers=1
maxnumber.of.embedded.workers=16
##
## Distributed mode (Cluster mode)
##
#grid name
grid.name=lsf
#grid.name=other-cluster
#project name for this run - use user.digest
user.digest=i5GridRun
#grid jobs limit : number of jobs you are allowed to run on the cluster
grid.jobs.limit=1000
#time between each bjobs or qstat command to check the status of jobs on the cluster
grid.check.interval.seconds=120
#allow master interproscan to run binaries ()
master.can.run.binaries=true
#deal with unknown step states
recover.unknown.step.state=false
#Grid submission commands (e.g. LSF bsub or SGE qsub) for starting remote workers
#commands the master uses to start new remote workers
grid.master.submit.command=bsub -q QUEUE_NAME
grid.master.submit.high.memory.command=bsub -q QUEUE_NAME -M 8192
#commands a worker uses to start new remote workers
grid.worker.submit.command=bsub -q QUEUE_NAME
grid.worker.submit.high.memory.command=bsub -q QUEUE_NAME -M 8192
# command to start a new worker (new jvm)
worker.command=java -Xms32m -Xmx2048m -jar interproscan-5.jar
# This may be identical to the worker.command argument above, however you may choose to select
# a machine with a much larger available memory, for use when a StepExecution fails.
worker.high.memory.command=java -Xms32m -Xmx2048m -jar interproscan-5.jar
# Set the number of embedded workers to the number of processors that you would like to employ
# on the node machine on which the worker will run.
#number of embedded workers in a remote worker
worker.number.of.embedded.workers=4
worker.maxnumber.of.embedded.workers=4
# max number of connections to the master
master.maxconsumers=48
#number of connections to the worker
worker.maxconsumers=32
#throttled network?
grid.throttle=true
# max number of jobs a tier 1 worker is allowed on its queue
worker.maxunfinished.jobs=32
#network tier depth
max.tier.depth=1
# Active MQ JMS broker temporary data directory
jms.broker.temp.directory=activemq-data/localhost/tmp_storage