Difference between revisions of "Applications/guppy"

From HPC
Jump to: navigation , search
m
Line 3: Line 3:
 
==Application Details==
 
==Application Details==
  
* Description: ABySS is a de novo sequence assembler intended for short paired-end reads and large genomes.
+
* Description: Local accelerated basecalling for Nanopore data.
* Version: 1.5.2 (gcc-4.9.3)
+
* Version: guppy 3.1.5
* Modules: abyss/1.5.2/gcc-4.9.3
+
* Modules: guppy/cpu/3.1.5 and guppy/gpu/3.1.5 (GPU)
 
* Licence: Free, open-source
 
* Licence: Free, open-source
  
 
==Usage Examples==
 
==Usage Examples==
  
===Assemble a small synthetic data set===
+
===Batch example===
  
<pre style="background-color: #f5f5dc; color: black; font-family: monospace, sans-serif;">
+
This uses 2 GPUs presently on GPU05 (compute level is high than 6) which is required by guppy's minimum requirement.
 
 
[username@login] module add abyss/1.5.2/gcc-4.9.3
 
[username@login] wget http://www.bcgsc.ca/platform/bioinfo/software/abyss/releases/1.3.4/test-data.tar.gz
 
[username@login] tar xzvf test-data.tar.gz
 
[username@login} abyss-pe k=25 name=test \
 
    in='test-data/reads1.fastq test-data/reads2.fastq'
 
 
 
</pre>
 
 
 
===Calculate assembly contiguity statistics===
 
 
 
<pre style="background-color: #f5f5dc; color: black; font-family: monospace, sans-serif;">
 
 
 
[username@login] module add abyss/1.5.2/gcc-4.9.3
 
[username@login] abyss-fac test-unitigs.fa
 
 
 
</pre>
 
 
 
===Parallel processing===
 
 
 
The np option of abyss-pe specifies the number of processes to use for the parallel MPI job. Without any MPI configuration, this will allow you to use multiple cores on a single machine. To use multiple machines for assembly, you must create a hostfile for mpirun, which is described in the mpirun man page.
 
 
 
Do not run '''mpirun -np 8 abyss-pe'''. To run ABySS with 8 threads, use '''abyss-pe np=8'''. The abyss-pe driver script will start the MPI process, like so: '''mpirun -np 8 ABYSS-P.'''
 
 
 
The paired-end assembly stage is multithreaded, but must run on a single machine. The number of threads to use may be specified with the parameter j. The default value for j is the value of np.
 
 
 
'''Note''': this example is done on a high memory node, usually access would be achieved with the scheduler
 
 
 
<pre style="background-color: black; color: white; border: 2px solid black; font-family: monospace, sans-serif;">
 
 
 
[username@c230 ~]$ module add abyss/1.5.2/gcc-4.9.3
 
[username@c230 ~]$ mpirun abyss-pe np=40
 
 
 
</pre>
 
 
 
Through '''SLURM''' this would become the script:
 
 
 
<pre style="background-color: black; color: white; border: 2px solid black; font-family: monospace, sans-serif;">
 
  
 +
<pre>
 
#!/bin/bash
 
#!/bin/bash
 
+
#SBATCH -J guppy.job
#SBATCH -J abyss
+
#SBATCH --exclusive
#SBATCH -p highmem
+
#SBATCH -o gpu05.%j.out
#SBATCH -N 2
+
#SBATCH -e gpu05.%j.err
#SBATCH --ntasks-per-node=40
+
#SBATCH -p gpu05
#SBATCH -o %N.%j.%a.out
+
#SBATCH --gres=gpu:tesla:2
#SBATCH -e %N.%j.%a.err
 
 
#SBATCH --exclusive
 
#SBATCH --exclusive
#SBATCH -t 00:30:00
+
 
  
 
module purge
 
module purge
module add abyss/1.5.2/gcc-4.9.3
+
module load cuda/10.1.168
 +
module load guppy/gpu/3.1.5
  
#Run your ABySS commands
 
  
abyss-pe name=test k=48 n=8 in='test-1.fa test-3.fa'
+
guppy_basecaller --input_path /home/user/fast5 --save_path /home/user/guppy --flowcell FLO-MIN106 --kit SQK-RPB004 --min_qscore 7 --qscore_filtering -x cuda:0 cuda:1
  
 
</pre>
 
</pre>
  
The Quantum Package has been designed pretty much as an interactive environment for quantum-chemistry calculations, in order to facilitate the user experience.
 
 
* qp                         
 
* qp_e_conv_fci             
 
* qp_name                   
 
* qp_reset                   
 
* qp_stop                   
 
* qpsh
 
* qp_convert_output_to_ezfio  qp_export_as_tgz           
 
* qp_plugins                 
 
* qp_set_frozen_core         
 
* qp_test
 
* qp_create_ninja           
 
* qp_mpirun                 
 
* qp_prepend_export         
 
* qp_srun                   
 
* qp_update
 
* |>
 
  
  
Line 95: Line 39:
 
==Further Information==
 
==Further Information==
  
* [http://computing.bio.cam.ac.uk/local/doc/abyss.html http://computing.bio.cam.ac.uk/local/doc/abyss.html]
+
* https://community.nanoporetech.com/ (login required)
  
 
==Navigation==
 
==Navigation==

Revision as of 13:59, 14 June 2019

Application Details

  • Description: Local accelerated basecalling for Nanopore data.
  • Version: guppy 3.1.5
  • Modules: guppy/cpu/3.1.5 and guppy/gpu/3.1.5 (GPU)
  • Licence: Free, open-source

Usage Examples

Batch example

This uses 2 GPUs presently on GPU05 (compute level is high than 6) which is required by guppy's minimum requirement.

#!/bin/bash
#SBATCH -J guppy.job
#SBATCH --exclusive
#SBATCH -o gpu05.%j.out
#SBATCH -e gpu05.%j.err
#SBATCH -p gpu05
#SBATCH --gres=gpu:tesla:2
#SBATCH --exclusive


module purge
module load cuda/10.1.168
module load guppy/gpu/3.1.5


guppy_basecaller --input_path /home/user/fast5 --save_path /home/user/guppy --flowcell FLO-MIN106 --kit SQK-RPB004 --min_qscore 7 --qscore_filtering -x cuda:0 cuda:1



Further Information

Navigation