目录

Welcome to Cerberus

Paper Pre-print Downloads PePY overall Downloads PePy monthly Downloads PePy weekly BioConda Install Downloads Version Anaconda-Server Latest Release date Anaconda-Server Relative Release date Anaconda-Server Platforms Anaconda-Server License

About

Cerberus transforms raw sequencing (i.e. genomic, transcriptomics, metagenomics, metatranscriptomic) data into knowledge. It is a start to finish python code for versatile analysis of the Functional Ontology Assignments for Metagenomes (FOAM), KEGG, CAZy/dbCAN, VOG, pVOG, PHROG, COG, and a variety of other databases including user customized databases via Hidden Markov Models (HMM) for functional annotation for complete metabolic analysis across the tree of life (i.e., bacteria, archaea, phage, viruses, eukaryotes, and whole ecosystems). Cerberus also provides automatic differential statistics using DESeq2/EdgeR, pathway enrichments with GAGE, and pathway visualization with Pathview R.

GitHub Logo Art by Andra Buchan

Installing Cerberus

Option 1) Mamba

  • Mamba install from bioconda with all dependencies:

Linux/OSX-64

  1. Install mamba using conda
conda install mamba

[!NOTE] Make sure you install mamba in your base conda environment unless you have OSX with ARM architecture (M1/M2 Macs). Follow the OSX-ARM instructions below if you have a Mac with ARM architecture.

  1. Install Cerberus with mamba
mamba create -n cerberus -c conda-forge -c bioconda cerberus
conda activate cerberus
cerberus.py --setup
cerberus.py --download

OSX-ARM (M1/M2)

  1. Set up conda environment
    conda create -y -n cerberus
    conda activate cerberus
    conda config --env --set subdir osx-64
  2. Install mamba, python, and pydantic inside the environment
    conda install -y -c conda-forge mamba python=3.10 "pydantic<2"
  3. Install Cerberus with mamba
    mamba install -y -c conda-forge -c bioconda cerberus
    cerberus.py --setup
    cerberus.py --download

[!NOTE] Mamba is the fastest installer. Anaconda or miniconda can be slow. Also, install mamba from conda not from pip. The pip mamba doesn’t work for install.

Option 2) Anaconda - Linux/OSX-64 Only

  • Anaconda install from bioconda with all dependencies:
conda create -n cerberus -c conda-forge -c bioconda cerberus -y
conda activate cerberus
cerberus.py --setup
cerberus.py --download

Option 3) Manual with conda/mamba from Github

git clone https://github.com/raw-lab/cerberus.git 
cd cerberus
bash install_cerberus.sh
conda activate Cerberus
cerberus.py --download

Brief Overview

Cerberus Workflow

General Info

  • Cerberus has three basic modes:
    1. Quality Control (QC) for raw reads
    2. Formatting/gene prediction
    3. Annotation
  • Cerberus can use three different input files:
    1. Raw read data from any sequencing platform (Illumina, PacBio, or Oxford Nanopore)
    2. Assembled contigs, as MAGs, vMAGs, isolate genomes, or a collection of contigs
    3. Amino acid fasta (.faa), previously called pORFs
  • We offer customization, including running all databases together, individually or specifying select databases. For example, if a user wants to run prokaryotic or eukaryotic-specific KOfams, or an individual database alone such as dbCAN, both are easily customized within Cerberus.

  • In QC mode, raw reads are quality controlled via FastQC prior and post trim FastQC. Raw reads are then trimmed via data type; if the data is Illumina or PacBio, fastp is called, otherwise it assumes the data is Oxford Nanopore, then PoreChop is utilized.

  • If Illumina reads are utilized, an optional bbmap step to remove the phiX174 genome is available or user provided contaminate genome. Phage phiX174 is a common contaminant within the Illumina platform as their library spike-in control. We highly recommend this removal if viral analysis is conducted, as it would provide false positives to ssDNA microviruses within a sample.

  • We include a --skip_decon option to skip the filtration of phiX174, which may remove common k-mers that are shared in ssDNA phages.

  • In the formatting and gene prediction stage, contigs and genomes are checked for N repeats. These N repeats are removed by default.

  • We impute contig/genome statistics (e.g., N50, N90, max contig) via our custom module Metaome Stats.

  • Contigs can be converted to pORFs using Prodigal, FragGeneScanRs , and Prodigal-gv as specified by user preference.

  • Scaffold annotation is not recommended due to N’s providing ambiguous annotation.

  • Both Prodigal and FragGeneScanRs can be used via our --super option, and we recommend using FragGeneScanRs for samples rich in eukaryotes.

  • FragGeneScanRs found more ORFs and KOs than Prodigal for a stimulated eukaryote rich metagenome. HMMER searches against the above databases via user specified bitscore and e-values or our minimum defaults (i.e., bitscore = 25, e-value = 1 x 10-9 ).

Input file formats

  • From any NextGen sequencing technology (from Illumina, PacBio, Oxford Nanopore)
  • type 1 raw reads (.fastq format)
  • type 2 nucleotide fasta (.fasta, .fa, .fna, .ffn format), assembled raw reads into contigs
  • type 3 protein fasta (.faa format), assembled contigs which genes are converted to amino acid sequence

Output Files

  • If an output directory is given, that folder will be created where all files are stored.
  • If no output directory is specified, the ‘results_cerberus’ subfolder will be created in the current directory.
  • Gage/Pathview R analysis provided as separate scripts within R.

Visualization of Outputs

  • We use Plotly to visualize the data
  • Once the program is finished running, the html reports containing the visuals will be saved to the last step of the pipeline.
  • The HTML files require plotly.js to be present. One has been provided in the package and is saved to the report folder.

Annotation Rules

Cerberus Rules

  • Rule 1 is for finding high quality matches across databases. It is a score pre-filtering module for pORFs thresholds: which states that each pORF match to an HMM is recorded by default or a user-selected cut-off (i.e., e-value/bit scores) per database independently, or across all default databases (e.g, finding best hit), or per user specification of the selected database.
  • Rule 2 is to avoid missing genes encoding proteins with dual domains that are not overlapping. It is imputed for non-overlapping dual domain module pORF threshold: if two HMM hits are non-overlapping from the same database, both are counted as long as they are within the default or user selected score (i.e., e-value/bit scores).
  • Rule 3 is to ensure overlapping dual domains are not missed. This is the dual independent overlapping domain module for convergent binary domain pORFs. If two domains within a pORF are overlapping <10 amino acids (e.g, COG1 and COG4) then both domains are counted and reported due to the dual domain issue within a single pORF. If a function hits multiple pathways within an accession, both are counted, in pathway roll-up, as many proteins function in multiple pathways.
  • Rule 4 is the equal match counter to avoid missing high quality matches within the same protein. This is an independent accession module for a single pORF: if both hits within the same database have equal values for both e-value and bit score but are different accessions from the same database (e.g., KO1 and KO3) then both are reported.
  • Rule 5 is the ‘winner take all’ match rule for providing the best match. It is computed as the winner takes all module for overlapping pORFs: if two HMM hits are overlapping (>10 amino acids) from the same database the lowest resulting e-value and highest bit score wins.
  • Rule 6 is to avoid partial or fractional hits being counted. This ensures that only whole discrete integer counting (e.g., 0, 1, 2 to n) are computed and that partial or fractional counting is excluded.

Quick start examples

Genome examples

All databases

conda activate cerberus
cerberus.py --prodigal lambda.fna --hmm ALL --dir_out lambda_dir

Only KEGG/FOAM all

conda activate cerberus
cerberus.py --prodigal lambda.fna --hmm KOFam_all --dir_out lambda_ko-only_dir

Only KEGG/FOAM prokaryotic centric

conda activate cerberus
cerberus.py --prodigal ecoli.fna --hmm KOFam_prokaryote --dir_out ecoli_ko-only_dir

Only KEGG/FOAM eukaryotic centric

conda activate cerberus
cerberus.py --fraggenescan human.fna --hmm KOFam_eukaryote --dir_out human_ko-only_dir

Only Viral/Phage databases

conda activate cerberus
cerberus.py --prodigal lambda.fna --hmm VOG, PHROG --dir_out lambda_vir-only_dir
  • NOTE: You can pick any single database you want for your analysis including KOFam_all, COG, VOG, PHROG, CAZy or specific KO databases for eukaryotes and prokaryotes (KOFam_eukaryote or KOFam_prokaryote).

Custom HMM

conda activate cerberus
cerberus.py --prodigal lambda.fna --hmm Custom.hmm --dir_out lambda_vir-only_dir

Illumina data

Bacterial, Archaea and Bacteriophage metagenomes/metatranscriptomes

conda activate cerberus
cerberus.py --prodigal [input_folder] --illumina --meta --dir_out [out_folder] 

Eukaryotes and Viruses metagenomes/metatranscriptomes

conda activate cerberus
cerberus.py --fraggenescan [input_folder] --illumina --meta --dir_out [out_folder] 

Nanopore data

Eukaryotes

conda activate cerberus
cerberus.py --fraggenescan [input_folder] --nanopore --dir_out [out_folder] 

PacBio data

Eukaryotes

conda activate cerberus
cerberus.py --fraggenescan [input_folder] --pacbio --dir_out [out_folder]

SUPER (both methods)

conda activate cerberus
cerberus.py --super [input_folder] --pacbio/--nanopore/--illumina --dir_out [out_folder]
  • Note: Fraggenescan will work for prokaryotes and viruses/bacteriophage but prodigal will not work well for eukaryotes.

Prerequisites and dependencies

  • python >= 3.8

Available from Bioconda - external tool list

Tool Version Publication
Fastqc 0.12.1 None
Fastp 0.23.4 Chen et al. 2018
Porechop 0.2.4 None
bbmap 39.06 None
Prodigal 2.6.3 Hyatt et al. 2010
FragGeneScanRs v1.1.0 Van der Jeugt et al. 2022
Prodigal-gv 2.2.1 Camargo et al. 2023
Phanotate 1.5.0 McNair et al. 2019
HMMER 3.4 Johnson et al. 2010

Cerberus databases

All pre-formatted databases are present at OSF.

Database sources

Database Last Update Version Publication Cerberus Update Version
KEGG/KOfams 2024-01-01 Jan24 Aramaki et al. 2020 beta
FOAM/KOfams 2017 1 Prestat et al. 2014 beta
COG 2020 2020 Galperin et al. 2020 beta
dbCAN/CAZy 2023-08-02 12 Yin et al., 2012 beta
VOG 2017-03-03 80 Website beta
pVOG 2016 2016 Grazziotin et al. 2017 1.2
PHROG 2022-06-15 4 Terizan et al., 2021 1.2
PFAM 2023-09-12 36 Mistry et al. 2020 1.3
TIGRfams 2018-06-19 15 Haft et al. 2003 1.3
PGAPfams 2023-12-21 14 Tatusova et al. 2016 1.3
AMRFinder-fams 2024-02-05 2024-02-05 Feldgarden et al. 2021 1.3
NFixDB 2024-01-22 2 Bellanger et al. 2024 1.3
GVDB 2021 1 Aylward et al. 2021 1.3
Pads Arsenal 2019-09-09 1 Zhang et al. 2020 Coming soon
efam-XC 2021-05-21 1 Zayed et al. 2021 Coming soon
NMPFams 2021 1 Baltoumas et al. 2024 Coming soon
MEROPS 2017 1 Rawlings et al. 2018 Coming soon
FESNov 2024 1 Rodríguez del Río et al. 2024 Coming soon

[!NOTE] The KEGG database contains KOs related to Human disease. It is possible that these will show up in the results, even when analyzing microbes. eggNOG and FunGene database are coming soon. If you want a custom HMM build please let us know by email or leaving an issue.

Custom Database

To run a custom database, you need a HMM containing the protein family of interest and a metadata sheet describing the HMM required for look-up tables and downstream analysis. For the metadata information you need an ID that matches the HMM and a function or hierarchy. See example below.

Example Metadata sheet

ID Function
HMM1 Sugarase
HMM2 Coffease

Cerberus Options

[!Important] If the Cerberus environment is not used, make sure the dependencies are in PATH or specified in the config file.

  • Run cerberus.py with the options required for your project.

Usage of cerberus.py:

[!Note] The following are different options/arguments to modify the execution of Cerberus.

Setup arguments:

Argument/Option Function [Default] Usage Format Accepted format Example (Type as one line)
--setup Setup additional dependencies [False] --setup N/A cerberus.py --setup
--update Update downloaded databases [False] --update N/A cerberus.py --update
--list-db List available and downloaded databases [False] --list-db N/A cerberus.py --list-db
--download Downloads selected HMMs. Use the option --list-db for a list of available databases, default is to download all available databases --download [DOWNLOAD ...] --download [.HMM FILE] --download path/to/example/directory.hmm
--uninstall Remove downloaded databases and FragGeneScan+ [False] --uninstall N/A cerberus.py --uninstall

Input File Arguments:

[!Important] At least one sequence is required.

Accepted formats: [.fastq, .fq, .fasta, .fa, .fna, .ffn, .faa]

Example:

  • cerberus.py --prodigal file1.fasta
  • cerberus.py --config file.config

If a sequence is given in [.fastq, .fq] format, one of --nanopore, --illumina, or --pacbio is required.:

Option format interpretation:

  • --setup = accepts no additional options

  • --download DOWNLOAD = accepts one option, (represented by capitalized command ‘DOWNLOAD’)

  • --fraggenescan FRAGGENESCAN [FRAGGENESCAN...] = accepts one or greater options (represented by capitalized commands)

Argument/Option Function Usage Format Accepted format # Options Accepted Example (Type as one line)
-c or --config Path to config file, command line takes priority -c CONFIG or --config CONFIG Path to config file 1 -c path/to/config/file
--prodigal Prokaryote nucleotide sequence (includes microbes, bacteriophage) --prodigal PRODIGAL [PRODIGAL ...] Sequence file =>1 --prodigal FILE1 FILE2...
--fraggenescan Eukaryote nucleotide sequence (includes other viruses, works all around for everything) --fraggenescan FRAGGENESCAN [FRAGGENESCAN ...] Sequence file =>1 --fraggenescan FILE1 FILE2...
--super Run sequence in both --prodigal and --fraggenescan modes --super SUPER [SUPER ...] Sequence file =>1 --super FILE1 FILE2...
--prodigalgv Giant virus nucleotide sequence --prodigalgv PRODIGALGV [PRODIGALGV ...] Sequence file =>1 --prodigalgv FILE1 FILE2...
--phanotate Phage sequence --phanotate PHANOTATE [PHANOTATE ...] Sequence file =>1 --phanotate FILE1 FILE2...
--protein or --amino Protein Amino Acid sequence --protein PROTEIN [PROTEIN ...] or --amino PROTEIN [PROTEIN ...] Sequence file =>1 --protein FILE1 FILE2... or --amino FILE1 FILE2...
--hmmer-tsv Annotations tsv file from HMMER (experimental) --hmmer-tsv HMMER_TSV [HMMER_TSV ...] Sequence file =>1 --hmmer-tsv FILE1 FILE2...
--class path to a tsv file which has class information for the samples. If this file is included, scripts will be included to run Pathview in R --class CLASS Path to TSV file 1 `–class TSV_FILE1
--illumina Specifies that the given FASTQ files are from Illumina --illumina N/A N/A cerberus.py --illumina
--nanopore Specifies that the given FASTQ files are from Nanopore --nanopore N/A N/A cerberus.py --nanopore
--pacbio Specifies that the given FASTQ files are from PacBio --pacbio N/A N/A cerberus.py --pacbio

Output options:

Argument/Option Function [DEFAULT] Usage Format Accepted format # Options Accepted Example (Type as one line)
--dir-out path to output directory, defaults to “results-cerberus” in current directory. [./results-cerberus] --dir-out DIR_OUT output file path 1 --dir-out path/to/output/file
--replace Flag to replace existing files. [False] --replace cerberus.py option N/A cerberus.py --replace
--keep Flag to keep temporary files. [False] --keep cerberus.py option N/A cerberus.py --keep
--tmpdir Temp directory for RAY (experimental) [system tmp dir] --tmpdir TMPDIR cerberus.py option 1 --tmpdir TEMPFILE1

Database options:

Argument/Option Function [DEFAULT] Usage Format Accepted format # Options Accepted Example (Type as one line)
--hmm A list of databases for HMMER. Use the option --list-db for a list of available databases [KOFam_all] --hmm HMM [HMM ...] cerberus.py option =>1 cerberus.py --hmm DATABASE1 DATABASE2...
--db-path Path to folder of databases [Default: under the library path of Cerberus] --db-path DB_PATH path to databases folder 1 --db-path path/to/databases/folder

Optional Arguments:

Argument/Option Function [DEFAULT] Usage Format Accepted format # Options Accepted Example (Type as one line)
--scaffolds Sequences are treated as scaffolds [False] --scaffolds cerberus.py option N/A cerberus.py --scaffolds
--minscore Score cutoff for parsing HMMER results [60] --minscore MINSCORE whole integer value 1 cerberus.py --minscore 50
--evalue E-value cutoff for parsing HMMER results [1e-09] --evalue EVALUE E-value 1 cerberus.py --evalue [E-value]
--skip-decon Skip decontamination step. [False] --skip-decon cerberus.py option N/A cerberus.py --skip-decon
--skip-pca Skip PCA. [False] --skip-pca cerberus.py option N/A cerberus.py --skip-pca
--cpus Number of CPUs to use per task. System will try to detect available CPUs if not specified [Auto Detect] --cpus CPUS whole integer value 1 cerberus.py --cpus 16
--chunker Split files into smaller chunks, in Megabytes [Disabled by default] --chunker CHUNKER whole integer value 1 cerberus.py --chunker 300
--grouped Group multiple fasta files into a single file before processing. When used with --chunker (see above) can improve speed --grouped cerberus.py option N/A cerberus.py --grouped
--version or -v show the version number and exit --version or -v cerberus.py option N/A cerberus.py --version
-h or --help show this help message and exit -h or --help cerberus.py option N/A cerberus.py -h
--adapters FASTA File containing adapter sequences for trimming --adapters ADAPTERS FASTA file 1 cerberus.py --adapters /path/to/FASTA/file
--qc_seq FASTA File containing control sequences for decontamination --qc_seq QC_SEQ FASTA file 1 cerberus.py --qc_seq /path/to/FASTA/file

[!NOTE] Arguments/options that start with -- can also be set in a config file (specified via -c). Config file syntax allows: key=value, flag=true, stuff=[a,b,c] (for details, see syntax. In general, command-line values override config file values which override defaults.

OUTPUTS (/final folder)

File Extension Description Summary Cerberus Update Version
.gff General Feature Format 1.3
.gbk GenBank Format 1.3
.fna Nucleotide FASTA file of the input contig sequences. 1.3
.faa Protein FASTA file of the translated CDS/ORFs sequences. 1.3
.ffn FASTA Feature Nucleotide file, the Nucleotide sequence of translated CDS/ORFs. 1.3
.html Summary statistics and/or visualizations, in step 10 folder 1.3
.txt Statistics relating to the annotated features found. 1.3
level.tsv Various levels of hierachical steps that is tab-separated file from various databases 1.3
rollup.tsv All levels of hierachical steps that is tab-separated file from various databases 1.3
.tsv Final Annotation summary, Tab-separated file of all features from various databases 1.3

GAGE / PathView

After processing the HMM files, Cerberus calculates a KO (KEGG Orthology) counts table from KEGG/FOAM for processing through GAGE and PathView. GAGE is recommended for pathway enrichment followed by PathView for visualize the metabolic pathways. A “class” file is required through the --class option to run this analysis.

[!Tip] As we are unsure which comparisons you want to make thus, you have to make a class.tsv so the code will know the comparisons you want to make.

For example (class.tsv):

Sample Class
1A rhizobium
1B non-rhizobium

The output is saved under the step_10-visualizeData/combined/pathview folder. Also, at least 4 samples need to be used for this type of analysis.

  • GAGE and PathView also require internet access to be able to download information from a database.
  • Cerberus will save a bash script run_pathview.sh in the step_10-visualizeData/combined/pathview directory along with the KO Counts tsv files and the class file for running manualy in case Cerberus was run on a cluster without access to the internet.

Multiprocessing / Multi-Computing with RAY

  • Cerberus uses Ray for distributed processing. This is compatible with both multiprocessing on a single node (computer) or multiple nodes in a cluster.
  • Cerberus has been tested on a cluster using Slurm.

[!Important] A script has been included to facilitate running Cerberus on Slurm. To use Cerberus on a Slurm cluster, setup your slurm script and run it using sbatch.

sbatch example_script.sh

example script:

#!/usr/bin/env bash

#SBATCH --job-name=test-job
#SBATCH --nodes=3
#SBATCH --tasks-per-node=1
#SBATCH --cpus-per-task=16
#SBATCH --mem=128MB
#SBATCH -e slurm-%j.err
#SBATCH -o slurm-%j.out
#SBATCH --mail-type=END,FAIL,REQUEUE

echo "====================================================="
echo "Start Time  : $(date)"
echo "Submit Dir  : $SLURM_SUBMIT_DIR"
echo "Job ID/Name : $SLURM_JOBID / $SLURM_JOB_NAME"
echo "Node List   : $SLURM_JOB_NODELIST"
echo "Num Tasks   : $SLURM_NTASKS total [$SLURM_NNODES nodes @ $SLURM_CPUS_ON_NODE CPUs/node]"
echo "======================================================"
echo ""

# Load any modules or resources here
conda activate cerberus
# source the slurm script to initialize the Ray worker nodes
source ray-slurm-cerberus.sh
# run Cerberus
cerberus.py --prodigal [input_folder] --illumina --dir_out [out_folder]

echo ""
echo "======================================================"
echo "End Time   : $(date)"
echo "======================================================"
echo ""

DESeq2 and Edge2 Type I errors

Both edgeR and DeSeq2 R have the highest sensitivity when compared to other algorithms that control type-I error when the FDR was at or below 0.1. EdgeR and DESeq2 all perform fairly well in simulation and via data splitting (so no parametric assumptions). Typical benchmarks will show limma having stronger FDR control across all types of datasets (it’s hard to beat the moderated t-test), and edgeR and DESeq2 having higher sensitivity for low counts (makes sense as limma has to filter these out / down-weight them to use the normal model on log counts). Further information about type I errors are present from Mike Love’s vignette here.

Contributing to Cerberus and Fungene

Cerberus as a community resource as recently acquired FunGene, we welcome contributions of other experts expanding annotation of all domains of life (viruses, bacteria, archaea, eukaryotes). Please send us an issue on our Cerberus GitHub open an issue; or email us we will fully annotate your genome, add suggested pathways/metabolisms of interest, make custom HMMs to be added to Cerberus and FunGene.

This is copyrighted by University of North Carolina at Charlotte, Jose L Figueroa III, Eliza Dhungal, Madeline Bellanger, Cory R Brouwer and Richard Allen White III. All rights reserved. Cerberus is a bioinformatic tool that can be distributed freely for academic use only. Please contact us for commerical use. The software is provided “as is” and the copyright owners or contributors are not liable for any direct, indirect, incidental, special, or consequential damages including but not limited to, procurement of goods or services, loss of use, data or profits arising in any way out of the use of this software.

Citing Cerberus

If you are publishing results obtained using Cerberus, please cite:

Publication

Figueroa III JL, Dhungel E, Bellanger M, Brouwer CR, White III RA. 2024. Cerberus: distributed highly parallelized HMM-based processing for robust functional annotation across the tree of life. Bioinformatics

Pre-print

Figueroa III JL, Dhungel E, Brouwer CR, White III RA. 2023.
Cerberus: distributed highly parallelized HMM-based processing for robust functional annotation across the tree of life. bioRxiv

CONTACT

The informatics point-of-contact for this project is Dr. Richard Allen White III.
If you have any questions or feedback, please feel free to get in touch by email.
Dr. Richard Allen White III
Jose Luis Figueroa III
Or open an issue.

关于

长读长转录本结构整合与注释工具,用于异构体比较和转录复杂性分析。

97.7 MB
邀请码
    Gitlink(确实开源)
  • 加入我们
  • 官网邮箱:gitlink@ccf.org.cn
  • QQ群
  • QQ群
  • 公众号
  • 公众号

版权所有:中国计算机学会技术支持:开源发展技术委员会
京ICP备13000930号-9 京公网安备 11010802032778号