Phynteny is annotation tool for bacteriophage genomes that integrates protein language models and gene synteny. phynteny-transformer leverages a transformer architecture with attention mechanisms and long short term memory to capture the positional information of genes.
phynteny-transformer takes a genbank file with PHROG annotations as input. If you haven’t already annotated your phage(s) with Pharokka and Phold go do that and then come right back here!
Dependencies
To run phynteny-transformer, you need the following dependencies:
phynteny_transformer.gbk contains a GenBank format file that has been updated to include annotations generated using Phynteny along with their Phynteny score and confidence.
phynteny_per_cds_funcions.tsv provides a table of the annotations generated (similar to the pharokka_cds_functions.tsv from Pharokka)
Advanced Usage
Phynteny Transformer provides an advanced mode for specifying the parameters of a model that you trained yourself. To see all advanced options:
phynteny_transformer --help --advanced
How do I run Phynteny with other phage annotation tools?
Phynteny leverages existing phage annotations to predict unknown genes. For this reason, we recommend using Phynteny after annotating using Pharokka and Phold.
The plot below shows the annotation rate of different tools across 4 benchmarked datasets ((a) INPHARED 1419, (b) Cook, (c) Crass and (d) Tara - see the Phold preprint for more information). The final Phynteny plots combine the benefits of annotation with Pharokka (with HMM, the second violin) followed by Phold (with structures, the fourth violin)followed by Phynteny.
Thank you to @gbouras13 for the awesome figure!
You can run pharokka + phold + phynteny in an interactive notebook using this link
How does Phynteny work?
The following figure summarises how Phynteny was created!
Training Custom Models
Phynteny Transformer allows you to train your own custom models. To train a model, you need to provide a dataset in the required format and specify the training parameters. For more details, refer to the documentation in the train_transformer directory.
Bugs and Suggestions
If you break Phynteny or would like to make any suggestions please open an issue or email me at susie.grigson@gmail.com and I’ll try to get back to you.
Acknowledgements
Thankyou to Laura Inglis for designing the Phynteny logo!
Phynteny was trained using resources provided by the Pawsey Supercomputing Research Centre (Perth, Australia) which is funded by the Australian Government. Analysis was performed using the Flinders University DeepThought High Performance Cluster (https://doi.org/10. 25957/FLINDERS.HPC.DEEPTHOUGHT).
Wow! how can I cite this?
Preprint for Phynteny is available here.
You can cite Phynteny as:
Grigson, S.R., Bouras, G., Papudeshi, B., Mallawaarachchi, V., Roach, M.J., Decewicz, P., & Edwards, R.A. (2025). Synteny-aware functional annotation of bacteriophage genomes with Phynteny. bioRxiv, 2025.07.28.667340. https://doi.org/10.1101/2025.07.28.667340.
Phynteny-Transformer
Phynteny is annotation tool for bacteriophage genomes that integrates protein language models and gene synteny.
phynteny-transformerleverages a transformer architecture with attention mechanisms and long short term memory to capture the positional information of genes.phynteny-transformertakes a genbank file with PHROG annotations as input. If you haven’t already annotated your phage(s) with Pharokka and Phold go do that and then come right back here!Dependencies
To run
phynteny-transformer, you need the following dependencies:You can install the dependencies using pip:
How do I use Phynteny?
Installation
Option 1: Installing Phynteny using conda
You can install Phynteny from bioconda at https://anaconda.org/bioconda/phynteny. Make sure you have
condainstalled.Now you can go to Install Models to install pre-trained phynteny-transformer models.
Option 2: Installing Phynteny using pip
You can install Phynteny from PyPI at https://pypi.org/project/phynteny/.
Now you can go to Install Models to install pre-trained phynteny models.
Option 3: Installing Phynteny from source
You can install Phynteny Transformer from source.
NOTE: Source installation is recommended if you would like to train your own
phynteny-trasformermodels.Install Models
Before you can run
phynteny-transformeryou’ll need to install some databasesIf you would like to install them to a specific location
If this doesn’t work you can download the models directly from Zenodo and untar them yourself and point Phynteny to them with the
-mflag.Quick Start
Output
phynteny_transformer.gbkcontains a GenBank format file that has been updated to include annotations generated using Phynteny along with their Phynteny score and confidence.phynteny_per_cds_funcions.tsvprovides a table of the annotations generated (similar to thepharokka_cds_functions.tsv from Pharokka)Advanced Usage
Phynteny Transformer provides an advanced mode for specifying the parameters of a model that you trained yourself. To see all advanced options:
How do I run Phynteny with other phage annotation tools?
Phynteny leverages existing phage annotations to predict unknown genes. For this reason, we recommend using Phynteny after annotating using Pharokka and Phold.
The plot below shows the annotation rate of different tools across 4 benchmarked datasets ((a) INPHARED 1419, (b) Cook, (c) Crass and (d) Tara - see the Phold preprint for more information). The final Phynteny plots combine the benefits of annotation with Pharokka (with HMM, the second violin) followed by Phold (with structures, the fourth violin)followed by Phynteny.
Thank you to @gbouras13 for the awesome figure!
You can run
pharokka+phold+phyntenyin an interactive notebook using this linkHow does Phynteny work?
The following figure summarises how Phynteny was created!
Training Custom Models
Phynteny Transformer allows you to train your own custom models. To train a model, you need to provide a dataset in the required format and specify the training parameters. For more details, refer to the documentation in the train_transformer directory.
Bugs and Suggestions
If you break Phynteny or would like to make any suggestions please open an issue or email me at susie.grigson@gmail.com and I’ll try to get back to you.
Acknowledgements
Thankyou to Laura Inglis for designing the Phynteny logo!
Phynteny was trained using resources provided by the Pawsey Supercomputing Research Centre (Perth, Australia) which is funded by the Australian Government. Analysis was performed using the Flinders University DeepThought High Performance Cluster (https://doi.org/10. 25957/FLINDERS.HPC.DEEPTHOUGHT).
Wow! how can I cite this?
Preprint for Phynteny is available here.
You can cite Phynteny as:
Grigson, S.R., Bouras, G., Papudeshi, B., Mallawaarachchi, V., Roach, M.J., Decewicz, P., & Edwards, R.A. (2025). Synteny-aware functional annotation of bacteriophage genomes with Phynteny. bioRxiv, 2025.07.28.667340. https://doi.org/10.1101/2025.07.28.667340.