Next Article in Journal
Epidemiological Analysis of Cassava Mosaic and Brown Streak Diseases, and Bemisia tabaci in the Comoros Islands
Next Article in Special Issue
Small RNA Analyses of a Ceratobasidium Isolate Infected with Three Endornaviruses
Previous Article in Journal
Flavivirus–Host Interaction Landscape Visualized through Genome-Wide CRISPR Screens
Previous Article in Special Issue
Abundance of Poleroviruses within Tasmanian Pea Crops and Surrounding Weeds, and the Genetic Diversity of TuYV Isolates Found
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Development and Validation of a Bioinformatic Workflow for the Rapid Detection of Viruses in Biosecurity

1
Plant Health and Environment Laboratory, Ministry for Primary Industries, P.O. Box 2095, Auckland 1140, New Zealand
2
Animal Health Laboratory, Ministry for Primary Industries, Upper Hutt 5018, New Zealand
*
Author to whom correspondence should be addressed.
Viruses 2022, 14(10), 2163; https://doi.org/10.3390/v14102163
Submission received: 15 July 2022 / Accepted: 25 September 2022 / Published: 30 September 2022
(This article belongs to the Special Issue State-of-the-Art Plant Virus Research in Australasia)

Abstract

:
The field of biosecurity has greatly benefited from the widespread adoption of high-throughput sequencing technologies, for its ability to deeply query plant and animal samples for pathogens for which no tests exist. However, the bioinformatics analysis tools designed for rapid analysis of these sequencing datasets are not developed with this application in mind, limiting the ability of diagnosticians to standardise their workflows using published tool kits. We sought to assess previously published bioinformatic tools for their ability to identify plant- and animal-infecting viruses while distinguishing from the host genetic material. We discovered that many of the current generation of virus-detection pipelines are not adequate for this task, being outperformed by more generic classification tools. We created synthetic MinION and HiSeq libraries simulating plant and animal infections of economically important viruses and assessed a series of tools for their suitability for rapid and accurate detection of infection, and further tested the top performing tools against the VIROMOCK Challenge dataset to ensure that our findings were reproducible when compared with international standards. Our work demonstrated that several methods provide sensitive and specific detection of agriculturally important viruses in a timely manner and provides a key piece of ground truthing for method development in this space.

1. Introduction

The New Zealand (NZ) export market is dominated by high quality agricultural products, with six of the ten top annual exports in 2021 coming from primary industries [1]. Over NZD 47 billion worth of exports, from a total of NZD 63.3 billion, in 2021 came from primary production after a significant growth in many of the major export commodities such as dairy products, meat, and wood [1,2]. In 2021 (the year to 30 June 2022) the NZ dairy sector generated NZD 19.1 billion in export revenue, roughly equal to that revenue of meat and wool ($10.4 billion NZD) and horticulture (NZD 6.6 billion) exports. NZ is currently free from many of the plant and animal viruses which compromise production and export value. In a country whose economy is so tightly linked to primary production, employing effective biosecurity measures to maintain this state is greatly desired amongst growers and exporters [3]. Safeguarding these industries is therefore a top priority for the Ministry for Primary Industries, to maintain and expand the productivity of NZ.
The introduction of exotic pathogens is a constant concern for NZ as international trade, while such an integral part of the economy, is also a major pathway for the introduction of economically significant viruses [4]. Incursion and subsequent establishment of regulated viruses would result in a decline of yield and quality of domestically grown, high-value products. Beyond the obvious reduction in marketable value for lost or disfigured crops and animals, invasions carry implications for trade agreements, as preferential access to international markets is often contingent upon the absence of specific pathogens. For example, an outbreak of Foot-and-mouth disease (FMD) could freeze NZ meat exports for months, resulting in billions of dollars of lost export value [5]. Similar consequences may occur with the introduction of the viruses responsible for African swine fever and Lumpy skin disease—losses of production and export value, as well as eradication, compensation, and surveillance costs may have considerable economic impact on primary industries [6,7]. Although these viruses are currently absent from NZ there are increasing concerns for their introduction into Australia, a major trading partner of NZ [8,9,10,11]. To address the ever-increasing risk of virus incursions, high-throughput sequencing (HTS) is now becoming routine within the NZ biosecurity sector, particularly when performing plant virus diagnostics [12]. The benefits of HTS as a diagnostic tool are widely acknowledged [13,14,15,16,17,18] and this approach has gained favour within NZ primarily for its broad target range when compared with traditional diagnostic techniques.
The use of HTS platforms for the sequencing of shotgun metagenomic (and metatranscriptomic) libraries is a technique commonly applied to the study of microorganisms [19,20], particularly in instances where cultivation of microorganisms is challenging. The applicability of high-throughput shotgun sequencing to biosecurity has long been recognised [13] and there are now numerous examples of this technology being used in surveillance and diagnostic settings [15,21,22,23,24,25,26,27]. Unlike traditional diagnostic techniques, such as ELISA and PCR, there is currently little standardisation of HTS laboratory methods and bioinformatic analyses when they are applied to pathogen diagnostics. This is particularly true when an HTS approach is used in the absence of prior knowledge of virus identity and simply mapping reads to a reference genome is not appropriate. For example, of the exemplar references above, each of those published since 2018 performed significantly different techniques for classifying sequence reads and obtaining a final diagnosis.
The variety of analysis methods and classification databases provides flexibility when performing analyses, allowing diagnosticians to customise their approach to specific pathogens of interest. However, this lack of standardisation results in a fragmentation of practices and workflows between institutions and countries, despite efforts to create harmonized approaches [28]. While efforts have been made to develop tools to automate the detection of viral sequences from HTS libraries, such tools often perform poorly when applied beyond their intended use. In a recent literature review, Nooij and colleagues [29] compared performance reports for 25 virus detection workflows and reported an average sensitivity of just 0.67 across all tools and datasets evaluated. Superior results are reported when general-purpose metagenome classifier tools have been applied to virus detection in clinical samples [30,31,32]. Owing to such observations, when performing virus discovery in plant HTS libraries there remains an overwhelming bias towards the use of virus-agnostic analysis pipelines. For example, from 71 HTS-based virus detections summarised by Ibaba and Gubba [33], 65 performed their main detection of viral sequences using short-read assembly, BLASTn or BLASTx queries to the NCBI GenBank database, and screening for viral proteins using Kaiju [34]. Similarly, a 2017 review of HTS applications in plant diagnostics highlighted seven instances of virus identification being performed by HTS, all of which relied on a manuscript-specific BLAST-based approach rather than a published virus pipeline [35]. Consequently, HTS-based initiatives focusing on virus discovery still favour alignment-based search tools or protein fold models for performing initial detections [36,37,38].
The discrepancy between the performance of custom-built virus detection tools and the more generic forms of sequence annotation may be explained by the observation that sequence databases are dominated by particular lineages, which likely results in bias. Many recently developed tools are not specifically calibrated on pathogens or risk organisms [39,40,41,42,43,44] and the datasets used in training are often rich in phage sequence [45,46,47,48,49] which contribute towards global sensitivity and specificity measures without being relevant to diagnoses in eukaryotes. We have previously reported the development of an in-house laboratory protocol for sequencing viruses using the Oxford Nanopore Technologies (ONT) MinION device, which allows us to generate cDNA libraries from infected plant material without the need for polyadenylated RNA capture [12]. However, with uncertainty surrounding the efficacy of existing high-throughput analysis pipelines, the subsequent bioinformatic workflow relied on a BLASTn query against the full NCBI nucleotide database. While undoubtedly thorough, this approach requires significant computational resources, can potentially require days of computation time to generate initial findings, and is not appropriate for samples sequenced using short read technologies without some form of genomic assembly to reduce the number of query sequences to be examined.
As biosecurity diagnoses are performed within tight reporting timeframes, the long classification times, coupled with the inherent uncertainty around the run time of any specific BLAST job led us to explore alternative methods for generating sensitive and accurate classifications with more timely and predictable results. We sought to identify tools which could perform tasks such as short-read assembly and prediction of protein coding regions quickly and robustly, so that an end-to-end diagnostic pipeline could be constructed to reliably screen sequence libraries from a range of hosts to provide timely indications of virus presence in samples. We identified a variety of existing tools and classification databases with the express purpose of detecting virus sequences in HTS libraries while assessing both their ability to identify plant and vertebrate viruses, and their specificity against host or other sequence contaminants. We constructed candidate analysis workflows from the best-performing tools and then tested said workflows against 165 synthetic libraries. Each synthetic library consisted of a plant or animal sample spiked with a virus of interest to the NZ agricultural or horticultural sector, and sequences were simulated for both Oxford Nanopore and Illumina sequencing platforms. In addition, we applied our methods to 18 samples available through the VIROMOCK Challenge [50] a recognised international platform to benchmark bioinformatic pipelines for virus detection.

2. Materials and Methods

2.1. Data Collection and Creation of Test Data

Viral genomes were downloaded from the NCBI RefSeq and GenBank database on the 9th of April 2020. Downloads from RefSeq were performed by iterating through the contents of the ‘ftp.ncbi.nih.gov://genomes/refseq/viral/’ and downloading each species entry via rsync. For each species record downloaded, genome sequence files and GenBank files were preferentially selected from the ‘latest_assembly_versions/’ folder if present, or the ‘all_assembly_versions/’ folder if the first option was empty. Genome records marked as ‘suppressed’ were ignored along with any GenBank records which had a corresponding RefSeq version of the assembly. Viral species were matched to a broad host lineage using the information available at the NCBI assembly reports. In the case of 3709 entries a host record was not provided but could be inferred from the virus name. These genomes are collectively referred to as the ‘target’ dataset.
To test the specificity of databases and analysis pipelines, genomes from the fungus, invertebrate, plant, protozoan, vertebrate_mammalian, and vertebrate_other locations in the NCBI FTP server were downloaded and parsed in the same manner. The sequence data for each lineage was additionally screened to remove identical contigs between records. A dereplicated set of prokaryotic (bacterial and archaeal) genomes were obtained from the Genome Taxonomy Database (release 95) [51,52]. These genomes are collectively referred to as the ‘non-target’ dataset. Protein sequences obtained from the non-target dataset were further dereplicated by clustering of protein sequences using the ‘linclust’ function of MMSeqs2 (version 12.113e3) [53]. Clustering was performed using 80% sequence similarity for the prokaryote lineage, and 95% sequence similarity for the remaining non-target datasets, to reduce the number of queries submitted for profiling tools.

2.2. Creation of Custom Viral Sequence Databases

To achieve faster classification rates than that achieved with BLAST queries against the NCBI nt and nr databases, we constructed virus-specific databases from the viral genomes and gene predictions obtained when constructing the target dataset. As there are many viral sequences for which a full genome are not yet available additional sequences were obtained by extracting all viral sequences from the NCBI nt and nr databases (July 2021 release) then dereplicating the results to a more manageable size. For nucleotide data, the nt sequences were grouped according to NCBI species name and clustering with vsearch. We initially performed grouping by NCBI species name and clustered intra-species sequences at 90% sequence identity, then also created datasets by truncating species names at the key words ‘virus’, ’viroid’, and ’phage’ to produce broader groupings, followed by intra-species sequence clustering with 90%, 97%, and 99% sequence identity thresholds.
When performing nucleotide clustering, two lineages were removed from the input data as they comprised a disproportionately large percentage of reads in the raw data. The first of these corresponded to the SARS-CoV-2 lineage from which genomes have been sequenced at a disproportionately high rate in the last two years. In this instance, we inserted a single representative sequence (NC_045512.2) to represent this lineage. Clustering was also impeded by a large grouping of unclassified viruses belonging to the family Myoviridae. In this instance, the cluster was removed from the database as; (a) viruses of the family Myoviridae infect bacteria and are therefore unlikely to be relevant to the aims of this work, and (b) there were many named isolates of this virus in the database, so the loss of this particular lineage was not deemed detrimental to the taxonomic coverage of the clustered sequence set. Protein clustering was performed by filtering protein sequences fewer than 25 amino acid residues in length, then grouping by either species name or truncated species name grouping and clustering proteins at 90% sequence identity using usearch (11.0.667) [54].

2.3. Creation of Custom Viral HMM Databases

There are numerous high-quality HMM databases of viral protein clusters currently available, however, many are rarely updated and several of the more recent databases comprise tens of thousands of models, making annotation a slow process. We sought to improve upon these resources by producing a database which could compete in terms of specificity and sensitivity while minimising the time required to classify proteins. Protein sequences were extracted from GenBank record files downloaded for the target dataset and identical proteins were removed using seqmagick. The dereplicated proteins were then purged of polyprotein sequences using an all-against-all BLASTP search with DIAMOND (version 0.9.32) [55]. Results were filtered to remove self-hits, and polyproteins were identified using the working definition of the vFAM database project, namely that a polyprotein is one in which greater than 70% of the protein sequence is covered by two or more non-overlapping shorter protein sequences. From an input set of 686,558 unique protein sequences, 33,562 were identified as polyproteins by this workflow and removed from the dataset.
Protein clusters were produced using multiple methods, based on either static of dynamic cluster definitions. Clustering was initially performed with usearch, clustering proteins into groups using identity thresholds of 50%, 60%, 70%, 80%, and 90%. A parallel clustering effort was performed using mcl (version 14.137) [56] to cluster the results of an all-against-all DIAMOND BLASTp query. Inflation values of 1.2, 1.4, 2.0, 4.0, and 6.0 were used to cluster the output using the ‘--stream-neg-log10′ parameter and transformation function ‘-stream-tf ‘ceil(200)’’ to convert e-values into edge weights for the input graph. Inspection of the outputs using the ‘clm dist’ and ‘clm info’ tools was performed and from the initial clustering results, a value of 4.0 was selected to define protein clusters. A hybrid clustering method was also applied, in which the proteins not successfully clustered by usearch at each identity threshold were then clustered by mcl. The resulting protein clusters were appended to the original usearch result as supplemental protein clusters.
To reduce the number of candidate protein clusters required to produce a comprehensive sampling of diagnostically relevant viral proteins, the output of each clustering iteration was evaluated. Protein clusters were scored in terms of their coverage of viral families and subfamilies using a modified scoring metric developed by Bobay and Ochman [57]. For every (sub)family grouping, each protein cluster was scored according to how many members of the grouping had contributed a protein into the cluster. If a cluster covered a sufficient portion of any (sub)family grouping it was retained as a sufficiently conserved for diagnostic purposes. The threshold for retention was set as:
m i n ( 0.85 1 1 n )
Thus, for large groupings the coverage requirement was 85%. For groupings with fewer than 7 genomes, it is not possible to score greater than 85% unless all members contribute to the cluster, so the threshold was set to retain clusters which were not represented by a single genome in a grouping. A multiple sequence alignment was performed on each remaining cluster using a range of tools, as there is no single alignment tool which consistently outperforms all others [58]. MAFFT (version 7.310) [59,60], MAFFT-L-INS-I [61], Clustal-Omega (version 1.2.4) [62], ClustalW2 (version 2.1) [63], and MUSCLE (version 3.8.31) [64] were applied to each cluster and the alignment which introduced the fewest gaps was selected as the representative alignment. Each sequence alignment was then converted to an HMM profile using hmmbuild (version 3.2.1) [65,66].
The target and non-target protein sequences were classified against each database using hmmsearch. For comparative purposes, we downloaded and tested the following publicly available databases using the same datasets; pVOGs [67] (June 2017 release), VOGDB (https://vogdb.org/, release 99), RVDB (version 18.0) [68], vFAM-A and vFAM-B [69] (February 2014 release), and EggNOG [70] (viral OGs, October 2015 release). We also downloaded the classification databases used in viral detection tools IMG/VR (version 3.0) [71], CheckV, (version 0.6) [72], and VirTool (https://www.virtool.ca/, November 2017 release). Hits were filtered to retain matches with a sequence-level e-value of ≤0.001 for all databases except for CheckV, for which we applied the cluster-specific bitscore thresholds specified in the documentation.

2.4. Testing of Sequence-Based Classification Tools

The dereplicated nucleotide and protein databases were tested for sensitivity and specificity. We defined sensitivity as the proportion of viral sequences which were correctly identified as such, and specificity as the proportion of the nontarget sequences correctly rejected by the classifier following appropriate filtering for each nontarget lineage (Table 1). For alignment-based, nucleotide/nucleotide classification we tested BLASTn (version 2.10.0) [73,74], usearch, vsearch (version 2.15.1) [75], and MMSeqs2 (version 12.113e3). Nucleotide/protein (translated) searches were performed using DIAMOND and MMSeqs2, and protein/protein searching was performed using BLASTp, DIAMOND, MMSeqs2, and usearch with both the usearch_global and ublast algorithms. Nucleotide/nucleotide and protein/protein searches required 70% identity over 50% of the query length and an e-value of less than 0.001 to be accepted as valid matches. Translated searches were performed using the same identity and e-value requirements, but without a minimum coverage value to account for instances in which open reading frames may be identified within a subset of a nucleotide sequence (i.e., where a read spans the start or stop boundary of the transcript). For nucleotide/nucleotide searches, query contigs were broken on poly-N scaffolding regions to prevent the interference of artificially induced ambiguous regions on alignment scores.
Nucleotide classification via k-mer frequency profiles was performed using Kraken2 (version 2.1.1) [76], KrakenUniq (version 0.5.7) [77], CLARK (version 1.2.6.1) [78], MetaCache (version 1.1.1) [79], and ganon (version 0.3.4) [80]. For Kraken2 the pre-formatted k2_standard_20201202 and k2_viral_20201202 databases were downloaded for classification. For all other tools, the tool-specific database creation commands were used to build the databases. Downloads for database construction commenced on the 23rd of December 2020, corresponding to release 203 of GenBank.
Pipelines and software suites developed for the detection of viral sequences in HTS libraries were identified through a literature review and tested against the target dataset. The tools installed were as follows: MiCOP (unversioned) [41], FastViromeExplorer (unversioned) [40], ViromeScan (not versioned) [39], VirFinder (version 1.1) [48], VirSorter (version 1.0.6) [46], VirSorter2 (version 2.1) [47], VirTool (https://www.virtool.ca/, version 3.9.6), and DeepVirFinder (version 1.0) [49]. All software was downloaded from the respective sources in January 2021, so for tools which do not publish with version information the current release at that time point is the version used in this analysis.
Each tool was used to classify the target and non-target datasets using their default detection thresholds. When no score cut-off was specified, the highest ranked hit per query sequence was retained as the result of classification. Multiple testing correction was not performed DeepVirFinder predictions, as the target dataset is exclusively viral sequences and correcting for multiple testing would suppress legitimate identifications. For nucleotide analysis with VirTool, mapping was performed using bwa (version 0.7.17) with the ‘-x intractg’ flag because long read mapping is not recommended by bowtie2, the tool used internally by the VirTool workflow.

2.5. Construction of Mock Libraries for Workflow Evaluation

A selection of plant and vertebrate viruses deemed of high importance to the NZ agricultural and horticultural sectors were identified. A total of 156 viruses comprising both RNA and DNA genomes were selected, and genomes from 16 host species were downloaded to simulate libraries of infected host material (Table S1). To model the skew of transcript abundance observed in RNA sequencing data, a negative binomial distribution of transcript abundance was produced by mapping ONT sequence data obtained during diagnostic sequencing against the NCBI tomato transcript reference (version GCF_000188115.4). The reference transcriptome was clustered using CD-HIT-EST (version 4.8.1) [81] with the parameters ‘-c 0.95 -n 10 -d 0 -T 10’ then ONT sequencing data produced from a previously published library preparation technique [12] were then mapped to the transcript reference using minimap2 (version 2.18) [82] using the ‘map-ont’ method. A negative binomial distribution of transcript coverage was estimated using maximum likelihood estimate of parameters (n = 0.052, p = 0.011) with the SciPy statistical package [83]. To create simulated libraries, the transcriptome reference sequences were obtained for each host species which were similarly dereplicated using CD-HIT-EST. Five randomly sampled transcript abundance libraries were created using the negative binomial model, and randomly assigned to each virus of that host.
Mock sequencing libraries were constructed to mimic the results of the ONT MinION sequencing platform using BadRead (version 0.2.0) [84] with the ‘nanopore2020′ error and q-score models. For DNA viruses and host genomes, libraries were simulated using default parameters. For RNA viruses and transcriptomes, one of the five transcript abundance libraries were sampled at random and BadReads were run using the parameters ‘--chimeras 0 --length 250,180′. Length parameters correspond to the mean and standard deviation of sequence lengths in the tomato sample used in mapping above. All libraries were simulated to produce at least 1 million host reads and 10,000 virus reads. Illumina HiSeq library simulation was performed using the same reference data as for ONT, using InSilicoSeq (version 1.5.2) [85] with the default HiSeq error model to produce 100 million paired end reads per host and 50,000 per virus.
For each virus/host pairing, the ONT library size was set as 400,000 sequence reads, approximately equivalent to the result of a 24 h ONT Flongle run [12]. For each ONT library the composition of viral reads was randomly sampled as an absolute value from a normal distribution with a mean of 0.06 and standard deviation of 1.1. These parameters were obtained by fitting a distribution to the percentage of viral reads identified in virus-positive plant material diagnosed at the Plant Health and Environment Laboratory during 2021 with the method described by Liefting et al. (2021). For HiSeq libraries, a sequencing depth of 40 million paired-end reads was selected. The same distribution was used to generate viral abundances, but values were reduced by a factor of 10 to simulate a lower viral titre. As an additional means of comparison, 18 datasets from the publicly available VIROMOCK challenge (https://gitlab.com/ilvo/VIROMOCKchallenge, accessed 05/08/2021) were downloaded and tested to assess the robustness of the workflows.

2.6. Application of a Diagnostic Workflows to Mock Libraries

The top performing classification tools were selected from the results of sensitivity and specificity testing based upon these two factors, and their average run duration. This selection of tools comprised BLASTn for nucleotide alignment search, DIAMOND (BLASTx) for translated nucleotide alignment search, DIAMOND (BLASTp) and usearch (usearch_global) for protein alignment search, and Kraken2 for k-mer frequency profiling. None of the virus detection pipelines performed sufficiently well to be considered for further testing (see Results below). To simulate a complete diagnostic event, a series of analysis options for each library were arranged to perform the following steps upon the raw libraries: quality filtering and adapter removal (mandatory), host mapping and removal of host reads (optional), assembly (Illumina only), and protein coding sequence prediction prior to classification of reads (Figure 1).
When working with ONT sequence data, quality filtering was performed with NanoFilt (version 2.6.0) [86] and adapter removal with Porechop (version 0.2.4) [87]. Sequences were mapped to the host genome using minimap2 (version 2.20) with the ‘map-ont’ or ‘splice’ settings for DNA and RNA libraries, respectively. For predicting protein coding regions for protein annotation, the tools prodigal (version 2.6.3) [88] and TransDecoder (https://github.com/TransDecoder, version 5.5.0) were tested. Prodigal was run using anonymous (metagenomic) mode, to avoid training sample-specific gene models. TransDecoder was applied in two ways, using the basic pairing of the TransDecoder.LongOrfs and TransDecoder.Predict functions (hereafter referred to as TransDecoder Base) and with the use of homology searches to guide ORF retention between the LongOrfs and Predict functions (TransDecoder Guided). For guided prediction, the initial set of candidate peptides identified by TransDecoder.LongOrfs were annotated using hmmsearch (v3.3) against the Pfam (version 32.0) [89] database and DIAMOND BLASTp (v2.0.6) against the SwissProt (release 2020_06) [90] database prior to prediction.
Quality filtering and adapter removal were performed for simulated HiSeq and VIROMOCK libraries using fastp (version 0.20.0) [91]. Host sequence removal was performed by mapping reads to the reference genome using bowtie2 (version 2.4.4) for DNA and HISAT2 (version 2.2.1) for RNA libraries [92,93]. Successfully mapped reads were identified and removed using samtools (version 1.12). Assembly was performed with three different assembly tools for each library type, selected for their documented strengths when assembling metagenomic datasets [94]. DNA libraries were assembled using MegaHIT (version 1.1.4) [95], and the metaSPAdes and metaviralSPAdes routines from the SPAdes assembler (version 3.15.0) [96,97,98]. For RNA libraries MegaHIT, rnaSPAdes [99], and rnaviralSPAdes [100] were applied. Gene prediction was performed using the same tools and parameters as the ONT libraries.
Each traversal from raw sequence to results (Figure 1) is hereafter referred to as an analysis iteration. Each analysis iteration was assessed in terms of each tools ability to recover the exact strain and species of the virus spiked into each library.

3. Results and Discussion

3.1. Creation of Target and Non-Target Datasets for Tool Testing

A total of 26,206 viral genomes were obtained from NCBI, consisting of 9011 from RefSeq and 17,195 from GenBank. When accounting for entries which were duplicated between repositories, a final count of 17,876 genomes from 17,231 viral species were obtained, which included 47 viroid genomes. Host information from NCBI was used to organise the viruses and viroids into broad host categories; 1992 plant, 2494 vertebrate, 1931 invertebrate, 313 fungus, 72 protozoan viruses, and 47 plant viroids. The remaining genomes were collapsed to the category of ‘other’ host consisting of viruses infecting algae, diatoms, prokaryotes, and viruses for which host information could not be established. From these viruses, 27,340 contigs were extracted and dereplicated into 27,029 unique sequences with seqmagick (https://fhcrc.github.io/seqmagick/, version 0.7.0). Protein coding sequences were extracted from the GenBank files to produce an initial dataset of 1,088,425 amino acid sequences which were similarly dereplicated to 686,559 unique sequences. Nucleotide sequences for the non-target dataset were extracted and reduced to unique contigs, and non-target protein sequences were dereplicated by clustering with the MMSeqs2 ‘linclust’ function using thresholds of 80% identity over 90% alignment coverage for prokaryotic proteins, and 95% identity for all other lineages (Table 1).

3.2. Creation of Custom Sequence Databases

In addition to the genomes obtained when constructing the target dataset, a total of 3,308,823 viral nucleic acid sequences were obtained from the NCBI nt database, along with 8,160,005 protein sequences from nr. These reference sets of viruses were filtered and dereplicated into candidate classification databases. The number of lineage clusters and representative sequences for each database and the working title for each candidate database can be found in Table 2.

3.3. Creation and Assessment of Custom Viral HMM Databases

A total of 12 candidate viral HMM databases were created from the target dataset. Following deduplication and removal of polyprotein sequences, 652,996 sequences remained for cluster generation. Clustering of these sequences was performed using usearch (50% to 90% identity thresholds, 10% increments) and mcl (inflation parameter 1.4 and 4.0). Five additional cluster sets were created by combining these techniques, applying mcl clustering (I = 4.0) to the proteins which were not clustered in each usearch iteration. Clustering methods yielded between 130,000 and 550,000 sequence clusters depending on the method used, although once singleton clusters and clusters with insufficient viral coverage (see Section 2.3) were removed from the result this number was significantly reduced (Table 3).
For each cluster which passed coverage filtering, a multiple sequence alignment was performed using four different tools, with two alignment algorithms applied using MAFFT. The alignment which introduced the fewest gaps per cluster was selected as the ‘optimal’ alignment and an HMM profile was generated from the alignment using hmmbuild. The complete set of HMMs for each clustering approach were then pooled and a complete database produced using the hmmpress. To assess the sensitivity of each method, the target dataset was classified using each candidate HMM database. As the cluster selection method deliberately discarded many clusters of viral protein, sensitivity was scored at the genome level rather than the level of protein. Specifically, we counted the number of viral genomes from which at least one protein was identified (Table S2).
Raw sensitivity values were highest across mcl and mixed clustering methods but promising from all approaches. Four candidate databases were selected for further evaluation—usearch_50, mcl_1.4, mcl_4.0, and mixed_90. Specificity scores were assessed at the level of protein coverage, rather than genome coverage, and results were lower than the sensitivity scores across all databases (Table S2). This finding is not surprising, as the HMM approach is a method aimed at recovering ancestral relationships which can capture protein motifs and signatures which are distantly related through evolutionary time or may have been the result of convergent evolution. This functionality is desirable in the context of our workflow, as we envision HMM searching as a secondary means of analysis to identity uncharacterised and new-to-science viral proteins rather than a means for obtaining exact matches to characterised pathogens.
Between the four candidates there was no clearly superior clustering method for producing the final classification database. Each technique produced variable results in terms of sensitivity and specificity, but not surprisingly these values were negatively correlated—clusters which were able to capture a larger variety of viral proteins produced a greater number of false positive hits amongst non-target proteins. We ultimately favoured databases with higher sensitivity as results of these searches would always be validated, whereas poor sensitivity will result in false negative results which are unlikely to be further investigated. Between the options tested both the mixed_90 and mcl_4.0 databases demonstrated the highest sensitivity for plant and vertebrate viruses (Table 4). Between these two options, average sensitivity was equivalent when e-value filtering was applied to the results and we selected the mcl_4.0 database due to its slightly higher specificity and drastically shorter run times.

3.4. Classification and Database Profiling

Classification of nucleotide sequence data was performed using alignment-based and k-mer frequency profiling tools, as well as a set of software tools designed specifically for identifying viral sequences in metagenomic datasets. The sensitivity of each tool was tested against the target dataset and specificity was assessed using the non-target data (Figure 2, Table S3). For the alignment-based tools, the BLASTn tool provided the greatest sensitivity and specificity, returning near-perfect coverage of the test datasets. A strong result was also observed with the MMSeqs2 easy-search functionality. Classifications with both usearch and vsearch were unable to complete, as run times when classifying with these tools were so prohibitively long that it was not deemed feasible to utilise them in a practical setting. We speculate that the reason for these poor times is a consequence of the indexing process performed by these tools, as both were originally developed for working with amplicon sequencing datasets in which databases are generally smaller and much more homogeneous. We did observe that usearch performed extremely well when classifying protein sequence classifications, where a more heterogeneous population of target sequences would be expected. Classification of plant virus sequences using k-mer profiling tools was generally successful across all tools tested, with a median sensitivity of 0.9584. However, sensitivity towards vertebrate viruses was notably lower, ranging from 0.753 to 0.866 (Figure 2). Specificity values were more variable, but Kraken2 was a clear leader with an average specificity score of 0.9996 when using the standard database against both plant and vertebrate viruses (Figure 2, Table S3).
For the sequence alignment and k-mer tools the time needed to complete analysis was also considered. Although time is not necessarily the most important factor when performing an analysis, it is still a consideration due to reporting windows. Run times were low for most k-mer classification tools, typically requiring less than 1 min to classify the target dataset (Table S4). Surprisingly, BLASTn proved quicker in performing classification than MMSeqs2, requiring as little as half the time to classify the target dataset (Table S4). Both usearch and vsearch performed so poorly under the time metric that they were discarded from the nucleotide classification section of our analysis as they would not be feasible in a real diagnostic setting.
In addition to standalone classification tools, a series of software pipelines designed for the identification of viral sequences from HTS libraries were trialled. Across the board, these tools performed poorly for sensitivity of plant and animal viruses, with only five out of 13 tool/database combinations scoring a sensitivity greater than 0.8 for either plant or vertebrate viruses (Table 5). We believe that this poor sensitivity is a consequence of the large number of phage sequences used in their development, either by design or as a consequence of NCBI database skew. For example, DeepVirFinder achieved high sensitivity for the ‘other’ group (Table 5) which included prokaryote-infecting viruses despite performing poorly in other lineages.
Protein classification was performed using both alignment-based search tools and HMM searches using hmmer. Sensitivity of all alignment-based search tools was high, with over 98% of viral proteins correctly identified as viral across all host types (Figure 3, Table S3). Translated searches were slightly less sensitive when considering plant viruses but performed equally well as protein/protein searches for vertebrate viruses.
Based on these findings we identified a trusted set of tools and databases which provided the best trade-off between sensitivity, specificity, and computational run time. We then sought to test these tools against noisier datasets, representative of what would likely be encountered during real diagnostic settings.

3.5. Recovery of Nucleotide and Gene Sequences from Simulated Libraries

Mock infections were simulated for 24 DNA virus infections (20 vertebrate, 4 plant) and 141 RNA virus infections (46 vertebrate, 95 plant). The selection of viruses was chosen to include viruses which are routinely identified in the Ministry for Primary Industries virology laboratories and emerging viruses which may prove to be of economic importance in the future. For each infection, sequences representing the outcome of Oxford Nanopore Technologies MinION and Illumina HiSeq platform were simulated. Mock ONT libraries were created at a depth of 400,000 sequences, mainly from the host organisms, and HiSeq libraries contained 40 million sequences, with a viral load approximately 10% that of the ONT libraries. These values were selected to reflect our current practice of performing ONT sequencing exclusively on symptomatic plant material, whereas Illumina sequencing is favoured when virus symptoms are not always apparent.
Each mock library was analysed through a variety of analysis pathways (Figure 1) which comprised the removal of low-quality sequences, followed by optional removal of host sequences using a workflow appropriate for the library. For HiSeq libraries, assembly was performed using MegaHIT and SPAdes [94]. DNA libraries were assembled using the metaSPAdes and metaviralSPAdes routines, and RNA libraries were assembled using rnaSPAdes and rnaviralSPAdes. When working with VIROMOCK libraries, the Illumina workflow was followed without host removal, as not all samples in the dataset contain host sequence. Assembly was not performed with ONT libraries as, to our knowledge, there is currently no de novo metatranscriptome assembler available. While we have successfully assembled cDNA Oxford Nanopore datasets to produce viral genomes [12] the performance of any given tool is variable when working with such data, and this instability does not lend itself well to a hands-off approach. As the primary objective of this work is to identify methods for the identification of viral sequences, not to produce representative genomes, we opted to skip a general-purpose assembly step for the ONT libraries. For these samples, fasta sequences were produced directly from the quality-filtered sequences using seqmagick.
To perform protein sequence analysis, prediction of open reading frames in nucleic acid sequences and subsequent translation into an amino acid sequence is required. However, in a diagnostic setting the organism to be detected is not known in advance and therefore a tailored gene prediction model cannot be applied to the data. While much work has been done to address this issue in the field of environmental metagenomics and prediction of coding sequences in prokaryotic lineages, viruses exhibit features not always encountered in prokaryotes, such as non-canonical start codons, readthrough transcription, leaky-scanning, and in some cases gene splicing [101,102,103,104,105,106,107]. We therefore compared prodigal, a tool designed for ab initio prediction of coding sequence in metagenomic datasets, with TransDecoder, designed for prediction of coding sequences with RNA-Seq data (Figure 1).
When assessing the results of the HiSeq libraries, it was quickly apparent that not all assembly options were equally robust against variations in the input data. The SPAdes assembler occasionally failed to produce an assembly, but this was particularly noticeable in the case of the metaviral assembly routine, which did not produce a successful assembly for DNA viruses when the host sequence was removed prior to assembly (Table 6). Similarly, protein coding prediction was not equally robust across all methods as only prodigal was able to produce protein coding sequence predictions from all input libraries (Table 6). We observed that prediction of protein coding sequences by TransDecoder suffered when host sequences were removed from Illumina assemblies. We attribute this observation to the TransDecoder algorithm attempting to generate a training model from the input data [108], as it is likely that when there is insufficient sequence data to produce initial models the subsequent prediction will suffer. In instances where assembly and gene calling were successful the results were still evaluated, but the less reliable performance of these tools meant that results were primarily focused on the prodigal predictions.

3.6. Performance of Nucleotide-Based Classification Approaches

Following the analysis of each mock infection through each analysis pathway the results were screened to determine whether sequences from the library were correctly identified as belonging to the strain and species of the spiked virus. Classification of the nucleotide sequences was performed using BLASTn with each database described in Table 2, Kraken2 with both databases, and DIAMOND BLASTx using the protein databases tested in Table 2.
Amongst nucleotide classification methods we noted that performance across the five in-house databases was not equal. For the ONT dataset, the nt_derep4 candidate database correctly identified all DNA viruses and 138 of 141 RNA viruses at the strain level but at the species level, the nt_derep1, nt_derep2, and nt_derep3 databases all achieved 100% sensitivity across RNA virus samples. The sensitivity of the ONT method was not affected by removal of host sequences but amongst the HiSeq libraries performance was variable and clearly influenced by host removal. Despite its superior performance in terms of completing assemblies, MegaHIT performed slightly poorer than SPAdes when run in RNA mode for identifying the exact strain of virus, although for species-level identification both tools performed equally well. MegaHIT performed marginally better than SPAdes (metagenomic mode) when assessing DNA samples at the strain level but following the removal of host sequences this result was reversed and metaSPAdes was able to perform equally well or better depending on the taxonomic resolution (Figure 4). We therefore tentatively conclude that for assessing DNA viruses MegaHIT is the superior option, as is SPAdes in transcriptomic mode for RNA viruses, but the performance of both tools is extremely close, and the effect of database and classification tool plays a far greater role in determining sensitivity.
K-mer profiling classification did not perform as well in simulated data as the initial comparisons against the target dataset would suggest. Across all library combinations, Kraken2 demonstrated a median sensitivity of 81.7% for strain-level and 89.6% for species-level detections (Figure 4). These values were comparable to the classification sensitivity of the BLASTn genome_nt database (81.2% and 91.0%, respectively) but inferior to results obtained by the more comprehensive versions derived from the NCBI nt database. This lower sensitivity may be a consequence of the Kraken2 and genome_nt databases only including representative genome sequences, as new and emerging sequences are typically first detected as amplicons and partial sequences.
Translated protein queries, performed via the DIAMOND BLASTx method, demonstrated variable performance for DNA virus samples when compared with direct nucleotide classification. Again, regardless of host sequence retention MegaHIT yielded slightly higher sensitivity than metaSPAdes assemblies, but differences were overall minor and more frequent in strain than species matching. For assessing RNA viruses, we first discounted two samples which were spiked with the viroids Peach latent mosaic viroid (PLMVd) and Hop stunt viroid (HSVd) as these entities do not encode protein sequences and therefore could not be detected with protein matching searches. The median sensitivity for nucleic acid databases ranged from 81.6% to 98.9% for strain matches and 91.0% to 99.6% for species-level matches (Figure 4). By comparison, after accounting for the viroid samples, strain-level sensitivity for BLASTx searches was 81.4% for the genome_prot database and 90.9% for nr_derep1 and nr_derep2 for strain matches, and 90.9%, 99.6%, and 98.9% for species matching (Figure 4). We attribute the finding of equivalent species sensitivity but inferior strain sensitivity to the wider range of target sequences reported by the BLASTx search, which likely fulfilled the maximum target sequence requirement of the BLASTx search before the correct strain targets could be encountered.

3.7. Performance of Protein–Protein Classification Approaches

Instead of BLASTx searching, an alternative method for acquiring rigorous protein annotations is to perform the translation step independently of the classification process to make more explicit use of the known features of protein-coding regions to refine the query sequences. Initial comparison between prodigal and TransDecoder revealed that a more stable gene prediction outcome across samples was produced by prodigal (Table 6). Protein classification was performed for each set of predicted proteins using usearch and DIAMOND BLASTp against the protein databases (Table 2), and HMM searching was performed against the mcl_4.0 database.
At the species level, protein predictions from DNA and RNA viruses (excluding viroids) were sufficient to produce 100% sensitivity under most assembly and host removal conditions (Figure 5). We observed slightly greater sensitivity when using DIAMOND compared with usearch for DNA virus classification, but results were identical for RNA viruses. Sensitivities were poorer when examining the predictions made by TransDecoder even when accounting for the number of samples from which protein-coding sequences were successfully produced (Tables S5 and S6).
HMM searches against the mcl_4.0 database yielded strong detection sensitivity, especially when considering that the protein models were produced from only the genome_prot sequences and lacks most of the sequences in the nr_derep1 and nr_derep2 databases (Table 2). The mcl_4.0 database outperformed all other protein databases at the strain level (Figure 5) and performed markedly better than the genome_prot database, from which it was generated, at the species level (median sensitivity 98.2% versus ~90.5%). At the resolution of species, the nr_derep1 and nr_derep2 databases both outperformed mcl_4.0 by a small margin, with a sensitivity of 99.3%. However, it is worth noting that the mcl_4.0 database alone cannot predict the species of the proteins detected by the models and a secondary BLASTp was required to assign taxonomy to the proteins identified by the HMM approach. Due to the need for this secondary step, we do not believe that the HMM approach would be ideal for routine diagnostic cases, but it’s superior ability to capture viral protein sequences which are not already present in the database, as shown by the improved sensitivity when compared with the genome_prot database (Figure 5), demonstrates the utility of this approach for detection of novel viral sequences.

3.8. Performance of Analysis Workflows against VIROMOCK Data

To further validate our results, we obtained 18 datasets from the VIROMOCK challenge (https://gitlab.com/ilvo/VIROMOCKchallenge, accessed 05/08/2021). These data comprise a mixture of real, synthetic, and artificially spiked sequence libraries to provide an independent test panel for the development of bioinformatic pipelines for virus detection. All datasets were downloaded, quality-filtered, and assembled using MegaHIT, SPAdes, and the appropriate viral SPAdes workflow for the viruses in the dataset. This yielded three assemblies for each dataset, except for dataset 4 which contains a mixture of DNA and RNA viruses, so was assembled using MegaHIT, metaSPAdes, rnaSPAdes, and rnaviralSPAdes. Each set of assembled contigs were then classified using all nucleic acid methods, and coding sequence predictions were made with prodigal for protein annotation. Detection of the expected viruses was achieved in most cases for at least one assembly option (Table S7) although as with the simulated libraries, there was no clearly superior assembly tool, with both MegaHIT and rnaSPAdes achieving greater sensitivity in different datasets.
With the previously noted exception of the protein annotation methods being unable to detect viroid sequences, there were only two instances where a detection could not be made. In the case of dataset 1, we were unable to identify Apple hammerhead viroid (AHVd) sequences in any assembly with any classification tool (Table S7). However, we note that the documentation for this dataset contains a disclaimer that the presence of this viroid could not be validated by qPCR and therefore may be the result of contamination. Investigation of the unassembled reads revealed just a single read pair which could be classified as AHVd. Due to this finding, and the fact that this dataset is from Citrus × sinensis and AHVd has an extremely limited host range [109,110], we deem this unlikely to be a true instance of the viroid. For dataset 9 we were unable to produce a strain-level match for the Pistacia emaravirus B, but we did identify sequences classifying to the genus Emaravirus within our assemblies. Under standard diagnostic settings such a finding would prompt further investigation, so we considered this a successful detection event although it did not meet the exact requirement of the challenge.

3.9. Challenges in Creating a Stand-Alone Analysis Pipeline

Despite its clear value in the field of biosecurity, there are still no internationally standardised pipelines for the detection of viruses using high-throughput sequencing. This challenge stems primarily from the difficulty in determining a ‘best’ tool for each stage of analysis. Variation is one of the foundational concepts of biology and a consequence of genetic variation within and between species means that there is typically no single best tool for a job, but a set of tools which outperform each other under different conditions. We therefore considered not only the ability of our selected tools to achieve accurate results, but also how robust they were to different starting libraries to ensure that our final workflow could be universally applied to diagnostic samples without the need to customise analysis parameters or repeat analysis steps due to computational crashes.
There is also a need to validate HTS-based detections, as false-positive results are possible when such large numbers of sequence data are surveyed, particularly when using reduced-representation classification approaches. Within our workflow there was no need for performing downstream validation as the viral species within each sample was known in advance. In application, however, some form of verification would be required following from the initial detection, either purely bioinformatic (i.e., recovering a complete genome) or using traditional diagnostic assays to provide an independent observation of the virus in question.

4. Conclusions

We assessed a total of 60 tool and database combinations, either sourced from the published literature or developed in house, for their efficacy as methods for the detection of plant and animal viruses. Despite the development of a plethora of tools for the express purpose of recovering viral sequences from metagenomic and RNAseq datasets we found that more generic alignment and k-mer based classification approaches provided superior sensitivity. We attempted to improve upon the speed required for classification with these tools by creating more streamlined virus databases for classification. To benchmark the performance of candidate databases and classification approaches we tested nearly 350 mock datasets using a combination of assembly, gene prediction, and classification approaches to establish an end-to-end analysis workflow which is rapid to run and robust across library preparations. Our work provides a strong foundation for the creation of an automated analysis pipeline and will be of value as HTS-based diagnostics become increasingly common in the biosecurity field.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/v14102163/s1, Table S1: Host genome identity and virus sequence(s) used to construct simulated mock libraries for Oxford Nanopore Technology and Illumina HiSeq sequencing. Table S2: Sensitivity values for each candidate database, as determined by their coverage of the target dataset genomes. Table S3: Sensitivity scores for viruses infecting plant, vertebrate, invertebrate, and other (protozoa, fungi, prokaryote) organisms, and specificity scores for plant- and vertebrate-infecting viruses. Table S4: Time required to complete classification of the target (viral) dataset for each tool using 10 threads for analysis. Table S5: Successful identifications of spiked viral species in mock libraries when TransDecoder is used to perform protein prediction. Table S6: Successful identifications of spiked viral species in mock libraries when TransDecoder was used to perform protein prediction with Pfam and UniProt annotation of long ORFs was used to guide the prediction process. Table S7: Detection results for each dataset in the VIROMOCK Challenge repository.

Author Contributions

Conceptualization, C.D., D.W.W., L.L., H.J.H.; methodology, D.W.W.; software, D.W.W.; validation, A.C., C.D., D.W.W., J.R.T., L.L.; formal analysis, D.W.W.; resources, C.D., H.J.H., J.R.T.; data curation, D.W.W.; writing—original draft preparation, D.W.W.; writing—review and editing, A.C., C.D., H.J.H., J.R.T., L.L.; funding acquisition, C.D., H.J.H., L.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Ministry for Primary Industries Operational Research Programme (project 406084). This research received no external funding.

Acknowledgments

We are grateful to Dongmei Li and Luciano Rigano for providing additional reviews of this manuscript and suggestions which improved the overall quality of the analysis.

Conflicts of Interest

The authors declare that this work was conducted in the absence of any conflict of interest.

References

  1. An Overview of New Zealand’s Trade in 2021. Available online: https://www.mfat.govt.nz/assets/Trade/MFAT-Market-reports/An-overview-of-New-Zealands-trade-in-2021-March-2021.pdf (accessed on 3 June 2022).
  2. Situation and Outlook for Primary Industries (SOPI) Data. Available online: https://www.mpi.govt.nz/resources-and-forms/economic-intelligence/data/ (accessed on 1 July 2022).
  3. Stone, M. Biosecurity is critical to New Zealand’s national security, economy and way of life. N. Z. Vet. J. 2021, 69, 309–312. [Google Scholar] [CrossRef] [PubMed]
  4. Souiri, A.; Khataby, K.; Kasmi, Y.; Zemzami, M.; Amzazi, S.; Ennaji, M.M. Chapter 16—Risk Assessment and Biosecurity Considerations in Control of Emergent Plant Viruses. In Emerging and Reemerging Viral Pathogens; Academic Press: Cambridge, MA, USA, 2020; pp. 287–311. [Google Scholar]
  5. Foot-and-Mouth Disease Economic Impact Assessment: What it means for New Zealand. Available online: https://www.mpi.govt.nz/dmsdocument/4406-Foot-and-Mouth-Disease-Economic-Impact-Assessment-What-it-means-for-New-Zealand (accessed on 1 July 2022).
  6. Tuppurainen, E.S.M.; Venter, E.H.; Shisler, J.L.; Gari, G.; Mekonnen, G.A.; Juleff, N.; Lyons, N.A.; De Clercq, K.; Upton, C.; Bowden, T.R.; et al. Review: Capripoxvirus Diseases: Current Status and Opportunities for Control. Transbound. Emerg. Dis. 2015, 64, 729–745. [Google Scholar] [CrossRef] [PubMed]
  7. Brown, V.R.; Miller, R.S.; McKee, S.C.; Ernst, K.H.; Didero, N.M.; Maison, R.M.; Grady, M.J.; Shwiff, S.A. Risks of intro-duction and economic consequences associated with African swine fever, classical swine fever and foot-and-mouth disease: A review of the literature. Transbound. Emerg. Dis. 2021, 68, 1910–1965. [Google Scholar] [CrossRef] [PubMed]
  8. Kurian, A.; Hall, W.; Neumann, E. African swine fever: A New Zealand perspective on epidemiological risk factors for its occurrence. N. Z. Vet. J. 2021, 69, 135–146. [Google Scholar] [CrossRef] [PubMed]
  9. Jiang, D.; Ma, T.; Hao, M.; Ding, F.; Sun, K.; Wang, Q.; Kang, T.; Di Wang, D.; Zhao, S.; Li, M.; et al. Quantifying risk factors and potential geographic extent of African swine fever across the world. PLoS ONE 2022, 17, e0267128. [Google Scholar] [CrossRef] [PubMed]
  10. Keeping African Swine Fever Out of Australia. Available online: https://www.agriculture.gov.au/biosecurity-trade/pests-diseases-weeds/animal/asf (accessed on 1 July 2022).
  11. Technical Advice: Risk of Lumpy Skin Disease Via Import of Cattle and Buffalo Meat and Meat Products for Human and Ani-mal Consumption. Available online: https://www.mpi.govt.nz/dmsdocument/51352-Technical-Advice-Risk-of-lumpy-skin-disease-via-import-of-cattle-and-buffalo-meat-and-meat-products-for-human-and-animal-consumption (accessed on 1 July 2022).
  12. Liefting, L.; Waite, D.; Thompson, J. Application of Oxford Nanopore Technology to Plant Virus Detection. Viruses 2021, 13, 1424. [Google Scholar] [CrossRef]
  13. Adams, I.P.; Glover, R.H.; Monger, W.A.; Mumford, R.; Jackeviciene, E.; Navalinskiene, M.; Samuitiene, M.; Boonham, N. Next-generation sequencing and metagenomic analysis: A universal diagnostic tool in plant virology. Mol. Plant Pathol. 2009, 10, 537–545. [Google Scholar] [CrossRef]
  14. Mumford, R.A.; Macarthur, R.; Boonham, N. The role and challenges of new diagnostic technology in plant biosecurity. Food Secur. 2015, 8, 103–109. [Google Scholar] [CrossRef]
  15. Boykin, L.; Ghalab, A.; De Marchi, B.R.; Savill, A.; Wainaina, J.M.; Kinene, T.; Lamb, S.; Rodrigues, M.; Kehoe, M.; Ndunguru, J.; et al. Real time portable genome sequencing for global food security. F1000Research 2018, 7, 1101. [Google Scholar] [CrossRef]
  16. Villamor, D.E.V.; Ho, T.; Al Rwahnih, M.; Martin, R.R.; Tzanetakis, I.E. High Throughput Sequencing for Plant Virus Detection and Discovery. Phytopathology 2019, 109, 716–725. [Google Scholar] [CrossRef]
  17. Hu, Y.; Wilson, S.; Schwessinger, B.; Rathjen, J.P. Blurred lines: Integrating emerging technologies to advance plant biosecurity. Curr. Opin. Plant Biol. 2020, 56, 127–134. [Google Scholar] [CrossRef] [PubMed]
  18. Mehetre, G.; Leo, V.; Singh, G.; Sorokan, A.; Maksimov, I.; Yadav, M.; Upadhyaya, K.; Hashem, A.; Alsaleh, A.; Dawoud, T.; et al. Current Developments and Challenges in Plant Viral Diagnostics: A Systematic Review. Viruses 2021, 13, 412. [Google Scholar] [CrossRef] [PubMed]
  19. Handelsman, J.; Rondon, M.R.; Brady, S.F.; Clardy, J.; Goodman, R.M. Molecular biological access to the chemistry of un-known soil microbes: A new frontier for natural products. Chem. Biol. 1998, 5, R245–R249. [Google Scholar] [CrossRef]
  20. Bowers, R.M.; Kyrpides, N.C.; Stepanauskas, R.; Harmon-Smith, M.; Doud, D.; Reddy, T.B.K.; Schulz, F.; Jarett, J.; Rivers, A.R.; Eloe-Fadrosh, E.A.; et al. Minimum information about a single amplified genome (MISAG) and a meta-genome-assembled genome (MIMAG) of bacteria and archaea. Nat. Biotechnol. 2017, 35, 725–731. [Google Scholar] [CrossRef]
  21. Candresse, T.; Filloux, D.; Muhire, B.; Julian, C.; Galzi, S.; Fort, G.; Bernardo, P.; Daugrois, J.-H.; Fernandez, E.; Martin, D.P.; et al. Appearances Can Be Deceptive: Revealing a Hidden Viral Infection with Deep Sequencing in a Plant Quarantine Context. PLoS ONE 2014, 9, e102945. [Google Scholar] [CrossRef] [PubMed]
  22. Badial, A.B.; Sherman, D.; Stone, A.; Gopakumar, A.; Wilson, V.; Schneider, W.; King, J. Nanopore Sequencing as a Surveillance Tool for Plant Pathogens in Plant and Insect Tissues. Plant Dis. 2018, 102, 1648–1652. [Google Scholar] [CrossRef] [PubMed]
  23. Chalupowicz, L.; Dombrovsky, A.; Gaba, V.; Luria, N.; Reuven, M.; Beerman, A.; Lachman, O.; Dror, O.; Nissan, G.; Manulis-Sasson, S. Diagnosis of plant diseases using the Nanopore sequencing platform. Plant Pathol. 2018, 68, 229–238. [Google Scholar] [CrossRef]
  24. Filloux, D.; Fernandez, E.; Loire, E.; Claude, L.; Galzi, S.; Candresse, T.; Winter, S.; Jeeva, M.L.; Makeshkumar, T.; Martin, D.P.; et al. Nanopore-based detection and characterization of yam viruses. Sci. Rep. 2018, 8, 1–11. [Google Scholar] [CrossRef]
  25. Fellers, J.P.; Webb, C.; Fellers, M.C.; Rupp, J.S.; De Wolf, E. Wheat Virus Identification within Infected Tissue Using Nanopore Sequencing Technology. Plant Dis. 2019, 103, 2199–2203. [Google Scholar] [CrossRef]
  26. Della Bartola, M.; Byrne, S.; Mullins, E. Characterization of Potato Virus Y Isolates and Assessment of Nanopore Sequencing to Detect and Genotype Potato Viruses. Viruses 2020, 12, 478. [Google Scholar] [CrossRef] [Green Version]
  27. Monger, W.A.; Goodfellow, H.A.; Back, E.J. First report of Arabis mosaic virus in potato (Solanum tuberosum), identified by nanopore sequencing. New Dis. Rep. 2020, 41, 29. [Google Scholar] [CrossRef]
  28. Ziebell, H.; De Jonghe, K.; Rott, M.; Nicolaisen, M.; Gentit, P.; Renvoise, J.-P.; Candresse, T.; Fox, A.; Varveri, C.; Melika, G.; et al. The application of Next-Generation Sequencing technology for the detection and diagnosis of non-culturable organisms: Viruses and viroids (NGSdetect). Zenodo 2021. [Google Scholar] [CrossRef]
  29. Nooij, S.; Schmitz, D.; Vennema, H.; Kroneman, A.; Koopmans, M. Overview of Virus Metagenomic Classification Methods and Their Biological Applications. Front. Microbiol. 2018, 9, 749. [Google Scholar] [CrossRef]
  30. Wilson, M.R.; Sample, H.A.; Zorn, K.C.; Arevalo, S.; Yu, G.; Neuhaus, J.; Federman, S.; Stryke, D.; Briggs, B.; Langelier, C. Clinical Metagenomic Sequencing for Di-agnosis of Meningitis and Encephalitis. N. Engl. J. Med. 2019, 380, 2327–2340. [Google Scholar] [CrossRef] [PubMed]
  31. de Vries, J.J.C.; Brown, J.R.; Fischer, N.; Sidorov, I.A.; Morfopoulou, S.; Huang, J.; Munnink, B.B.O.; Sayiner, A.; Bulgurcu, A.; Rodriguez, C.; et al. Benchmark of thirteen bioinformatic pipelines for metagenomic virus diagnostics using datasets from clinical samples. J. Clin. Virol. 2021, 141, 104908. [Google Scholar] [CrossRef]
  32. Carbo, E.C.; Sidorov, I.A.; van Rijn-Klink, A.L.; Pappas, N.; van Boheemen, S.; Mei, H.; Hiemstra, P.S.; Eagan, T.M.; Claas, E.C.J.; Kroes, A.C.M.; et al. Performance of Five Metagenomic Classifiers for Virus Pathogen Detection Using Respiratory Samples from a Clinical Cohort. Pathogens 2022, 11, 340. [Google Scholar] [CrossRef]
  33. Ibaba, J.D.; Gubba, A. High-Throughput Sequencing Application in the Diagnosis and Discovery of Plant-Infecting Viruses in Africa, A Decade Later. Plants 2020, 9, 1376. [Google Scholar] [CrossRef]
  34. Menzel, P.; Ng, K.L.; Krogh, A. Fast and sensitive taxonomic classification for metagenomics with Kaiju. Nat. Commun. 2016, 7, 11257. [Google Scholar] [CrossRef]
  35. Jones, S.; Baizan-Edge, A.; MacFarlane, S.; Torrance, L. Viral Diagnostics in Plants Using Next Generation Sequencing: Computational Analysis in Practice. Front. Plant Sci. 2017, 8, 1770. [Google Scholar] [CrossRef]
  36. Susi, H.; Filloux, D.; Frilander, M.J.; Roumagnac, P.; Laine, A.-L. Diverse and variable virus communities in wild plant populations revealed by metagenomic tools. PeerJ 2019, 7, e6140. [Google Scholar] [CrossRef]
  37. Edgar, R.C.; Taylor, J.; Lin, V.; Altman, T.; Barbera, P.; Meleshko, D.; Lohr, D.; Novakovsky, G.; Buchfink, B.; Al-Shayeb, B.; et al. Petabase-scale sequence alignment catalyses viral discovery. Nature 2022, 602, 142–147. [Google Scholar] [CrossRef] [PubMed]
  38. Gauthier, M.A.; Lelwala, R.V.; Elliott, C.E.; Windell, C.; Fiorito, S.; Dinsdale, A.; Whattam, M.; Pattemore, J.; Barrero, R.A. Side-by-Side Comparison of Post-Entry Quarantine and High Throughput Sequencing Methods for Virus and Viroid Diagnosis. Biology 2022, 11, 263. [Google Scholar] [CrossRef] [PubMed]
  39. Rampelli, S.; Soverini, M.; Turroni, S.; Quercia, S.; Biagi, E.; Brigidi, P.; Candela, M. ViromeScan: A new tool for meta-genomic viral community profiling. BMC Genom. 2016, 17, 165. [Google Scholar] [CrossRef]
  40. Tithi, S.S.; Aylward, F.; Jensen, R.V.; Zhang, L. FastViromeExplorer: A pipeline for virus and phage identification and abundance profiling in metagenomics data. PeerJ 2018, 6, e4227. [Google Scholar] [CrossRef] [PubMed]
  41. LaPierre, N.; Mangul, S.; Alser, M.; Mandric, I.; Wu, N.C.; Koslicki, D.; Eskin, E. MiCoP: Microbial community profiling method for detecting viral and fungal organisms in metagenomic samples. BMC Genom. 2019, 20, 423. [Google Scholar] [CrossRef] [PubMed]
  42. Laffy, P.W.; Wood-Charlson, E.M.; Turaev, D.; Weynberg, K.; Botté, E.; van Oppen, M.; Webster, N.; Rattei, T. HoloVir: A Workflow for Investigating the Diversity and Function of Viruses in Invertebrate Holobionts. Front. Microbiol. 2016, 7, 822. [Google Scholar] [CrossRef]
  43. Garretto, A.; Hatzopoulos, T.; Putonti, C. virMine: Automated detection of viral sequences from complex metagenomic samples. PeerJ 2019, 7, e6695. [Google Scholar] [CrossRef]
  44. Lefebvre, M.; Theil, S.; Ma, Y.; Candresse, T. The VirAnnot Pipeline: A Resource for Automated Viral Diversity Estimation and Operational Taxonomy Units Assignation for Virome Sequencing Data. Phytobiomes J. 2019, 3, 256–259. [Google Scholar] [CrossRef]
  45. Ho, T.; Tzanetakis, I.E. Development of a virus detection and discovery pipeline using next generation sequencing. Virology 2014, 471–473, 54–60. [Google Scholar] [CrossRef]
  46. Roux, S.; Enault, F.; Hurwitz, B.L.; Sullivan, M.B. VirSorter: Mining viral signal from microbial genomic data. PeerJ 2015, 3, e985. [Google Scholar] [CrossRef]
  47. Guo, J.; Bolduc, B.; Zayed, A.A.; Varsani, A.; Dominguez-Huerta, G.; Delmont, T.O.; Pratama, A.A.; Gazitúa, M.C.; Vik, D.; Sullivan, M.B.; et al. VirSorter2: A multi-classifier, expert-guided approach to detect diverse DNA and RNA viruses. Microbiome 2021, 9, 1–13. [Google Scholar] [CrossRef] [PubMed]
  48. Ren, J.; Ahlgren, N.A.; Lu, Y.Y.; Fuhrman, J.; Sun, F. VirFinder: A novel k-mer based tool for identifying viral sequences from assembled metagenomic data. Microbiome 2017, 5, 1–20. [Google Scholar] [CrossRef] [PubMed]
  49. Ren, J.; Song, K.; Deng, C.; Ahlgren, N.A.; Fuhrman, J.A.; Li, Y.; Xie, X.; Poplin, R.; Sun, F. Identifying viruses from meta-genomic data using deep learning. Quant. Biol. 2020, 8, 64–77. [Google Scholar] [CrossRef] [PubMed]
  50. Tamisier, L.; Haegeman, A.; Foucart, Y.; Fouillien, N.; Al Rwahnih, M.; Buzkan, N.; Candresse, T.; Chiumenti, M.; De Jonghe, K.; Lefebvre, M.; et al. Semi-artificial datasets as a resource for validation of bioinformatics pipelines for plant virus detection. Peer Community J. 2021, 1, e53. [Google Scholar] [CrossRef]
  51. Parks, D.H.; Chuvochina, M.; Waite, D.W.; Rinke, C.; Skarshewski, A.; Chaumeil, P.-A.; Hugenholtz, P. A standardized bacterial taxonomy based on genome phylogeny substantially revises the tree of life. Nat. Biotechnol. 2018, 36, 996–1004. [Google Scholar] [CrossRef]
  52. Parks, D.H.; Chuvochina, M.; Rinke, C.; Mussig, A.J.; Chaumeil, P.A.; Hugenholtz, P. GTDB: An ongoing census of bacterial and archaeal diversity through a phylogenetically consistent, rank normalized and complete genome-based taxonomy. Nucleic Acids Res. 2022, 50, D785–D794. [Google Scholar] [CrossRef]
  53. Steinegger, M.; Söding, J. MMseqs2 enables sensitive protein sequence searching for the analysis of massive data sets. Nat. Biotechnol. 2017, 35, 1026–1028. [Google Scholar] [CrossRef]
  54. Edgar, R.C. Search and clustering orders of magnitude faster than BLAST. Bioinformatics 2010, 26, 2460–2461. [Google Scholar] [CrossRef]
  55. Buchfink, B.; Xie, C.; Huson, D.H. Fast and sensitive protein alignment using DIAMOND. Nat. Methods 2014, 12, 59–60. [Google Scholar] [CrossRef]
  56. Enright, A.J.; Van Dongen, S.; Ouzounis, C.A. An efficient algorithm for large-scale detection of protein families. Nucleic Acids Res. 2002, 30, 1575–1584. [Google Scholar] [CrossRef]
  57. Bobay, L.-M.; Ochman, H. Biological species in the viral world. Proc. Natl. Acad. Sci. USA 2018, 115, 6040–6045. [Google Scholar] [CrossRef] [PubMed]
  58. Pervez, M.T.; Babar, M.E.; Nadeem, A.; Aslam, M.; Awan, A.R.; Aslam, N.; Hussain, T.; Naveed, N.; Qadri, S.; Waheed, U.; et al. Evaluating the Accuracy and Efficiency of Multiple Sequence Alignment Methods. Evol. Bioinform. 2014, 10, 205–217. [Google Scholar] [CrossRef] [PubMed]
  59. Katoh, K.; Misawa, K.; Kuma, K.; Miyata, T. MAFFT: A novel method for rapid multiple sequence alignment based on fast Fourier transform. Nucleic Acids Res. 2002, 30, 3059–3066. [Google Scholar] [CrossRef] [PubMed]
  60. Katoh, K.; Standley, D.M. MAFFT Multiple Sequence Alignment Software Version 7: Improvements in Performance and Usability. Mol. Biol. Evol. 2013, 30, 772–780. [Google Scholar] [CrossRef] [PubMed]
  61. Katoh, K.; Kuma, K.; Toh, H.; Miyata, T. MAFFT version 5: Improvement in accuracy of multiple sequence alignment. Nucleic Acids Res. 2005, 33, 511–518. [Google Scholar] [CrossRef] [PubMed]
  62. Sievers, F.; Higgins, D.G. Clustal Omega for making accurate alignments of many protein sequences. Protein Sci. 2017, 27, 135–145. [Google Scholar] [CrossRef] [PubMed]
  63. Larkin, M.A.; Blackshields, G.; Brown, N.P.; Chenna, R.; McGettigan, P.A.; McWilliam, H.; Valentin, F.; Wallace, I.M.; Wilm, A.; Lopez, R.; et al. Clustal W and Clustal X version 2.0. Bioinformatics 2007, 23, 2947–2948. [Google Scholar] [CrossRef]
  64. Edgar, R.C. MUSCLE: Multiple sequence alignment with high accuracy and high throughput. Nucleic Acids Res. 2004, 32, 1792–1797. [Google Scholar] [CrossRef]
  65. Eddy, S.R. Profile hidden Markov models. Bioinformatics 1998, 14, 755–763. [Google Scholar] [CrossRef]
  66. Eddy, S.R. Accelerated Profile HMM Searches. PLOS Comput. Biol. 2011, 7, e1002195. [Google Scholar] [CrossRef] [Green Version]
  67. Grazziotin, A.L.; Koonin, E.V.; Kristensen, D.M. Prokaryotic Virus Orthologous Groups (pVOGs): A resource for comparative genomics and protein family annotation. Nucleic Acids Res 2017, 45, D491–D498. [Google Scholar] [CrossRef] [PubMed]
  68. Bigot, T.; Temmam, S.; Perot, P.; Eloit, M. RVDB-prot, a reference viral protein database and its HMM profiles. F1000Res 2019, 8, 530. [Google Scholar] [CrossRef] [PubMed]
  69. Skewes-Cox, P.; Sharpton, T.J.; Pollard, K.S.; DeRisi, J.L. Profile Hidden Markov Models for the Detection of Viruses within Metagenomic Sequence Data. PLoS ONE 2014, 9, e105067. [Google Scholar] [CrossRef] [PubMed]
  70. Huerta-Cepas, J.; Szklarczyk, D.; Forslund, K.; Cook, H.; Heller, D.; Walter, M.C.; Rattei, T.; Mende, D.R.; Sunagawa, S.; Kuhn, M.; et al. eggNOG 4.5: A hierarchical orthology framework with improved functional annotations for eukaryotic, prokaryotic and viral sequences. Nucleic Acids Res. 2016, 44, D286–D293. [Google Scholar] [CrossRef] [PubMed]
  71. Roux, S.; Páez-Espino, D.; Chen, I.-M.A.; Palaniappan, K.; Ratner, A.; Chu, K.; Reddy, T.B.K.; Nayfach, S.; Schulz, F.; Call, L.; et al. IMG/VR v3: An integrated ecological and evolutionary framework for interrogating genomes of uncultivated viruses. Nucleic Acids Res. 2020, 49, D764–D775. [Google Scholar] [CrossRef]
  72. Nayfach, S.; Camargo, A.P.; Schulz, F.; Eloe-Fadrosh, E.; Roux, S.; Kyrpides, N.C. CheckV assesses the quality and completeness of metagenome-assembled viral genomes. Nat. Biotechnol. 2021, 39, 578–585. [Google Scholar] [CrossRef]
  73. Altschul, S.F.; Gish, W.; Miller, W.; Myers, E.W.; Lipman, D.J. Basic local alignment search tool. J. Mol. Biol. 1990, 215, 403–410. [Google Scholar] [CrossRef]
  74. Camacho, C.; Coulouris, G.; Avagyan, V.; Ma, N.; Papadopoulos, J.; Bealer, K.; Madden, T.L. BLAST+: Architecture and ap-plications. BMC Bioinform. 2009, 10, 421. [Google Scholar] [CrossRef]
  75. Rognes, T.; Flouri, T.; Nichols, B.; Quince, C.; Mahé, F. VSEARCH: A versatile open source tool for metagenomics. PeerJ 2016, 2016, e2584. [Google Scholar] [CrossRef]
  76. Wood, D.E.; Lu, J.; Langmead, B. Improved metagenomic analysis with Kraken 2. Genome Biol. 2019, 20, 257. [Google Scholar] [CrossRef] [Green Version]
  77. Breitwieser, F.P.; Baker, D.N.; Salzberg, S.L. KrakenUniq: Confident and fast metagenomics classification using unique k-mer counts. Genome Biol. 2018, 19, 1–10. [Google Scholar] [CrossRef]
  78. Ounit, R.; Wanamaker, S.; Close, T.J.; Lonardi, S. CLARK: Fast and accurate classification of metagenomic and genomic se-quences using discriminative k-mers. BMC Genom. 2015, 16, 236. [Google Scholar] [CrossRef] [PubMed]
  79. Müller, A.; Hundt, C.; Hildebrandt, A.; Hankeln, T.; Schmidt, B. MetaCache: Context-aware classification of metagenomic reads using minhashing. Bioinformatics 2017, 33, 3740–3748. [Google Scholar] [CrossRef] [PubMed]
  80. Piro, V.C.; Dadi, T.H.; Seiler, E.; Reinert, K.; Renard, B.Y. ganon: Precise metagenomics classification against large and up-to-date sets of reference sequences. Bioinformatics 2020, 36 (Suppl. S1). [Google Scholar] [CrossRef]
  81. Fu, L.; Niu, B.; Zhu, Z.; Wu, S.; Li, W. CD-HIT: Accelerated for clustering the next-generation sequencing data. Bioinformatics 2012, 28, 3150–3152. [Google Scholar] [CrossRef]
  82. Li, H. Minimap2: Pairwise alignment for nucleotide sequences. Bioinformatics 2018, 34, 3094–3100. [Google Scholar] [CrossRef] [PubMed]
  83. Virtanen, P.; Gommers, R.; Oliphant, T.E.; Haberland, M.; Reddy, T.; Cournapeau, D.; Burovski, E.; Peterson, P.; Weckesser, W.; Bright, J.; et al. SciPy 1.0 Contributors, SciPy 1.0: Fundamental algorithms for scientific computing in Python. Nat. Methods 2020, 17, 261–272. [Google Scholar] [CrossRef]
  84. Wick, R. Badread: Simulation of error-prone long reads. J. Open Source Softw. 2019, 4, 1316. [Google Scholar] [CrossRef]
  85. Gourlé, H.; Karlsson-Lindsjö, O.; Hayer, J.; Bongcam-Rudloff, E. Simulating Illumina metagenomic data with InSilicoSeq. Bioinformatics 2018, 35, 521–522. [Google Scholar] [CrossRef]
  86. De Coster, W.; D’Hert, S.; Schultz, D.T.; Cruts, M.; Van Broeckhoven, C. NanoPack: Visualizing and processing long-read sequencing data. Bioinformatics 2018, 34, 2666–2669. [Google Scholar] [CrossRef]
  87. Wick, R.R.; Judd, L.M.; Gorrie, C.L.; Holt, K.E. Completing bacterial genome assemblies with multiplex MinION sequencing. Microb. Genom. 2017, 3, e000132. [Google Scholar] [CrossRef] [PubMed]
  88. Hyatt, D.; Chen, G.-L.; Locascio, P.F.; Land, M.L.; Larimer, F.W.; Hauser, L.J. Prodigal: Prokaryotic gene recognition and translation initiation site identification. BMC Bioinform. 2010, 11, 119. [Google Scholar] [CrossRef] [PubMed]
  89. El-Gebali, S.; Mistry, J.; Bateman, A.; Eddy, S.R.; Luciani, A.; Potter, S.C.; Qureshi, M.; Richardson, L.J.; Salazar, G.A.; Smart, A.; et al. The Pfam protein families database in 2019. Nucleic Acids Res. 2019, 47, D427–D432. [Google Scholar] [CrossRef] [PubMed]
  90. The UniProt Consortium. UniProt: The universal protein knowledgebase in 2021. Nucleic Acids Res. 2021, 49, D480–D489. [Google Scholar] [CrossRef]
  91. Chen, S.; Zhou, Y.; Chen, Y.; Gu, J. fastp: An ultra-fast all-in-one FASTQ preprocessor. Bioinformatics 2018, 34, i884–i890. [Google Scholar] [CrossRef]
  92. Langmead, B.; Salzberg, S.L. Fast gapped-read alignment with Bowtie 2. Nat. Methods 2012, 9, 357–359. [Google Scholar] [CrossRef]
  93. Kim, D.; Paggi, J.M.; Park, C.; Bennett, C.; Salzberg, S.L. Graph-based genome alignment and genotyping with HISAT2 and HISAT-genotype. Nat. Biotechnol. 2019, 37, 907–915. [Google Scholar] [CrossRef]
  94. Vollmers, J.; Wiegand, S.; Kaster, A.-K. Comparing and Evaluating Metagenome Assembly Tools from a Microbiologist’s Perspective—Not Only Size Matters! PLoS ONE 2017, 12, e0169662. [Google Scholar] [CrossRef]
  95. Li, D.; Liu, C.-M.; Luo, R.; Sadakane, K.; Lam, T.-W. MEGAHIT: An ultra-fast single-node solution for large and complex metagenomics assembly via succinct de Bruijn graph. Bioinformatics 2015, 31, 1674–1676. [Google Scholar] [CrossRef]
  96. Bankevich, A.; Nurk, S.; Antipov, D.; Gurevich, A.A.; Dvorkin, M.; Kulikov, A.S.; Lesin, V.M.; Nikolenko, S.I.; Pham, S.; Prjibelski, A.D.; et al. SPAdes: A new ge-nome assembly algorithm and its applications to single-cell sequencing. J. Comput. Biol. 2012, 19, 455–477. [Google Scholar] [CrossRef] [Green Version]
  97. Nurk, S.; Meleshko, D.; Korobeynikov, A.; Pevzner, P.A. metaSPAdes: A new versatile metagenomic assembler. Genome Res. 2017, 27, 824–834. [Google Scholar] [CrossRef] [PubMed]
  98. Antipov, D.; Raiko, M.; Lapidus, A.; Pevzner, P.A. Metaviral SPAdes: Assembly of viruses from metagenomic data. Bioinformatics 2020, 36, 4126–4129. [Google Scholar] [CrossRef] [PubMed]
  99. Bushmanova, E.; Antipov, D.; Lapidus, A.; Prjibelski, A.D. rnaSPAdes: A de novo transcriptome assembler and its application to RNA-Seq data. Gigascience 2019, 8, 9. [Google Scholar] [CrossRef] [PubMed]
  100. Prjibelski, A.D.; Antipov, D.; Meleshko, D.; Lapidus, A.; Korobeynikov, A. Using SPAdes De Novo Assembler. Curr. Pro-Tocols Bioinform. 2020, 70, e102. [Google Scholar] [CrossRef]
  101. Collins, P.L.; Olmsted, R.A.; Spriggs, M.K.; Johnson, P.R.; Buckler-White, A.J. Gene overlap and site-specific attenuation of transcription of the viral polymerase L gene of human respiratory syncytial virus. Proc. Natl. Acad. Sci. USA 1987, 84, 5134–5138. [Google Scholar] [CrossRef]
  102. Bertram, G.; Innes, S.; Minella, O.; Richardson, J.P.; Stansfield, I. Endless possibilities: Translation termination and stop co-don recognition. Microbiology 2001, 147, 255–269. [Google Scholar] [CrossRef]
  103. Kozak, M. Pushing the limits of the scanning mechanism for initiation of translation. Gene 2002, 299, 1–34. [Google Scholar] [CrossRef]
  104. Dreher, T.W.; Miller, W.A. Translational control in positive strand RNA plant viruses. Virology 2005, 344, 185–197. [Google Scholar] [CrossRef]
  105. Firth, A.E.; Wills, N.M.; Gesteland, R.F.; Atkins, J.F. Stimulation of stop codon readthrough: Frequent presence of an ex-tended 3’ RNA structural element. Nucleic Acids Res. 2011, 39, 6679–6691. [Google Scholar] [CrossRef]
  106. Dubois, J.; Terrier, O.; Rosa-Calatrava, M. Influenza Viruses and mRNA Splicing: Doing More with Less. mBio 2014, 5, e00070-14. [Google Scholar] [CrossRef] [Green Version]
  107. Jaafar, Z.A.; Kieft, J.S. Viral RNA structure-based strategies to manipulate translation. Nat. Rev. Genet. 2018, 17, 110–123. [Google Scholar] [CrossRef] [PubMed]
  108. Haas, B.J.; Papanicolaou, A.; Yassour, M.; Grabherr, M.; Blood, P.D.; Bowden, J.; Couger, M.B.; Eccles, D.; Li, B.; Lieber, M.; et al. De novo transcript sequence reconstruction from RNA-seq using the Trinity platform for reference generation and analysis. Nat. Protoc. 2013, 8, 1494–1512. [Google Scholar] [CrossRef] [PubMed]
  109. EFSA Panel on Plant Health; Bragard, C.; Dehnen-Schmutz, K.; Gonthier, P.; Jacques, M.A.; Jaques Miret, J.A.; Justesen, A.F.; MacLeod, A.; Magnusson, C.S.; Milonas, P.; et al. Pest categorisation of non-EU viruses and viroids of Cydonia Mill., Malus Mill. and Pyrus L. EFSA J. 2019, 17, e05590. [Google Scholar] [CrossRef] [PubMed]
  110. Canales, C.; Moran, F.; Olmos, A.; Ruiz-Garcia, A.B. First Detection and Molecular Characterization of Apple Stem Grooving Virus, Apple Chlorotic Leaf Spot Virus, and Apple Hammerhead Viroid in Loquat in Spain. Plants 2021, 10, 2293. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The flow of analysis stages for each simulated library, depending on whether it was produced as an ONT or Illumina dataset. Each set of nucleotide and protein sequences were classified and screened for the virus of interest. Each traversal from raw sequences to a set of nucleotide and protein sequences is referred to as an analysis iteration.
Figure 1. The flow of analysis stages for each simulated library, depending on whether it was produced as an ONT or Illumina dataset. Each set of nucleotide and protein sequences were classified and screened for the virus of interest. Each traversal from raw sequences to a set of nucleotide and protein sequences is referred to as an analysis iteration.
Viruses 14 02163 g001
Figure 2. Sensitivity for detection of plant- and vertebrate-infecting viruses, and specificity against plant and vertebrate sequences, using nucleotide classification methods. (Top) BLASTn and MMSeqs2 classifications against each of the in-house nucleotide databases, after filtering for results with at least 70% identity over 50% of the query length and an e-value of less than 0.001. (Bottom) Classification using k-mer profiling tools after result filtering with the appropriate classification threshold for each tool. Classification was performed against a single database, except for Kraken2, in which case both the standard and viral databases were tested.
Figure 2. Sensitivity for detection of plant- and vertebrate-infecting viruses, and specificity against plant and vertebrate sequences, using nucleotide classification methods. (Top) BLASTn and MMSeqs2 classifications against each of the in-house nucleotide databases, after filtering for results with at least 70% identity over 50% of the query length and an e-value of less than 0.001. (Bottom) Classification using k-mer profiling tools after result filtering with the appropriate classification threshold for each tool. Classification was performed against a single database, except for Kraken2, in which case both the standard and viral databases were tested.
Viruses 14 02163 g002
Figure 3. Sensitivity for detection of plant- and vertebrate-infecting viruses, and specificity against the plant and vertebrate sequences, using protein annotation methods. Databases used for classification are provided under each tool name. (Top) BLASTp, DIAMOND BLASTp, MMSeqs2, and usearch (ublast and usearch_global methods), after filtering for results with at least 70% identity over 50% of the query length and an e-value of less than 0.001. (Bottom, left) Nucleotide/protein translated searches using DIAMOND BLASTx and MMSeqs2. (Bottom, right) Classification according to HMM searches against candidate profile databases. For translated searches and HMM profile searches filtering was performed specifying an e-value of less than 0.001. Arrows denote instances where classification was successful, but at a rate too low to be visible in the y-axis. Due to impractical run time requirements, BLASTp and translational MMSeqs2 approaches could not be assessed for specificity.
Figure 3. Sensitivity for detection of plant- and vertebrate-infecting viruses, and specificity against the plant and vertebrate sequences, using protein annotation methods. Databases used for classification are provided under each tool name. (Top) BLASTp, DIAMOND BLASTp, MMSeqs2, and usearch (ublast and usearch_global methods), after filtering for results with at least 70% identity over 50% of the query length and an e-value of less than 0.001. (Bottom, left) Nucleotide/protein translated searches using DIAMOND BLASTx and MMSeqs2. (Bottom, right) Classification according to HMM searches against candidate profile databases. For translated searches and HMM profile searches filtering was performed specifying an e-value of less than 0.001. Arrows denote instances where classification was successful, but at a rate too low to be visible in the y-axis. Due to impractical run time requirements, BLASTp and translational MMSeqs2 approaches could not be assessed for specificity.
Viruses 14 02163 g003
Figure 4. The rate of successful detections at either the strain or species level for each mock library analysis options for nucleic acid classification tool. Sequence sensitivity for each tool and database combination is reported as the proportion of successful identification events across all samples for the given sequencing platform, as assessed at strain- or species-level resolution.
Figure 4. The rate of successful detections at either the strain or species level for each mock library analysis options for nucleic acid classification tool. Sequence sensitivity for each tool and database combination is reported as the proportion of successful identification events across all samples for the given sequencing platform, as assessed at strain- or species-level resolution.
Viruses 14 02163 g004
Figure 5. The rate of successful detections at either the strain or species level for each mock library analysis from coding region predictions made using prodigal. Sequence sensitivity for each tool and database combination is reported as the proportion of successful identification events across all samples for the given sequencing platform, as assessed at strain- or species-level resolution.
Figure 5. The rate of successful detections at either the strain or species level for each mock library analysis from coding region predictions made using prodigal. Sequence sensitivity for each tool and database combination is reported as the proportion of successful identification events across all samples for the given sequencing platform, as assessed at strain- or species-level resolution.
Viruses 14 02163 g005
Table 1. A summary of the number of genomes, unique nucleotide contigs, and representative protein coding sequences present in each of the non-target lineages.
Table 1. A summary of the number of genomes, unique nucleotide contigs, and representative protein coding sequences present in each of the non-target lineages.
LineageGenomesUnique ContigsRepresentative Proteins
Fungus252314,840,5068,485,622
Invertebrate1137136,042,6985,593,319
Plant769100,686,5976,412,958
Protozoa4654,690,4391,701,139
Vertebrate (mammal)536198,838,3253,611,964
Vertebrate (other)79352,609,3005,459,166
Prokaryote31,9103,357,37263,089,651
Table 2. Summary statistics for the candidate nucleotide and protein sequence databases constructed for testing.
Table 2. Summary statistics for the candidate nucleotide and protein sequence databases constructed for testing.
Sequence TypeClustering MethodLineage GroupsSequencesDatabase Name
Nucleotide (genome)Unique contigs17,23127,029genome_nt
Nucleotide (nt)Species, 90%40,114181,612nt_derep1
Truncated species, 90%14,630157,271nt_derep2
Truncated species, 97%14,630559,726nt_derep3
Truncated species, 99%14,6301,076,325nt_derep4
Protein (genome)Unique sequences17,231686,559genome_prot
Protein (nr)Species, 90%38,4572,924,049nr_derep1
Truncated species, 90%14,5422,198,163nr_derep2
Table 3. Number of protein clusters yielded by each method following the removal of duplicate and polyprotein sequences. The final, filtered count of sequences was obtained by removing protein clusters whose membership did not sufficiently cover at least one viral subfamily (see methods for how coverage was assessed). Family and subfamily classifications were exported from the NCBI taxonomy record for each virus species.
Table 3. Number of protein clusters yielded by each method following the removal of duplicate and polyprotein sequences. The final, filtered count of sequences was obtained by removing protein clusters whose membership did not sufficiently cover at least one viral subfamily (see methods for how coverage was assessed). Family and subfamily classifications were exported from the NCBI taxonomy record for each virus species.
MethodDatabase NameInitial ClustersNon-Singleton
Clusters
Filtered Clusters
usearch (50%)usearch_50253,37990,1656612
usearch (60%)usearch_60283,61293,2866802
usearch (70%)usearch_70315,14094,8257310
usearch (80%)usearch_80350,57494,7516514
usearch (90%)usearch_90400,30391,9505918
mcl (I = 1.4) *mcl_1.4135,20658,272397 *
mcl (I = 4.0)mcl_4.0154,11477,0325950
Mixed (usearch 50%)mixed_50361,297110,5258569
Mixed (usearch 60%)mixed_60396,879119,2939468
Mixed (usearch 70%)mixed_70433,833126,78210,283
Mixed (usearch 80%)mixed_80475,194133,29310,519
Mixed (usearch 90%)mixed_90532,477139,7389498
Asterisk (*) denotes where clustering was filtered at the family level rather than subfamily.
Table 4. Sensitivity, specificity, and average scoring values for the four top performing candidate databases after e-value filtering (0.001) of hits.
Table 4. Sensitivity, specificity, and average scoring values for the four top performing candidate databases after e-value filtering (0.001) of hits.
DatabasePlant
Sensitivity
SpecificityF1 ScoreAnimal
Sensitivity
SpecificityF1 ScoreAverage
Sensitivity
Specificity
usearch_500.97950.88950.93230.94010.88690.91270.95980.8882
mcl_1.40.9490.92850.93860.84640.90220.87340.89770.9153
mcl_4.00.98220.86790.92150.99050.85410.91720.98640.861
mixed_900.98240.84680.90960.99030.81520.89430.98640.831
Table 5. Detection sensitivity for nucleotide sequences using virus detection pipelines. Note that the sensitivity of VirTool was assessed using only the sequence mapping portion of the workflow—the VirTool HMM database was assessed as part of the HMM testing later in this manuscript.
Table 5. Detection sensitivity for nucleotide sequences using virus detection pipelines. Note that the sensitivity of VirTool was assessed using only the sequence mapping portion of the workflow—the VirTool HMM database was assessed as part of the HMM testing later in this manuscript.
ToolDatabasePlantVertebrateInvertebrateOther
DeepVirFinder -0.66540.60390.72230.9104
FastViromeExplorereukvir0.02790.01020.00640.0030
gov0.00000.00000.00110.0002
imgvr0.00000.00000.00000.0000
ncbi0.04010.01380.00700.0115
MiCOP-0.88750.74580.89470.8073
VirFinderdefault0.44420.36540.46290.8437
modEPV0.91490.80830.86680.8539
ViromeScan-0.91490.86990.59940.7890
VirSorterdb10.00980.09780.01990.7397
db20.00980.09430.01990.7316
VirSorter2-0.86640.94110.96350.9843
VirTool-0.88550.00630.00860.0093
Table 6. The number of successfully completed assemblies for each assembler and subsequent coding region predictions with each approach. The number of samples processed were 24 for DNA and 141 for RNA libraries.
Table 6. The number of successfully completed assemblies for each assembler and subsequent coding region predictions with each approach. The number of samples processed were 24 for DNA and 141 for RNA libraries.
HostLibrary TypePlatformAssemblerProdigalTransDecoderTransDecoder (Guided)
RetainedDNAONT-241424
HiSeqMegaHIT242424
SPAdes (meta)212117
SPAdes (metaviral)242424
RNAONT-141141140
HiSeqMegaHIT141141141
SPAdes (rna)141141141
SPAdes (rnaviral)141141141
RemovedDNAONT-242424
HiSeqMegaHIT241414
SPAdes (meta)241717
SPAdes (metaviral)0--
RNAONT-141113113
HiSeqMegaHIT141130130
SPAdes (rna)141140140
SPAdes (rnaviral)141133133
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Waite, D.W.; Liefting, L.; Delmiglio, C.; Chernyavtseva, A.; Ha, H.J.; Thompson, J.R. Development and Validation of a Bioinformatic Workflow for the Rapid Detection of Viruses in Biosecurity. Viruses 2022, 14, 2163. https://doi.org/10.3390/v14102163

AMA Style

Waite DW, Liefting L, Delmiglio C, Chernyavtseva A, Ha HJ, Thompson JR. Development and Validation of a Bioinformatic Workflow for the Rapid Detection of Viruses in Biosecurity. Viruses. 2022; 14(10):2163. https://doi.org/10.3390/v14102163

Chicago/Turabian Style

Waite, David W., Lia Liefting, Catia Delmiglio, Anastasia Chernyavtseva, Hye Jeong Ha, and Jeremy R. Thompson. 2022. "Development and Validation of a Bioinformatic Workflow for the Rapid Detection of Viruses in Biosecurity" Viruses 14, no. 10: 2163. https://doi.org/10.3390/v14102163

APA Style

Waite, D. W., Liefting, L., Delmiglio, C., Chernyavtseva, A., Ha, H. J., & Thompson, J. R. (2022). Development and Validation of a Bioinformatic Workflow for the Rapid Detection of Viruses in Biosecurity. Viruses, 14(10), 2163. https://doi.org/10.3390/v14102163

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop