-
PDF
- Split View
-
Views
-
Cite
Cite
Andrea H L Bruning, Mariska M G Leeflang, Johanna M B W Vos, Rene Spijker, Menno D de Jong, Katja C Wolthers, Dasja Pajkrt, Rapid Tests for Influenza, Respiratory Syncytial Virus, and Other Respiratory Viruses: A Systematic Review and Meta-analysis, Clinical Infectious Diseases, Volume 65, Issue 6, 15 September 2017, Pages 1026–1032, https://doi.org/10.1093/cid/cix461
- Share Icon Share
Abstract
Rapid diagnosis of respiratory virus infections contributes to patient care. This systematic review evaluates the diagnostic accuracy of rapid tests for the detection of respiratory viruses. We searched Medline and EMBASE for studies evaluating these tests against polymerase chain reaction as the reference standard. Of 179 studies included, 134 evaluated rapid tests for influenza viruses, 32 for respiratory syncytial virus (RSV), and 13 for other respiratory viruses. We used the bivariate random effects model for quantitative meta-analysis of the results. Most tests detected only influenza viruses or RSV. Summary sensitivity and specificity estimates of tests for influenza were 61.1% and 98.9%. For RSV, summary sensitivity was 75.3%, and specificity, 98.7%. We assessed the quality of studies using the Quality Assessment of Diagnostic Accuracy Studies (QUADAS-2) checklist. Because of incomplete reporting, the risk of bias was often unclear. Despite their intended use at the point of care, 26.3% of tests were evaluated in a laboratory setting. Although newly developed tests seem more sensitive, high-quality evaluations of these tests are lacking.
Acute respiratory tract infections (RTIs) are a leading cause of disease and death worldwide [1]. Recent studies suggest more than half of RTIs are caused by viruses, even in severely ill patients [2–4]. Commonly detected viruses include influenza viruses, respiratory syncytial virus (RSV), adenovirus, human metapneumovirus (hMPV), human parainfluenza viruses, coronaviruses, and rhinoviruses [5, 6].
Clinical signs and symptoms of viral RTIs overlap with those of bacterial infections. It is challenging to clinically distinguish bacterial from viral infections and different viral pathogens [7]. This diagnostic uncertainty leads to overprescription of antibiotics and extra diagnostic testing (along with costs) to rule out bacterial infections [8, 9]. Rapid detection of viral pathogens could overcome these disadvantages. Besides, prompt viral diagnosis may lead to rapid implementation of infection control measures, early administration of antiviral medication, if available, and shorter hospital stays, resulting in reduced healthcare costs [10–12].
For this reason, rapid diagnostic or point-of-care tests have been developed. Compared with other diagnostic modalities—culture, polymerase chain reaction (PCR), or immunofluorescence testing—point-of-care tests are often faster, less expensive, easier to use and accessible to staff without laboratory training. They have the potential to be carried out at or near the point of care.
There is a clear trend toward point-of-care testing, and the number and quality of rapid diagnostic tests for respiratory viruses has rapidly increased [13]. For clinicians, it is important to be aware of the diagnostic accuracy of the different rapid viral tests, the factors that affect their accuracies, and test performances in daily practice. Previous systematic reviews addressing the diagnostic accuracy of respiratory tests either evaluated only a single respiratory virus (influenza or RSV) or were conducted in specific populations [14–16].
The aim of our review was to provide a state-of-the-art overview of all available rapid tests for the detection of respiratory viruses in patients of all ages with RTIs. We systematically summarized the available evidence on their diagnostic accuracy for virus detection compared with PCR testing. We assessed the quality of included studies and quantitatively summarized the results in a meta-analysis.
MATERIALS AND METHODS
Protocol
This systematic review was built on a protocol based on the Preferred Reporting Items for Systematic Review and Meta-Analyses for Protocols 2015 (PRISMA-P 2015), which is registered in the Prospero database (registration No. CRD42015024581).
Search Strategy
A systematic literature review was conducted by searching Medline and EMBASE electronic databases, through the Ovid interface, from inception to 18 January 2016. The reference lists of all articles were hand searched for additional studies.
The search strategy was developed in collaboration with a medical information specialist (R. S.) and contained search terms for the most common respiratory viruses (influenza or respiratory syncytial virus or metapneumovirus or parainfluenza virus or human adenovirus or human rhinovirus or human bocavirus or human coronavirus) combined with search terms for rapid diagnostic tests (diagnostic kit or antigen [test or detection] or reagent or immunotest or point-of-care systems or rapid, simple, easy, or quick test), including brand names for the most common commercial rapid tests. The complete search strategy is shown in Appendix A. After removal of duplicates, the search results were imported into Covidence, a Web-based software platform that streamlines the production of systematic reviews (https://www.covidence.org/).
Selection Criteria
Studies were considered for inclusion if they were written in English or Dutch and reported original data regarding the accuracy of a rapid test for ≥1 respiratory virus compared with PCR. Although standardization across PCR methods is currently lacking, we selected PCR as the appropriate reference standard because of its current status as the reference standard method for detecting respiratory viruses [17]. Rapid tests were defined (adapted from the World Health Organization simple/rapid tests definition [18]) as any commercially available quick (up to 2 hours) and easy-to-use test requiring little or no additional equipment or technological skills.
Only original studies were included. Studies evaluating in-house tests and precommercial versions of rapid tests were excluded, as well as studies using the rapid test result as part of a composite reference standard (to avoid incorporation bias) or performing PCR solely on samples in which rapid test results were negative (to avoid partial verification bias). Case-control studies (use of the rapid test on previously tested known positive or negative samples), case reports, conference abstracts, reviews, and veterinary studies were excluded.
Two reviewers (A. H. L. B. and D. P.) independently assessed inclusion eligibility. Initial selection was based on screening of title and abstract. Full-text versions were further assessed for eligibility, and reasons for exclusion were documented.
Data Extraction
A data-extraction-form was developed with a clinical epidemiologist (M. M. G. L.). Two authors (A. H. L. B. and J. M. B. W. V.) independently extracted data. Information collected included author, year of study and publication, country, study population, specimen type, setting, performance of the test in a laboratory or at the point of care, brand name, and data required to calculate sensitivity and specificity. With these data we constructed 2×2 contingency tables. If 2×2 tables could not be extracted or calculated from the published data, we contacted authors for additional information. In some articles, the same sample sets were tested with different rapid tests. Each test comparison was considered a separate analysis, resulting in a total number of included studies higher than the actual number of included articles.
Quality Assessment
The methodological quality of studies was independently reviewed by 2 authors (A. H. L. B. and J. M. B. W. V) using Quality Assessment of Diagnostic Accuracy Studies (QUADAS-2) criteria, as recommended by the Cochrane Collaboration [19]. Disagreement about inclusion, data extraction, and quality assessment was solved by consensus.
Statistical Analysis
With the 2 × 2 tables, we estimated the sensitivity and specificity of the rapid tests. Each respiratory virus can be regarded as a separate target condition. Therefore, we analyzed the accuracy of the tests for each virus separately. Studies evaluating rapid tests for influenza viruses sometimes reported multiple 2×2 tables for influenza A and B separately. However, most of the included tests detected both influenza subtypes and made no distinction between subtypes. We therefore used “any influenza” as our main analyses and evaluated test accuracy in subtypes in subgroup analyses.
Sensitivity and specificity estimates were pooled into receiver operating characteristic plots. Because sensitivity and specificity were negatively correlated, the meta-analysis should be performed on both outcome measures simultaneously. For this, we used a bivariate random effects meta-regression model [20]. To investigate the effect of potential sources of heterogeneity, we added the following covariates to the model: study population (children vs adults), commercial brand of the rapid test, and whether or not the test was performed at the point of care. We calculated summary sensitivity and specificity estimates for each covariate. To assess the effect of study quality, we performed sensitivity analyses. All analyses were performed using SAS software, version 9.4.
RESULTS
Search Results
Our literature searches identified 3527 articles. After screening, 383 articles were eligible for full-text review, of which 258 were excluded (Figure 1). The main reasons for exclusion were that the rapid test was evaluated in a case-control study, the test could not be performed rapid, the study did not include original data, a reference test other than PCR was used, or the study was not reported in English or Dutch. Because several reports evaluated >1 test, 179 separate studies were included in the systematic review and meta-analysis. Of these, 134 studies (74.9%) evaluated rapid tests for influenza viruses, 32 (17.9%) for RSV, and the remaining 13 (7.3%) for other respiratory viruses (hMPV, adenovirus, and parainfluenza virus; Table 1).

Preferred Reporting Items for Systematic Review and Meta-Analyses (PRISMA) flow diagram.
Characteristic . | Studies, No. (%) . |
---|---|
Virus evaluated | |
Influenza virus (A, B, H1N1, A, and B) | 134 (74.9) |
RSV | 32 (17.9) |
Adenovirus | 6 (3.4) |
Human metapneumovirus | 5 (2.8) |
Parainfluenza virus type 1–3 | 2 (1.1) |
Population | |
Children | 74 (41.3) |
Adults | 14 (7.8) |
Mixed/not reported | 91 (50.8) |
Most frequently studied tests (company)a | |
Influenza virus | |
QuickVue Influenza A+B (Quidel) | 26 (14.5) |
BinaxNOW Influenza A&B (Alere Scarborough) | 19 (10.6) |
Directigen EZ FluA+B (Directigen-EZ, Becton Dickinson) | 16 (8.9) |
Sofia Influenza A+B (Quidel) | 9 (5.0) |
RSV | |
BinaxNOW RSV (Alere Scarborough) | 6 (3.3) |
BD Veritor RSV (BD Diagnostics) | 5 (2.8) |
Sofia RSV (Quidel) | 4 (2.2) |
Directigen EZ RSV (Becton Dickinson) | 3 (1.7) |
Multiplex test | |
mariPOC (ArcDia International) | 15 (8.4) |
Other rapid tests | 77 (43.0) |
Point-of-care testing | 47 (26.3) |
Characteristic . | Studies, No. (%) . |
---|---|
Virus evaluated | |
Influenza virus (A, B, H1N1, A, and B) | 134 (74.9) |
RSV | 32 (17.9) |
Adenovirus | 6 (3.4) |
Human metapneumovirus | 5 (2.8) |
Parainfluenza virus type 1–3 | 2 (1.1) |
Population | |
Children | 74 (41.3) |
Adults | 14 (7.8) |
Mixed/not reported | 91 (50.8) |
Most frequently studied tests (company)a | |
Influenza virus | |
QuickVue Influenza A+B (Quidel) | 26 (14.5) |
BinaxNOW Influenza A&B (Alere Scarborough) | 19 (10.6) |
Directigen EZ FluA+B (Directigen-EZ, Becton Dickinson) | 16 (8.9) |
Sofia Influenza A+B (Quidel) | 9 (5.0) |
RSV | |
BinaxNOW RSV (Alere Scarborough) | 6 (3.3) |
BD Veritor RSV (BD Diagnostics) | 5 (2.8) |
Sofia RSV (Quidel) | 4 (2.2) |
Directigen EZ RSV (Becton Dickinson) | 3 (1.7) |
Multiplex test | |
mariPOC (ArcDia International) | 15 (8.4) |
Other rapid tests | 77 (43.0) |
Point-of-care testing | 47 (26.3) |
Abbreviation: RSV, respiratory syncytial virus.
Characteristic . | Studies, No. (%) . |
---|---|
Virus evaluated | |
Influenza virus (A, B, H1N1, A, and B) | 134 (74.9) |
RSV | 32 (17.9) |
Adenovirus | 6 (3.4) |
Human metapneumovirus | 5 (2.8) |
Parainfluenza virus type 1–3 | 2 (1.1) |
Population | |
Children | 74 (41.3) |
Adults | 14 (7.8) |
Mixed/not reported | 91 (50.8) |
Most frequently studied tests (company)a | |
Influenza virus | |
QuickVue Influenza A+B (Quidel) | 26 (14.5) |
BinaxNOW Influenza A&B (Alere Scarborough) | 19 (10.6) |
Directigen EZ FluA+B (Directigen-EZ, Becton Dickinson) | 16 (8.9) |
Sofia Influenza A+B (Quidel) | 9 (5.0) |
RSV | |
BinaxNOW RSV (Alere Scarborough) | 6 (3.3) |
BD Veritor RSV (BD Diagnostics) | 5 (2.8) |
Sofia RSV (Quidel) | 4 (2.2) |
Directigen EZ RSV (Becton Dickinson) | 3 (1.7) |
Multiplex test | |
mariPOC (ArcDia International) | 15 (8.4) |
Other rapid tests | 77 (43.0) |
Point-of-care testing | 47 (26.3) |
Characteristic . | Studies, No. (%) . |
---|---|
Virus evaluated | |
Influenza virus (A, B, H1N1, A, and B) | 134 (74.9) |
RSV | 32 (17.9) |
Adenovirus | 6 (3.4) |
Human metapneumovirus | 5 (2.8) |
Parainfluenza virus type 1–3 | 2 (1.1) |
Population | |
Children | 74 (41.3) |
Adults | 14 (7.8) |
Mixed/not reported | 91 (50.8) |
Most frequently studied tests (company)a | |
Influenza virus | |
QuickVue Influenza A+B (Quidel) | 26 (14.5) |
BinaxNOW Influenza A&B (Alere Scarborough) | 19 (10.6) |
Directigen EZ FluA+B (Directigen-EZ, Becton Dickinson) | 16 (8.9) |
Sofia Influenza A+B (Quidel) | 9 (5.0) |
RSV | |
BinaxNOW RSV (Alere Scarborough) | 6 (3.3) |
BD Veritor RSV (BD Diagnostics) | 5 (2.8) |
Sofia RSV (Quidel) | 4 (2.2) |
Directigen EZ RSV (Becton Dickinson) | 3 (1.7) |
Multiplex test | |
mariPOC (ArcDia International) | 15 (8.4) |
Other rapid tests | 77 (43.0) |
Point-of-care testing | 47 (26.3) |
Abbreviation: RSV, respiratory syncytial virus.
Characteristics of Included Studies
Supplementary Table 1 provides the main characteristics of included studies. Supplementary Figure 1 displays the accompanying forest plots of the estimated sensitivities and specificities. Table 1 shows that a total of 50 different rapid tests were evaluated. The most frequently studied tests for influenza were Alere BinaxNOW Influenza A&B Test, Quidel QuickVue Influenza A+B, BD Directigen EZ Flu A+B, and Quidel Sofia Influenza A+B FIA. For RSV, BD Veritor System, Alere BinaxNOW RSV Card, and Quidel Sofia RSV FIA were evaluated most frequently. Most studies were conducted in either children or populations including both adults and children. Of the studied rapid tests, 26.3% were evaluated at the point of care.
Quality of Included Studies
An overview of the risk of bias and concerns regarding applicability of included articles is presented in Figure 2. Because of our inclusion and exclusion criteria, studies were free of spectrum, incorporation, and partial verification bias, and all included studies used an appropriate reference standard. Nevertheless, several study characteristics might incorporate a risk of bias. Because inclusion and exclusion criteria were often defined unclearly or not at all, the risk of bias regarding patient selection was difficult to assess. In at least 37 articles (29.6%), the index test was performed at the point of care, suggesting that index test results were interpreted without knowledge of reference standard results. In 8 of these articles, additional information was available on the clinical feasibility or logistical organization of the test. Owing to frequent incomplete description of study characteristics that might result in a risk of bias, these items were difficult to assess, resulting in an unclear risk. The concerns regarding applicability were generally low.

Quality assessment of included studies using Quality Assessment of Diagnostic Accuracy Studies (QUADAS-2) criteria.
Overall Accuracy of Rapid Tests for Influenza and Respiratory Syncytial Virus
For influenza, sensitivity ranged from 4.4% to 100.0%. For all rapid tests that can detect influenza, that summary estimate was 61.1% for sensitivity (95% confidence interval [CI], 53.3%–68.3%) and 98.9% for specificity (98.4%–99.3%). Operating test characteristics of the studies regarding influenza are shown in Figure 3A.

Hierarchical summary receiver operating characteristic curve plots of studies evaluating rapid tests for influenza (A) and respiratory syncytial virus (RSV) (B). Each circle indicates a single included study, with the size of the circle proportionate to the size of the study.
For RSV, sensitivity had more variation than specificity, but sensitivity estimates were higher than for influenza, ranging from 41.2% to 88.6% (Figure 3B). The summary sensitivity and specificity for all studies evaluating RSV were 75.3% (95% CI, 72.6%–77.8%) and 98.7% (97.3%–99.4%), respectively. Owing to the small number of included studies for adenovirus, hMPV, and parainfluenza virus types 1–3, meta-analysis was not performed.
Investigation of Heterogeneity
The sensitivities of rapid tests for influenza A (68.1%; 95% CI, 58.9%–76.0%) and influenza B (71.0%; 56.8%–82.1%) were comparable. Several studies (n=58) only evaluated the diagnostic accuracy of rapid tests for detecting H1N1. Most of these rapid tests were not developed to detect H1N1 specifically. The sensitivity for H1N1 was generally lower (54.0%, 95% CI 47.6–60.3) than for the other virus subtypes (Table 2).
. | Pooled Sensitivity (95% CI) . | Pooled Specificity (95% CI) . |
---|---|---|
Influenza | ||
Virus type | ||
Influenza A | 68.1 (58.9–76.0) | 99.2 (98.5–99.6) |
H1N1 | 54.0 (47.6–60.3) | 99.1 (98.5–99.5) |
Influenza B | 71.0 (56.8–82.1) | 99.6 (99.2–99.8) |
Influenza A+B | 61.1 (53.3–68.3) | 98.9 (98.4–99.3) |
Population | ||
Children | 66.1 (52.9–79.3) | 98.3 (97.2–99.5) |
Adults | 34.1 (14.0–54.1) | 99.2 (98.2–100.0) |
Point-of-care testing | 62.1 (47.6–74.7) | 98.4 (96.7–99.2) |
Rapid test | ||
QuickVue Influenza A+B | 44.6 (29.1–60.0) | 99.3 (98.8–99.9) |
Sofia Influenza A+B | 75.3 (59.2–91.5) | 95.3 (91.5–99.2) |
BinaxNow Influenza A&B | 44.1 (23.3–64.9) | 99.4 (98.6–100.0) |
Directigen Flu A+B | 35.8 (11.8–59.7) | 99.2 (98.0–99.4) |
mariPOC | 76.1 (53.5–98.7) | 99.4 (98.3–100.0) |
RSV | ||
Population | ||
Children | 75.9 (73.1–78.5) | 98.5 (968–99.4) |
Mixed | 70.9 (63.0–77.8) | 99.1 (95.9–99.8) |
Point-of-care testing | 76.0 (69.8–81.2) | 99.1 (95.5–99.8) |
Rapid test | ||
BD Veritor RSV | 76.9 (71.0–82.8) | 98.9 (97.1–100.0) |
BinaxNOW RSV | 72.2 (65.2–79.1) | 98.6 (96.5–100.0) |
Sofia RSV | 80.0 (73.0–86.9) | 97.8 (93.8–100.0) |
. | Pooled Sensitivity (95% CI) . | Pooled Specificity (95% CI) . |
---|---|---|
Influenza | ||
Virus type | ||
Influenza A | 68.1 (58.9–76.0) | 99.2 (98.5–99.6) |
H1N1 | 54.0 (47.6–60.3) | 99.1 (98.5–99.5) |
Influenza B | 71.0 (56.8–82.1) | 99.6 (99.2–99.8) |
Influenza A+B | 61.1 (53.3–68.3) | 98.9 (98.4–99.3) |
Population | ||
Children | 66.1 (52.9–79.3) | 98.3 (97.2–99.5) |
Adults | 34.1 (14.0–54.1) | 99.2 (98.2–100.0) |
Point-of-care testing | 62.1 (47.6–74.7) | 98.4 (96.7–99.2) |
Rapid test | ||
QuickVue Influenza A+B | 44.6 (29.1–60.0) | 99.3 (98.8–99.9) |
Sofia Influenza A+B | 75.3 (59.2–91.5) | 95.3 (91.5–99.2) |
BinaxNow Influenza A&B | 44.1 (23.3–64.9) | 99.4 (98.6–100.0) |
Directigen Flu A+B | 35.8 (11.8–59.7) | 99.2 (98.0–99.4) |
mariPOC | 76.1 (53.5–98.7) | 99.4 (98.3–100.0) |
RSV | ||
Population | ||
Children | 75.9 (73.1–78.5) | 98.5 (968–99.4) |
Mixed | 70.9 (63.0–77.8) | 99.1 (95.9–99.8) |
Point-of-care testing | 76.0 (69.8–81.2) | 99.1 (95.5–99.8) |
Rapid test | ||
BD Veritor RSV | 76.9 (71.0–82.8) | 98.9 (97.1–100.0) |
BinaxNOW RSV | 72.2 (65.2–79.1) | 98.6 (96.5–100.0) |
Sofia RSV | 80.0 (73.0–86.9) | 97.8 (93.8–100.0) |
Abbreviations: CI, confidence interval; RSV, respiratory syncytial virus.
. | Pooled Sensitivity (95% CI) . | Pooled Specificity (95% CI) . |
---|---|---|
Influenza | ||
Virus type | ||
Influenza A | 68.1 (58.9–76.0) | 99.2 (98.5–99.6) |
H1N1 | 54.0 (47.6–60.3) | 99.1 (98.5–99.5) |
Influenza B | 71.0 (56.8–82.1) | 99.6 (99.2–99.8) |
Influenza A+B | 61.1 (53.3–68.3) | 98.9 (98.4–99.3) |
Population | ||
Children | 66.1 (52.9–79.3) | 98.3 (97.2–99.5) |
Adults | 34.1 (14.0–54.1) | 99.2 (98.2–100.0) |
Point-of-care testing | 62.1 (47.6–74.7) | 98.4 (96.7–99.2) |
Rapid test | ||
QuickVue Influenza A+B | 44.6 (29.1–60.0) | 99.3 (98.8–99.9) |
Sofia Influenza A+B | 75.3 (59.2–91.5) | 95.3 (91.5–99.2) |
BinaxNow Influenza A&B | 44.1 (23.3–64.9) | 99.4 (98.6–100.0) |
Directigen Flu A+B | 35.8 (11.8–59.7) | 99.2 (98.0–99.4) |
mariPOC | 76.1 (53.5–98.7) | 99.4 (98.3–100.0) |
RSV | ||
Population | ||
Children | 75.9 (73.1–78.5) | 98.5 (968–99.4) |
Mixed | 70.9 (63.0–77.8) | 99.1 (95.9–99.8) |
Point-of-care testing | 76.0 (69.8–81.2) | 99.1 (95.5–99.8) |
Rapid test | ||
BD Veritor RSV | 76.9 (71.0–82.8) | 98.9 (97.1–100.0) |
BinaxNOW RSV | 72.2 (65.2–79.1) | 98.6 (96.5–100.0) |
Sofia RSV | 80.0 (73.0–86.9) | 97.8 (93.8–100.0) |
. | Pooled Sensitivity (95% CI) . | Pooled Specificity (95% CI) . |
---|---|---|
Influenza | ||
Virus type | ||
Influenza A | 68.1 (58.9–76.0) | 99.2 (98.5–99.6) |
H1N1 | 54.0 (47.6–60.3) | 99.1 (98.5–99.5) |
Influenza B | 71.0 (56.8–82.1) | 99.6 (99.2–99.8) |
Influenza A+B | 61.1 (53.3–68.3) | 98.9 (98.4–99.3) |
Population | ||
Children | 66.1 (52.9–79.3) | 98.3 (97.2–99.5) |
Adults | 34.1 (14.0–54.1) | 99.2 (98.2–100.0) |
Point-of-care testing | 62.1 (47.6–74.7) | 98.4 (96.7–99.2) |
Rapid test | ||
QuickVue Influenza A+B | 44.6 (29.1–60.0) | 99.3 (98.8–99.9) |
Sofia Influenza A+B | 75.3 (59.2–91.5) | 95.3 (91.5–99.2) |
BinaxNow Influenza A&B | 44.1 (23.3–64.9) | 99.4 (98.6–100.0) |
Directigen Flu A+B | 35.8 (11.8–59.7) | 99.2 (98.0–99.4) |
mariPOC | 76.1 (53.5–98.7) | 99.4 (98.3–100.0) |
RSV | ||
Population | ||
Children | 75.9 (73.1–78.5) | 98.5 (968–99.4) |
Mixed | 70.9 (63.0–77.8) | 99.1 (95.9–99.8) |
Point-of-care testing | 76.0 (69.8–81.2) | 99.1 (95.5–99.8) |
Rapid test | ||
BD Veritor RSV | 76.9 (71.0–82.8) | 98.9 (97.1–100.0) |
BinaxNOW RSV | 72.2 (65.2–79.1) | 98.6 (96.5–100.0) |
Sofia RSV | 80.0 (73.0–86.9) | 97.8 (93.8–100.0) |
Abbreviations: CI, confidence interval; RSV, respiratory syncytial virus.
The diagnostic test accuracy for influenza was significantly decreased when tests were performed in adults (sensitivity, 34.1%; 95% CI, 14.0%–54.1%), compared with performance in children or a mixed population (P < .01). For RSV, age did not significantly influence performance characteristics (P = .20).
To identify the rapid test with the best performance characteristics, we evaluated pooled summary estimates for each test separately. This was only possible for tests with ≥4 studies. For influenza, mariPOC and Sofia Influenza A+B FIA had the best overall performance, with summary sensitivities of 76.1% and 75.3%, respectively, and summary specificities of 99.4% and 95.3%, although in some studies [21] the specificity for Sofia Influenza A+B FIA was lower. For RSV, an estimated sensitivity and specificity per test could be calculated only for BD Veritor RSV, Sofia RSV FIA, and BinaxNOW RSV. Sofia RSV FIA had the best overall performance. For both influenza and RSV, diagnostic test accuracy was not influenced by setting—that is, by whether the test was performed at the point of care or in the laboratory.
The risk of bias for studies evaluating influenza was generally low. We assessed the difference between studies with low risk of bias for the patient domain and those with an unclear risk of bias in this domain. Sensitivity was generally lower in studies with a low risk of bias (53.3%; 95% CI, 42.1%–64.1%). For RSV, the investigated quality criteria did not have a statistically significant effect on diagnostic test accuracy.
DISCUSSION
Main Findings of the Study
In this systematic review and meta-analysis we provided an overview of the rapid tests that are available for the detection of respiratory viruses in patients with RTIs. The sensitivity of these tests varied considerably, but specificity was high. With the result of the rapid test it is thus possible to rule in a respiratory viral infection, but false-negative results are common. The performance of rapid tests for RSV was generally superior to the performance for influenza. A major advantage of rapid tests is their potential to be performed in a nonlaboratory setting, but only 26.3% were evaluated at the point of care.
Although diagnosing RTIs requires a syndromic approach because symptoms of respiratory viruses infection overlap, only few rapid tests currently available can simultaneously detect multiple viruses. Most rapid tests solely detect influenza viruses or RSV. This is historically understandable, because these viruses were considered the most important respiratory viruses. Furthermore, availability of antivirals for influenza renders rapid diagnostic testing of influenza of high priority. However, recent studies indicate that other respiratory viruses, such as rhinoviruses and hMPV, can also cause severe respiratory illness and are sometimes detected at higher frequencies than influenza and RSV [22, 23].
The results of our heterogeneity investigation show that rapid test performance is comparable for influenza A and influenza B, which is in line with findings of a previous meta-analysis [14]. Most of the included studies were performed in children, especially those evaluating rapid tests for RSV, which at least partly explains the lack of influence of age on the diagnostic accuracy estimates for RSV. For influenza, however, rapid test performance was significantly better in children, as reported elsewhere [24, 25]. Age is inversely associated with viral load, which may explain better test results in children.
Testing at the point of care did not influence diagnostic test accuracy, although this finding should be interpreted with caution. In many studies, the test was not evaluated at the point of care, nor was the setting or personnel described. Direct head-to-head comparisons between tests performed at the point of care and in the laboratory were limited. Besides, only 8 of the 37 articles describing tests performed at the point of care included information on the clinical feasibility of the test—that is, whether the test was easy to use, how rapid test results were communicated, or how the personnel performing the rapid tests evaluated its usefulness.
By estimating the pooled sensitivities of the rapid tests, we aimed to determine which test performed best. The conclusions should be interpreted cautiously. Because of our random-effects model, we could not evaluate all rapid tests separately, and the newly developed molecular testing devices were evaluated only sporadically in prospective cross-sectional studies [26, 27]. In addition, rapid test kits from the same company might undergo manufacturing changes. When we analyzed only high-quality studies (ie, with a low risk of bias), the overall sensitivity decreased, implying an even lower diagnostic accuracy for rapid tests.
Strengths and Limitations
Diagnostic accuracy studies are described in different ways, and there is no standard terminology available [28]. To include all available evidence, we constructed our search strategy as broadly as possible. Data were extracted, and the quality of included studies was assessed by 2 independent reviewers, decreasing the risk of subjectivity to a minimum. In the event of methodological shortcomings—for example, incorporation bias or partial verification bias—diagnostic test performance might be underestimated or overestimated, which we tried to avoid by applying strict exclusion criteria.
We deliberately chose to include only studies in which PCR was used as the reference standard, because of its status as the reference standard method for detecting respiratory viruses [17]. This important strength of our systematic review provides a more realistic accuracy estimate for rapid tests. In previous studies using viral culture or immunofluorescence as a reference, pooled sensitivities were 9%–14% higher than in studies that used PCR, and test accuracy was overestimated [15]. We did not assess publication bias because no accurate reporting methods for diagnostic test accuracy studies exist [28].
The accuracy of a systematic review depends on the quality of the studies included. As demonstrated in our quality assessment, many items regarding the risk of bias were unclear. Sensitivity and specificity calculations are extremely sensitive to the design of a study and influenced by many factors, such as virus prevalence, predominant circulating virus strain, time from illness onset to sample collection, type and quality of respiratory specimen, age, and disease severity [11, 29]. In many studies that evaluated specific rapid tests, these items were lacking or unclear. This is a major limitation in diagnostic accuracy studies in general. Comparisons between tests should be ideally performed in the same study, against the same reference standard, and in the same patients.
Future Perspectives
In the current era of emerging novel respiratory viruses, there is a growing need for rapid, sensitive, and specific identification of viral pathogens to allow effective prompt antimicrobial therapy, decrease extra diagnostic testing, and implement pathogen-specific infection control measures. Rapid tests have the potential to fulfill these needs, but one should be aware of their current limitations in diagnostic performance and range of pathogens identified.
More sensitive and specific rapid multiplex molecular assays are in development. They have the potential to rapidly and accurately identify not only respiratory viruses but also bacteria. Fully automated molecular methods are commercially available and presented as designed to be operated at the point of care [30]. Although these newer tests seem to perform comparably to laboratory-operated PCRs, comprehensive state-of-the-art clinical evaluations are lacking. Besides, in addition to high costs and low sample throughput, major drawbacks of these newer diagnostic devices are the technical complexity of the tests and their dependence on electronic devices [31, 32], limiting their opportunities for direct point-of-care use.
In the upcoming years, the nonmolecular rapid tests will still have a role in practical patient care. It is important for clinicians to be aware of their availability and performance characteristics. Our systematic review provides a helpful tool in this understanding. Although some studies have already evaluated the impact of rapid diagnostics on patient management [33–35], randomized controlled trials are needed to assess the clinical relevance of rapid tests in terms of clinically relevant outcomes, such as antibiotic use, length of hospital stay, and cost efficiency. In addition, point-of-care testing requires novel strategies for logistics and organization [36]. For successful implementation of rapid tests in the clinic, a laboratory test validation is not sufficient and a high-quality evaluation at the point of care is required.
Supplementary Data
Supplementary materials are available at Clinical Infectious Diseases online. Consisting of data provided by the authors to benefit the reader, the posted materials are not copyedited and are the sole responsibility of the authors, so questions or comments should be addressed to the corresponding author.
Notes
Financial support. This work was supported by the European Union’s Seventh Framework People Programme (Research Executive Agency grant 612308).
Disclaimer. The funding source had no role in study design, collection, analysis, or interpretation of the data or writing of the report.
Potential conflicts of interest. All authors: No reported conflicts of interest. All authors have submitted the ICMJE Form for Disclosure of Potential Conflicts of Interest. Conflicts that the editors consider relevant to the content of the manuscript have been disclosed.
References