Cargando…
Accounting for seasonal patterns in syndromic surveillance data for outbreak detection
BACKGROUND: Syndromic surveillance (SS) can potentially contribute to outbreak detection capability by providing timely, novel data sources. One SS challenge is that some syndrome counts vary with season in a manner that is not identical from year to year. Our goal is to evaluate the impact of incon...
Autores principales: | , , , , , |
---|---|
Formato: | Texto |
Lenguaje: | English |
Publicado: |
BioMed Central
2006
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC1698911/ https://www.ncbi.nlm.nih.gov/pubmed/17144927 http://dx.doi.org/10.1186/1472-6947-6-40 |
_version_ | 1782131240749498368 |
---|---|
author | Burr, Tom Graves, Todd Klamann, Richard Michalak, Sarah Picard, Richard Hengartner, Nicolas |
author_facet | Burr, Tom Graves, Todd Klamann, Richard Michalak, Sarah Picard, Richard Hengartner, Nicolas |
author_sort | Burr, Tom |
collection | PubMed |
description | BACKGROUND: Syndromic surveillance (SS) can potentially contribute to outbreak detection capability by providing timely, novel data sources. One SS challenge is that some syndrome counts vary with season in a manner that is not identical from year to year. Our goal is to evaluate the impact of inconsistent seasonal effects on performance assessments (false and true positive rates) in the context of detecting anomalous counts in data that exhibit seasonal variation. METHODS: To evaluate the impact of inconsistent seasonal effects, we injected synthetic outbreaks into real data and into data simulated from each of two models fit to the same real data. Using real respiratory syndrome counts collected in an emergency department from 2/1/94–5/31/03, we varied the length of training data from one to eight years, applied a sequential test to the forecast errors arising from each of eight forecasting methods, and evaluated their detection probabilities (DP) on the basis of 1000 injected synthetic outbreaks. We did the same for each of two corresponding simulated data sets. The less realistic, nonhierarchical model's simulated data set assumed that "one season fits all," meaning that each year's seasonal peak has the same onset, duration, and magnitude. The more realistic simulated data set used a hierarchical model to capture violation of the "one season fits all" assumption. RESULTS: This experiment demonstrated optimistic bias in DP estimates for some of the methods when data simulated from the nonhierarchical model was used for DP estimation, thus suggesting that at least for some real data sets and methods, it is not adequate to assume that "one season fits all." CONCLUSION: For the data we analyze, the "one season fits all " assumption is violated, and DP performance claims based on simulated data that assume "one season fits all," for the forecast methods considered, except for moving average methods, tend to be optimistic. Moving average methods based on relatively short amounts of training data are competitive on all three data sets, but are particularly competitive on the real data and on data from the hierarchical model, which are the two data sets that violate the "one season fits all" assumption. |
format | Text |
id | pubmed-1698911 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2006 |
publisher | BioMed Central |
record_format | MEDLINE/PubMed |
spelling | pubmed-16989112006-12-19 Accounting for seasonal patterns in syndromic surveillance data for outbreak detection Burr, Tom Graves, Todd Klamann, Richard Michalak, Sarah Picard, Richard Hengartner, Nicolas BMC Med Inform Decis Mak Research Article BACKGROUND: Syndromic surveillance (SS) can potentially contribute to outbreak detection capability by providing timely, novel data sources. One SS challenge is that some syndrome counts vary with season in a manner that is not identical from year to year. Our goal is to evaluate the impact of inconsistent seasonal effects on performance assessments (false and true positive rates) in the context of detecting anomalous counts in data that exhibit seasonal variation. METHODS: To evaluate the impact of inconsistent seasonal effects, we injected synthetic outbreaks into real data and into data simulated from each of two models fit to the same real data. Using real respiratory syndrome counts collected in an emergency department from 2/1/94–5/31/03, we varied the length of training data from one to eight years, applied a sequential test to the forecast errors arising from each of eight forecasting methods, and evaluated their detection probabilities (DP) on the basis of 1000 injected synthetic outbreaks. We did the same for each of two corresponding simulated data sets. The less realistic, nonhierarchical model's simulated data set assumed that "one season fits all," meaning that each year's seasonal peak has the same onset, duration, and magnitude. The more realistic simulated data set used a hierarchical model to capture violation of the "one season fits all" assumption. RESULTS: This experiment demonstrated optimistic bias in DP estimates for some of the methods when data simulated from the nonhierarchical model was used for DP estimation, thus suggesting that at least for some real data sets and methods, it is not adequate to assume that "one season fits all." CONCLUSION: For the data we analyze, the "one season fits all " assumption is violated, and DP performance claims based on simulated data that assume "one season fits all," for the forecast methods considered, except for moving average methods, tend to be optimistic. Moving average methods based on relatively short amounts of training data are competitive on all three data sets, but are particularly competitive on the real data and on data from the hierarchical model, which are the two data sets that violate the "one season fits all" assumption. BioMed Central 2006-12-04 /pmc/articles/PMC1698911/ /pubmed/17144927 http://dx.doi.org/10.1186/1472-6947-6-40 Text en Copyright © 2006 Burr et al; licensee BioMed Central Ltd. http://creativecommons.org/licenses/by/2.0 This is an Open Access article distributed under the terms of the Creative Commons Attribution License ( (http://creativecommons.org/licenses/by/2.0) ), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. |
spellingShingle | Research Article Burr, Tom Graves, Todd Klamann, Richard Michalak, Sarah Picard, Richard Hengartner, Nicolas Accounting for seasonal patterns in syndromic surveillance data for outbreak detection |
title | Accounting for seasonal patterns in syndromic surveillance data for outbreak detection |
title_full | Accounting for seasonal patterns in syndromic surveillance data for outbreak detection |
title_fullStr | Accounting for seasonal patterns in syndromic surveillance data for outbreak detection |
title_full_unstemmed | Accounting for seasonal patterns in syndromic surveillance data for outbreak detection |
title_short | Accounting for seasonal patterns in syndromic surveillance data for outbreak detection |
title_sort | accounting for seasonal patterns in syndromic surveillance data for outbreak detection |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC1698911/ https://www.ncbi.nlm.nih.gov/pubmed/17144927 http://dx.doi.org/10.1186/1472-6947-6-40 |
work_keys_str_mv | AT burrtom accountingforseasonalpatternsinsyndromicsurveillancedataforoutbreakdetection AT gravestodd accountingforseasonalpatternsinsyndromicsurveillancedataforoutbreakdetection AT klamannrichard accountingforseasonalpatternsinsyndromicsurveillancedataforoutbreakdetection AT michalaksarah accountingforseasonalpatternsinsyndromicsurveillancedataforoutbreakdetection AT picardrichard accountingforseasonalpatternsinsyndromicsurveillancedataforoutbreakdetection AT hengartnernicolas accountingforseasonalpatternsinsyndromicsurveillancedataforoutbreakdetection |