Cargando…

Accounting for seasonal patterns in syndromic surveillance data for outbreak detection

BACKGROUND: Syndromic surveillance (SS) can potentially contribute to outbreak detection capability by providing timely, novel data sources. One SS challenge is that some syndrome counts vary with season in a manner that is not identical from year to year. Our goal is to evaluate the impact of incon...

Descripción completa

Detalles Bibliográficos
Autores principales: Burr, Tom, Graves, Todd, Klamann, Richard, Michalak, Sarah, Picard, Richard, Hengartner, Nicolas
Formato: Texto
Lenguaje:English
Publicado: BioMed Central 2006
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC1698911/
https://www.ncbi.nlm.nih.gov/pubmed/17144927
http://dx.doi.org/10.1186/1472-6947-6-40
_version_ 1782131240749498368
author Burr, Tom
Graves, Todd
Klamann, Richard
Michalak, Sarah
Picard, Richard
Hengartner, Nicolas
author_facet Burr, Tom
Graves, Todd
Klamann, Richard
Michalak, Sarah
Picard, Richard
Hengartner, Nicolas
author_sort Burr, Tom
collection PubMed
description BACKGROUND: Syndromic surveillance (SS) can potentially contribute to outbreak detection capability by providing timely, novel data sources. One SS challenge is that some syndrome counts vary with season in a manner that is not identical from year to year. Our goal is to evaluate the impact of inconsistent seasonal effects on performance assessments (false and true positive rates) in the context of detecting anomalous counts in data that exhibit seasonal variation. METHODS: To evaluate the impact of inconsistent seasonal effects, we injected synthetic outbreaks into real data and into data simulated from each of two models fit to the same real data. Using real respiratory syndrome counts collected in an emergency department from 2/1/94–5/31/03, we varied the length of training data from one to eight years, applied a sequential test to the forecast errors arising from each of eight forecasting methods, and evaluated their detection probabilities (DP) on the basis of 1000 injected synthetic outbreaks. We did the same for each of two corresponding simulated data sets. The less realistic, nonhierarchical model's simulated data set assumed that "one season fits all," meaning that each year's seasonal peak has the same onset, duration, and magnitude. The more realistic simulated data set used a hierarchical model to capture violation of the "one season fits all" assumption. RESULTS: This experiment demonstrated optimistic bias in DP estimates for some of the methods when data simulated from the nonhierarchical model was used for DP estimation, thus suggesting that at least for some real data sets and methods, it is not adequate to assume that "one season fits all." CONCLUSION: For the data we analyze, the "one season fits all " assumption is violated, and DP performance claims based on simulated data that assume "one season fits all," for the forecast methods considered, except for moving average methods, tend to be optimistic. Moving average methods based on relatively short amounts of training data are competitive on all three data sets, but are particularly competitive on the real data and on data from the hierarchical model, which are the two data sets that violate the "one season fits all" assumption.
format Text
id pubmed-1698911
institution National Center for Biotechnology Information
language English
publishDate 2006
publisher BioMed Central
record_format MEDLINE/PubMed
spelling pubmed-16989112006-12-19 Accounting for seasonal patterns in syndromic surveillance data for outbreak detection Burr, Tom Graves, Todd Klamann, Richard Michalak, Sarah Picard, Richard Hengartner, Nicolas BMC Med Inform Decis Mak Research Article BACKGROUND: Syndromic surveillance (SS) can potentially contribute to outbreak detection capability by providing timely, novel data sources. One SS challenge is that some syndrome counts vary with season in a manner that is not identical from year to year. Our goal is to evaluate the impact of inconsistent seasonal effects on performance assessments (false and true positive rates) in the context of detecting anomalous counts in data that exhibit seasonal variation. METHODS: To evaluate the impact of inconsistent seasonal effects, we injected synthetic outbreaks into real data and into data simulated from each of two models fit to the same real data. Using real respiratory syndrome counts collected in an emergency department from 2/1/94–5/31/03, we varied the length of training data from one to eight years, applied a sequential test to the forecast errors arising from each of eight forecasting methods, and evaluated their detection probabilities (DP) on the basis of 1000 injected synthetic outbreaks. We did the same for each of two corresponding simulated data sets. The less realistic, nonhierarchical model's simulated data set assumed that "one season fits all," meaning that each year's seasonal peak has the same onset, duration, and magnitude. The more realistic simulated data set used a hierarchical model to capture violation of the "one season fits all" assumption. RESULTS: This experiment demonstrated optimistic bias in DP estimates for some of the methods when data simulated from the nonhierarchical model was used for DP estimation, thus suggesting that at least for some real data sets and methods, it is not adequate to assume that "one season fits all." CONCLUSION: For the data we analyze, the "one season fits all " assumption is violated, and DP performance claims based on simulated data that assume "one season fits all," for the forecast methods considered, except for moving average methods, tend to be optimistic. Moving average methods based on relatively short amounts of training data are competitive on all three data sets, but are particularly competitive on the real data and on data from the hierarchical model, which are the two data sets that violate the "one season fits all" assumption. BioMed Central 2006-12-04 /pmc/articles/PMC1698911/ /pubmed/17144927 http://dx.doi.org/10.1186/1472-6947-6-40 Text en Copyright © 2006 Burr et al; licensee BioMed Central Ltd. http://creativecommons.org/licenses/by/2.0 This is an Open Access article distributed under the terms of the Creative Commons Attribution License ( (http://creativecommons.org/licenses/by/2.0) ), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Research Article
Burr, Tom
Graves, Todd
Klamann, Richard
Michalak, Sarah
Picard, Richard
Hengartner, Nicolas
Accounting for seasonal patterns in syndromic surveillance data for outbreak detection
title Accounting for seasonal patterns in syndromic surveillance data for outbreak detection
title_full Accounting for seasonal patterns in syndromic surveillance data for outbreak detection
title_fullStr Accounting for seasonal patterns in syndromic surveillance data for outbreak detection
title_full_unstemmed Accounting for seasonal patterns in syndromic surveillance data for outbreak detection
title_short Accounting for seasonal patterns in syndromic surveillance data for outbreak detection
title_sort accounting for seasonal patterns in syndromic surveillance data for outbreak detection
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC1698911/
https://www.ncbi.nlm.nih.gov/pubmed/17144927
http://dx.doi.org/10.1186/1472-6947-6-40
work_keys_str_mv AT burrtom accountingforseasonalpatternsinsyndromicsurveillancedataforoutbreakdetection
AT gravestodd accountingforseasonalpatternsinsyndromicsurveillancedataforoutbreakdetection
AT klamannrichard accountingforseasonalpatternsinsyndromicsurveillancedataforoutbreakdetection
AT michalaksarah accountingforseasonalpatternsinsyndromicsurveillancedataforoutbreakdetection
AT picardrichard accountingforseasonalpatternsinsyndromicsurveillancedataforoutbreakdetection
AT hengartnernicolas accountingforseasonalpatternsinsyndromicsurveillancedataforoutbreakdetection