Email updates

Keep up to date with the latest news and content from BMC Research Notes and BioMed Central.

Open Access Short Report

Perceived usefulness of a distributed community-based syndromic surveillance system: a pilot qualitative evaluation study

Blaine Reeder1*, Debra Revere2, Donald R Olson3 and William B Lober124

Author Affiliations

1 Biobehavioral Nursing and Health Systems School of Nursing, Box 359442 University of Washington, Seattle, WA 98195, USA

2 Department of Health Services School of Public Health, Box 354943 University of Washington, Seattle, WA 98195, USA

3 International Society for Disease Surveillance 26 Lincoln Street, 3 Brighton, MA 02135, USA

4 Medical Education and Biomedical Informatics School of Medicine, Box 357240 University of Washington, Seattle, WA 98195, USA

For all author emails, please log on.

BMC Research Notes 2011, 4:187  doi:10.1186/1756-0500-4-187


The electronic version of this article is the complete one and can be found online at: http://www.biomedcentral.com/1756-0500/4/187


Received:18 February 2011
Accepted:14 June 2011
Published:14 June 2011

© 2011 Reeder et al; licensee BioMed Central Ltd.

This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

Background

We conducted a pilot utility evaluation and information needs assessment of the Distribute Project at the 2010 Washington State Public Health Association (WSPHA) Joint Conference. Distribute is a distributed community-based syndromic surveillance system and network for detection of influenza-like illness (ILI). Using qualitative methods, we assessed the perceived usefulness of the Distribute system and explored areas for improvement. Nine state and local public health professionals participated in a focus group (n = 6) and in semi-structured interviews (n = 3). Field notes were taken, summarized and analyzed.

Findings

Several emergent themes that contribute to the perceived usefulness of system data and the Distribute system were identified: 1) Standardization: a common ILI syndrome definition; 2) Regional Comparability: views that support county-by-county comparisons of syndromic surveillance data; 3) Completeness: complete data for all expected data at a given time; 4) Coverage: data coverage of all jurisdictions in WA state; 5) Context: metadata incorporated into the views to provide context for graphed data; 6) Trusted Data: verification that information is valid and timely; and 7) Customization: the ability to customize views as necessary. As a result of the focus group, a new county level health jurisdiction expressed interest in contributing data to the Distribute system.

Conclusion

The resulting themes from this study can be used to guide future information design efforts for the Distribute system and other syndromic surveillance systems. In addition, this study demonstrates the benefits of conducting a low cost, qualitative evaluation at a professional conference.

Introduction

Distribute is a community-based, population-level public health information system for syndromic influenza-like illness (ILI) surveillance that displays aggregated, de-identified, public health surveillance data collected from emergency departments (EDs) by state and local health jurisdictions [1]. Distribute was first organized by the International Society for Disease Surveillance (ISDS) in 2006 as a proof of concept. With support from the Markle Foundation and the United States Centers for Disease Control and Prevention (CDC), Distribute grew from participation by 8 state and large metropolitan health jurisdictions, representing summarized data on 10% of all US emergency department (ED) visits, to a nation-wide system that currently receives data from 43 health jurisdictions that represents over 50% of the US population and summarizes more than 35% of all ED visits nationwide. Distribute currently has participation from all ten Health and Human Services (HHS) surveillance regions, includes data from over one million ED visits each week, and displays updated visualizations of ILI trends in the US on Public and Restricted access web sites.

Distribute serves as an example of a new paradigm in the collection and sharing of public health surveillance data [2]. The roots of automated syndromic surveillance systems began just prior to the 2001 Anthrax attacks [3] with systems which automatically classify clinic visits and other data according to loose "syndromic" criteria and present graphic and statistical views of summarized data based on counts of those visits, and their numbers in proportion to population and denominators derived from utilization of health care services. The literature describes both the early experience and growth of these systems [4] and their evolving design and implementation [5]. With the development of health information exchanges (HIEs), and of methods for the structuring of, and access to, regional data across multiple health care systems, public health gained access to larger sources of both visit level and summarized "syndromic" data [6]. In part, Distribute developed as a way for health departments to share and compare these summarized, syndromic data, regardless of whether those data were obtained from integration of data from individual providers or hospitals, or from a single large hospital network or HIE.

In the early development of the Distribute project, the apparent benefits of mandated standards for syndrome definitions were weighed against two often overlooked issues: barriers to entry and ability to compare data across jurisdictions. The need to adopt a mandated standard prior to joining the network created a potential technical barrier that could delay or prevent interested jurisdictions from participating. In addition, although mandated standard syndrome definitions could improve data comparison on average across the whole network, there was concern on the part of project participants that this practice might decrease accuracy and utility locally. That is, a local definition of a syndrome might best reflect local variations in coding or clinical practice that were reflected in the data, and might most accurately reflect the underlying disease being tracked. To address these issues, the Distribute project adopted the use of two separate syndromes: 1) a more narrow and specific definition, following a traditional clinical definition of ILI, and 2) another more sensitive definition, as a broad febrile, respiratory and influenza-like syndrome [7]. In a preliminary comparison, two Distribute participating sites shared local coding of their narrow and broad ILI syndrome definitions and applied each other's definitions to their own local data. The pilot findings suggested that data using locally applied syndromes were better correlated with population-level viral surveillance data [8,9].

Utility and Usability

Utility and usability issues are related and often difficult to separate in the evaluation of information systems [10,11]. Utility, or perceived usefulness, refers to the extent to which an information system or its output provides benefit or value [10-12]. Usability, or perceived ease of use, refers to the degree of effort required to use an information system or its output [10-12]. Many international standards for system design conflate usability and utility, incorporating aspects of utility and usability in a single definition [13]. Because this project is not an interaction study, we focus on the utility, or perceived usefulness, of the Distribute system and its data outputs while acknowledging that usability contributes to utility.

Current initiatives of the Distribute project place a high priority on improving the utility and usability of the information system and extending functionality to support public health decision-making and practice. Qualitative methods are important in the evaluation of health information systems [14-16]. It is important to engage practitioners in a discussion of their needs and proposed system features to mitigate common informatics risk factors for failed system adoption [17-19]. Following the idea that "the simplest way to assess usefulness is to ask those involved in public health practice"[20], we engaged epidemiologists and other public health practitioners in a pilot study to collect quality improvement feedback for the Distribute system. This pilot study was undertaken to inform the design of a larger quality improvement investigation by including participants who were current members of the Distribute community of practice and those interested in learning more about the Distribute system.

System Description

The objective of Distribute is to collect, analyze, and display ILI surveillance data from across the United States. Another objective is to provide ways to compare the progression of outbreaks of infectious disease between regions, and to enhance communication between health jurisdictions. Distribute displays summary level data from state and local health department ED surveillance systems in two views: Public and Restricted. The Public site provides public access graphs of weekly trends in ratios of ILI syndromes to all ED visits. The restricted site requires secure authentication for access and provides greater granularity in the time series data, primary counts as well as ratios, multiple syndromes, information about data upload history and transmission details, data timeliness, detailed visualizations based on user-specified queries and metadata that includes background information about each data provider and details of syndrome definitions. This secure view of information about all data-providing jurisdictions is available to each participating data provider on the restricted site. Distribute data providers upload data from existing surveillance systems, such as ESSENCE [21], EARS [22], RODS [23]and BioSense [24], at local and state health jurisdictions along with provider-specified syndrome definitions. Table 1 shows selected metadata elements available to data providers with restricted site access.

Table 1. Selected metadata elements from the Restricted site of the Distribute system

Methods

Setting

This study was conducted at the Washington State Public Health Association (WSPHA) Joint Conference on Health on October 11-12, 2010 in Yakima, WA USA. The conference is an annual meeting of public health practitioners that includes participants from the Washington State Department of Health and local health jurisdictions in the State of Washington. The study protocol received approval from the University of Washington Institutional Review Board.

Participants

A total of nine public health practitioners participated, representing state- (n = 5) and county- (n = 4) level public health organizations. Six participants attended a focus group and three participants engaged in brief interviews. Seven participants were current epidemiologists, health officers or other public health practitioners while two participants were former epidemiologists or health officers. Regarding prior use and/or familiarity with Distribute, four participants had never seen or used Distribute and of the five who were familiar with Distribute, only two participants had access to the Restricted site. All participants were familiar with how surveillance data are used for public health purposes. Table 2 shows a breakdown of study participants by role and data collection method.

Table 2. Study participants by role, agency level, familiarity with Distribute and method of data collection

Data Collection

We employed qualitative methods [25-27] to capture participants' perceived usefulness of data, data visualizations and features of the Distribute system and solicit quality improvement feedback from participants: 1) a focus group discussion of the Distribute system during a scheduled presentation and 2) three brief, semi-structured interviews with conference attendees who fit the profile of the target user group. Following a brief project background, participants were shown different views of graphed data for US Department of Human and Health Services (HHS) Regions from the Public site and Restricted site comparisons of graphed data from data providers in the Pacific Northwest for periods of time covering the 2009-2010 H1N1 and influenza seasonal time periods back to March 2009. Participants were asked to respond about perceived usefulness of views of graphed data and to discuss how they use surveillance data in their work. Focus group and interview participants were asked the same types of questions. Those familiar with Distribute were asked questions about their perceptions of the system. Table 3 shows the types of questions asked of participants.

Table 3. Types of questions asked of focus group and interview participants

Figure 1 shows a comparison graph used during data collection. The data-providing jurisdictions in the graph are de-identified for publication. The graph shows ILI visit time-series ratios of syndromic ED surveillance data during March 2009 through April 2010 from two Northwest US jurisdictions (Sites A and B) participating in Distribute. The ratios presented are ILI syndrome visits over total visits using syndrome definitions that are jurisdiction-specific and not standardized. While the absolute ILI levels are not directly comparable, the timing and relative magnitude show ILI trends representing the emergence and early spring wave of the influenza A/H1N1 pandemic and its autumn 2009 return.

thumbnailFigure 1. De-identified jurisdiction comparison showing the emergence and return of pandemic A/H1N1 influenza, March 2009-April 2010.

Data Analysis

Notes taken by Distribute team members during the focus group and interviews were summarized and analyzed to identify patterns and themes [28,29]. Study notes were stripped of identifying information before analysis. Conference attendees were referred to by role and an assigned study code. Specific data extracted from study notes pertained to opinions of conference attendees about usefulness, suggested features and other recommendations for improvement of the Distribute system. Themes were identified from focus group and interview notes by grouping similar responses and creating names and descriptions of the groupings [26,28].

Results

Overall, participants were engaged and positive about Distribute as a community-based network and syndromic surveillance system. Data analysis resulted in the emergent themes, which are displayed in Table 4 along with brief descriptions of each theme.

Table 4. Themes that contribute to information system and data usefulness

With regard to standardization, participants recognized a need for a common influenza-like illness (ILI) syndrome definition in order to make comparisons between data sets more meaningful and relevant. For regional comparability, participants wanted to see views of different regions to aid in comparisons. In particular, they expressed a desire for views that support county-by-county comparisons of syndromic surveillance data, separate regional views of Western Washington and Idaho and views by preparedness regions as an alternative if representative views of HHS Regions were unavailable.

The completeness theme is described by participant desire for completeness in the data sets submitted from each data provider. Coverage refers to participant desire to know that data are representative of a given population in order to generalize findings across the population. In particular, participants expressed a need for data coverage of all Washington State. Participants noted that during the second wave of H1N1, the eastern side of the state, which includes one-third of the population, initially saw two-thirds of all cases (consistent with the graph of data for Eastern Washington in Distribute). One participant noted minimal use of Distribute due to lack of close neighbors for comparison. Context refers to the expressed need for metadata incorporated into views to facilitate understanding of graphed data. Participants suggested inclusion of the number of hospitals, emergency departments, patients, data providers and denominators for the total number of ED visits as contextual information in the graphed data views.

Participants expressed a need for trusted data, that is, confidence that data are valid and timely. They wanted to know that data were verified through defined quality assurance processes that are conducted on a regular basis. Participants reported that these data are useful for consistent events such as seasonal influenza and local data could be used to declare a local epidemic. Consistent, reliable data were cited as more useful to stand-down from an emergency than to issue an initial alert; participants hypothesized that during H1N1 the data in Distribute might have been more useful after confirmation of an actual event occurrence. Customization of available data and data processing capabilities to meet the surveillance needs of each local and state health jurisdiction was requested by participants. The ability to overlay graphs with other graphs and create labels on request was envisioned as a useful feature.

Participants acknowledged the value of the Public site as a tool to view national trends. One participant cited the need for a surveillance system with a low-impact training cost that anyone can use and that is largely automated to minimize maintenance. Requests for additional data viewed as overlays to graphs included: metadata already available elsewhere in the system, county school absenteeism rates and a state view that includes data from all clinics in the Group Health Cooperative health care system. An additional result of the focus group was the expressed desire to participate as a data contributor by one participant from a county-level health jurisdiction.

Limitations

The limitations of this pilot study include its restricted time frame for data collection and the regional population from which the sample is drawn.

Conclusion

Our results suggest themes that can be used to guide future evaluation and design iterations to improve support for public health surveillance. These results are important for improvements to syndromic surveillance of influenza-like-illness in the Distribute system but can also help improve syndromic surveillance efforts overall, regardless of the disease or surveillance system. For example, gastrointestinal (GI) indicators are currently being piloted in the Distribute system as a demonstration of system extensibility for surveillance of other diseases. Themes from this qualitative evaluation study can inform GI syndromic surveillance efforts as they are expanded within Distribute or any other surveillance system. These themes should be further explored by including public health practitioners in information design efforts. In addition, this study demonstrates that the application of qualitative methods in an "evaluation of opportunity" at a public health practice gathering can be a simple way to solicit feedback for the improvement of a working public health information system. Lastly, we found that efforts of this type can be useful in recruiting new users to participate in the system and expand the community of practice.

The community-based approach employed in the Distribute project focuses on data use and has resulted in convergence toward a recognized need for a common influenza-like illness (ILI) syndrome definition to compare data sets across jurisdictions among the Distribute community of practice. The findings of this pilot study are consistent with this trend. However, to maintain local utility of data, existing data providers need not, and should not, abandon prior syndrome definitions, but rather should submit an additional common definition while continuing to send data aggregated by existing syndrome definitions that have local meaning.

Syndromic surveillance data, if available, are used by public health practitioners as early indicators of influenza outbreaks within their own jurisdictions and adjacent health jurisdictions. These data are used in conjunction with other data sources, such as laboratory results, to triangulate disease prevalence. To aid decision-making for interventions that help contain outbreaks, improved data access and visualizations for syndromic surveillance data are needed. The context of how data are used for individual tasks is important [30,31] and data quality cannot be assessed independent of the people who use them [32]. Information systems are part of the contexts of use for data; the utility and usability of these systems are factors in the utility and usability of data [13,33,34]. Three contexts of use for syndromic surveillance information systems - routine, anticipated threat and present threat - have been recognized as key inputs to tasks for analysis and characterization of syndromic surveillance data for decision-making [35]. In addition, our pilot results indicate contextual information - metadata related to hospitals, patients, data providers, ED visit counts, etc. - provides meaning of syndromic surveillance data to epidemiologists.

To better understand contexts of information system and data use, future efforts should identify the specific ways in which epidemiologists and others use metadata to discern meaning from data, the best ways to include annotations in data visualizations and different ways to display information for population health surveillance. Interviews with a larger number of participants will help refine the specific meanings of our themes, gauge reactions to anticipated results from common ILI syndrome definition efforts and explore specific needs around regionalization and other identified themes. Future work to engage a more geographically diverse population of participants will help validate these results outside the pilot region of Washington State. This future work will be informed by pilot results in two related areas: 1) usability studies to improve the design of Distribute as an information resource that an epidemiologist might check before making a phone call to a colleague in a different jurisdiction or region and 2) utility studies to assess the value of Distribute to participants, their organizations and community population health outcomes.

Competing interests

All authors declare financial support from the International Society of Disease Surveillance. The authors declare that they have no other competing interests.

Authors' contributions

BR conceived of and implemented the study design with DR, acted as investigator/observer, collected and analyzed data and authored the overall manuscript. DR conceived of and implemented the study design with BR, authored the protocol, acted as investigator, collected and analyzed data and authored the overall manuscript. DRO contributed expertise and manuscript content related to the Distribute project and common ILI definitions. WBL contributed expertise and manuscript content about informatics and syndromic surveillance. All authors read and approved the final manuscript.

Acknowledgements

The authors gratefully acknowledge the International Society for Disease Surveillance as the financial sponsor of this study and the Markle Foundation and CDC for supporting the Distribute project. We thank Rebecca Hills, PhD (Cand.), MSPH for her contributions during the study and Laura Streichert, PhD, MPH and Janet Baseman, PhD, MPH for their comments on drafts of the manuscript. We also thank the two reviewers of this manuscript for their helpful suggestions. This study would not have been possible without the public health professionals who donated their time to share information about their work and participation of the Distribute community of practice.

References

  1. Olson DR, Paladini M, Buehler J, Mostashari F: Review of the ISDS Distributed Surveillance Taskforce for Real-time Influenza Burden Tracking & Evaluation (DiSTRIBuTE) Project 2007/08 Influenza Season Proof-of-concept Phase.

    Advances in Disease Surveillance 2008, 5:185. OpenURL

  2. Diamond CC, Mostashari F, Shirky C: Collecting And Sharing Data For Population Health: A New Paradigm.

    Health affairs 2009, 28:454. PubMed Abstract | Publisher Full Text OpenURL

  3. Lober WB, Karras BT, Wagner MM, Overhage JM, Davidson AJ, Fraser H, Trigg LJ, Mandl KD, Espino JU, Tsui FC: Roundtable on bioterrorism detection: information system-based surveillance.

    J Am Med Inform Assoc 2002, 9:105-115. PubMed Abstract | Publisher Full Text | PubMed Central Full Text OpenURL

  4. Mandl KD, Overhage JM, Wagner MM, Lober WB, Sebastiani P, Mostashari F, Pavlin JA, Gesteland PH, Treadwell T, Koski E, et al.: The Practice of Informatics - Review Paper - Implementing Syndromic Surveillance: A Practical Guide Informed by the Early Experience.

    Journal of the American Medical Informatics Association: JAMIA 2004, 11:141. PubMed Abstract | Publisher Full Text | PubMed Central Full Text OpenURL

  5. Lober WB, Trigg L, Karras B: Information system architectures for syndromic surveillance.

    MMWR Morb Mortal Wkly Rep 2004, 53(Suppl):203-208. PubMed Abstract | Publisher Full Text OpenURL

  6. Hills RA, Lober WB, Painter IS, Workshop: Biosurveillance, Case Reporting, and Decision Support: Public Health Interactions with a Health Information Exchange.

    2008.

  7. Olson DR, Heffernan RT, Paladini M, Konty K, Weiss D, Mostashari F: Monitoring the Impact of Influenza by Age: Emergency Department Fever and Respiratory Complaint Surveillance in New York City.

    PLoS Med 2007, 4:e247. PubMed Abstract | Publisher Full Text | PubMed Central Full Text OpenURL

  8. Pendarvis J, L. ME, Paladini M, Gunn J, Olson DR: Age Specific Correlations between Influenza Laboratory Data and Influenza-like Syndrome Definitions in Boston and New York City.

    Advances in Disease Surveillance 2008, 5:53. OpenURL

  9. Paladini M, Pendarvis J, Murray EL, Gunn J, Olson DR: A Comparison of Locally Developed Influenza-like Syndrome Definitions Using Electronic Emergency Department Data in Boston and New York City.

    Advances in Disease Surveillance 2008, 5:50. OpenURL

  10. Grudin J: Utility and usability: research issues and development contexts.

    Interacting with Computers 1992, 4:209-217. Publisher Full Text OpenURL

  11. McLaughlin J, Skinner D: Developing Usability and Utility: A Comparative Study of the Users of New IT.

    Technology Analysis & Strategic Management 2000, 12:413-423.

    Routledge; 413-423

    PubMed Abstract | Publisher Full Text OpenURL

  12. Davis FD: Perceived Usefulness, Perceived Ease of Use, and User Acceptance of Information Technology.

    MIS Quarterly 1989., 13 OpenURL

  13. Bevan N: International standards for HCI and usability.

    International Journal of Human-Computer Studies 2001, 55:533-552. Publisher Full Text OpenURL

  14. Ash JS, Guappone KP: Qualitative evaluation of health information exchange efforts.

    Journal of Biomedical Informatics 2007, 40:S33-S39. PubMed Abstract | Publisher Full Text | PubMed Central Full Text OpenURL

  15. Rose AF, Schnipper JL, Park ER, Poon EG, Li Q, Middleton B: Using qualitative studies to improve the usability of an EMR.

    Journal of Biomedical Informatics 2005, 38:51-60. PubMed Abstract | Publisher Full Text OpenURL

  16. Hill HK, Stewart DC, Ash JS: Health Information Technology Systems profoundly impact users: a case study in a dental school.

    Journal of dental education 2009, 74:434-445. OpenURL

  17. Wells S, Bullen C: A near miss: the importance of context in a public health informatics project in a New Zealand case study.

    Journal of the American Medical Informatics Association: JAMIA 2008., 15 OpenURL

  18. Littlejohns P, Wyatt JC, Garvican L: Evaluating computerised health information systems: hard lessons still to be learnt.

    BMJ 2003, 326:860-863. PubMed Abstract | Publisher Full Text | PubMed Central Full Text OpenURL

  19. Heeks R: Health information systems: Failure, success and improvisation.

    International journal of medical informatics 2006, 75:125. PubMed Abstract | Publisher Full Text OpenURL

  20. Thacker SB, Parrish RG, Trowbridge FL: A method for evaluating systems of epidemiological surveillance.

    World health statistics quarterly Rapport trimestriel de statistiques sanitaires mondiales 1988, 41:11-18. PubMed Abstract OpenURL

  21. Lombardo J, Burkom H, Elbert E, Magruder S, Lewis SH, Loschen W, Sari J, Sniegoski C, Wojcik R, Pavlin J: A systems overview of the Electronic Surveillance System for the Early Notification of Community-Based Epidemics (ESSENCE II).

    Journal of urban health: bulletin of the New York Academy of Medicine 2003, 80:32-42. OpenURL

  22. Hutwagner L, Thompson W, Seeman GM, Treadwell T: The Bioterrorism Preparedness and Response Early Aberration Reporting System (EARS).

    Journal of Urban Health: Bulletin of the New York Academy of Medicine 2003, 80:89-96. OpenURL

  23. Tsui FC, Espino JU, Dato VM, Gesteland PH, Hutman J, Wagner MM: Technical description of RODS: a real-time public health surveillance system.

    Journal of the American Medical Informatics Association 2003, 10:399-408. PubMed Abstract | Publisher Full Text | PubMed Central Full Text OpenURL

  24. Bradley CA, Rolka H, Walker D, Loonsk J: BioSense: Implementation of a National Early Event Detection and Situational Awareness System.

    IMMW MMWR Morbidity & Mortality Weekly Report 2005, 11-19. PubMed Abstract | Publisher Full Text OpenURL

  25. Blomberg J, Giacomi J, Mosher A, Swenton-Wall P: Ethnographic Field Methods and Their Relation to Design. In Participatory design: principles and practices. Edited by Schuler D, Namioka A. Hillsdale, N.J.: L. Erlbaum Associates; 1993:123-155. OpenURL

  26. Krueger RA, Casey MA: Focus groups: a practical guide for applied research. Thousand Oaks, Calif.: Sage Publications; 2000. OpenURL

  27. Ulin PR, Robinson ET, Tolley EE: Qualitative methods in public health: a field guide for applied research. San Francisco, CA: Jossey-Bass; 2005. OpenURL

  28. Boyatzis RE: Transforming qualitative information: thematic analysis and code development. Thousand Oaks, CA: Sage Publications; 1998. OpenURL

  29. Miles MB, Huberman AM: Qualitative data analysis: an expanded sourcebook. Thousand Oaks: Sage Publications; 1994. OpenURL

  30. Wang RY, Strong DM: Beyond Accuracy: What Data Quality Means to Data Consumers.

    Journal of management information systems: JMIS 1996, 12:5. OpenURL

  31. Bevan N: Quality in use: Meeting user needs for quality.

    JSS Journal of Systems & Software 1999, 49:89-96. PubMed Abstract | Publisher Full Text OpenURL

  32. Strong DM, Lee YW, Wang RY: Data Quality in Context.

    Communications of the ACM 1997, 40:103. OpenURL

  33. Orr K: Data quality and systems theory.

    Commun ACM 1998, 41:66-71. OpenURL

  34. Boddy D, King G, Clark JS, Heaney D, Mair F: The influence of context and process when implementing e-health.

    BMC Medical Informatics & Decision Making 2009, 9:1-9.

    BioMed Central; 1-9

    PubMed Abstract | BioMed Central Full Text | PubMed Central Full Text OpenURL

  35. International Society for Disease Surveillance (ISDS) Meaningful Use Workgroup: Final Recommendation: Core Processes and EHR Requirements for Public Health Syndromic Surveillance.

    2010.