Article

Analysis of Web access logs for surveillance of influenza.

RODS Laboratory, Center for Biomedical Informatics, University of Pittsburgh, PA 15219, USA.
Studies in health technology and informatics 02/2004; 107(Pt 2):1202-6.
Source: PubMed

ABSTRACT The purpose of this study was to determine whether the level of influenza in a population correlates with the number of times that internet users access information about influenza on health-related Web sites. We obtained Web access logs from the Healthlink Web site. Web access logs contain information about the user and the information the user accessed, and are maintained electronically by most Web sites, including Healthlink. We developed weekly counts of the number of accesses of selected influenza-related articles on the Healthlink Web site and measured their correlation with traditional influenza surveillance data from the Centers for Disease Control and Prevention (CDC) using the cross-correlation function (CCF). We defined timeliness as the time lag at which the correlation was a maximum. There was a moderately strong correlation between the frequency of influenza-related article accesses and the CDC's traditional surveillance data, but the results on timeliness were inconclusive. With improvements in methods for performing spatial analysis of the data and the continuing increase in Web searching behavior among Americans, Web article access has the potential to become a useful data source for public health early warning systems.

1 Bookmark
 · 
250 Views
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: The internet and the World Wide Web have changed the ways that we function. As technologies grow and adapt, there is a huge potential for the internet to affect drug research and development, as well as many other aspects of clinical pharmacology. We review some of the areas of interest to date and discuss some of the potential areas in which internet-based technology can be exploited.Information retrieval from the web by health-care professionals is common, and bringing evidence-based medicine to the bedside affects the care of patients.As a primary research tool the web can provide a vast array of information in generating new ideas or exploring previous research findings. This has facilitated systematic reviewing, for example. The content of the web has become a subject of research in its own right.The web is also widely used as a research facilitator, including enhancement of communication between collaborators, provision of online research tools (such as questionnaires, management of large scale multicentre trials, registration of clinical trials) and distribution of information.Problems include information overload, ignorance of early data that are not indexed in databases, difficulties in keeping web sites up to date and assessing the validity of information retrieved. Some web-based activities are viewed with suspicion, including analysis by pharmaceutical companies of drug information to facilitate direct-to-consumer advertising of novel pharmaceuticals.Use of these technologies will continue to expand in often unexpected ways. Clinical pharmacologists must embrace internet technology and include it as a key priority in their research agenda.
    British Journal of Clinical Pharmacology 06/2012; 73(6). · 3.69 Impact Factor
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: This paper presents a survey of novel technologies for uncovering implicit knowledge through the analysis of user-contributed content in Web2.0 applications. The special features of emergent semantics are herein described, along with the various dimensions that the techniques should be able to handle. Consequently a series of application domains is given where the extracted information can be consumed. The relevant techniques are reviewed and categorised according to their capability for scaling, multi-modal analysis, social networks analysis, semantic representation, real-time and spatio-temporal processing. A showcase of such an emergent semantics extraction application, namely ClustTour, is also presented, and open issues and future challenges in this new field are discussed.
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: Infectious disease is a leading threat to public health, economic stability, and other key social structures. Efforts to mitigate these impacts depend on accurate and timely monitoring to measure the risk and progress of disease. Traditional, biologically-focused monitoring techniques are accurate but costly and slow; in response, new techniques based on social internet data, such as social media and search queries, are emerging. These efforts are promising, but important challenges in the areas of scientific peer review, breadth of diseases and countries, and forecasting hamper their operational usefulness. We examine a freely available, open data source for this use: access logs from the online encyclopedia Wikipedia. Using linear models, language as a proxy for location, and a systematic yet simple article selection procedure, we tested 14 location-disease combinations and demonstrate that these data feasibly support an approach that overcomes these challenges. Specifically, our proof-of-concept yields models with [Formula: see text] up to 0.92, forecasting value up to the 28 days tested, and several pairs of models similar enough to suggest that transferring models from one location to another without re-training is feasible. Based on these preliminary results, we close with a research agenda designed to overcome these challenges and produce a disease monitoring and forecasting system that is significantly more effective, robust, and globally comprehensive than the current state of the art.
    PLoS Computational Biology 11/2014; 10(11):e1003892. · 4.83 Impact Factor

Full-text (3 Sources)

Download
198 Downloads
Available from
May 26, 2014