Next Article in Journal
The Modulated Role of Toxoplasma gondii on Eosinophils in Psychiatric Disorders after Cannabis Cessation
Previous Article in Journal
A Systematic Review of Factors Associated with Mortality among Patients with Mycobacterium avium Complex Lung Disease
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Challenges and Opportunities in One Health: Google Trends Search Data

1
Richard A. Gillespie College of Veterinary Medicine, Lincoln Memorial University, Harrogate, TN 37752, USA
2
School of Mathematics and Science, Lincoln Memorial University, Harrogate, TN 37752, USA
*
Author to whom correspondence should be addressed.
Pathogens 2023, 12(11), 1332; https://doi.org/10.3390/pathogens12111332
Submission received: 9 August 2023 / Revised: 24 October 2023 / Accepted: 7 November 2023 / Published: 9 November 2023
(This article belongs to the Special Issue Data Science Challenges and Opportunities in One Health)

Abstract

:
Google Trends data can be informative for zoonotic disease incidences, including Lyme disease. However, the use of Google Trends for predictive purposes is underutilized. In this study, we demonstrate the potential to use Google Trends for zoonotic disease prediction by predicting monthly state-level Lyme disease case counts in the United States. We requested Lyme disease data for the years 2010–2021. We downloaded Google Trends search data on terms for Lyme disease, symptoms of Lyme disease, and diseases with similar symptoms to Lyme disease. For each search term, we built an expanding window negative binomial model that adjusted for seasonal differences using a lag term. Performance was measured by Root Mean Squared Errors (RMSEs) and the visual associations between observed and predicted case counts. The highest performing model had excellent predictive ability in some states, but performance varied across states. The highest performing models were for Lyme disease search terms, which indicates the high specificity of search terms. We outline challenges of using Google Trends data, including data availability and a mismatch between geographic units. We discuss opportunities for Google Trends data for One Health research, including prediction of additional zoonotic diseases and incorporating environmental and companion animal data. Lastly, we recommend that Google Trends be explored as an option for predicting other zoonotic diseases and incorporate other data streams that may improve predictive performance.

1. Introduction

Google Flu Trends (GFT) was a service operated by Google to predict outbreaks of flu and was discontinued in 2015 due to inaccurate predictions. GFT trends overestimated flu prevalence by over 50% in 2011–2012, which some researchers blamed on the increased media coverage and google searches for “swine flu” and “bird flu” [1]. A recent study indicated that a simple heuristic model predicted flu incidence better than the GFT black box algorithm [2]. However, Google Trends may still have potential to be an affordable, timely, robust, and sensitive surveillance system [3] with refinement of search terms, monitoring and updating of the algorithm, and use of additional data streams [1,4]. Google Trends data have been evaluated for their correlation with multiple zoonotic diseases, including Zika [5], salmonellosis [6], encephalitis [7], and Lyme disease [8]. These correlative studies show promise, although the use of Google Trends data for zoonotic disease prediction is underutilized.
Lyme borreliosis (Lyme disease) has been deemed a public health crisis and is reported at epidemic levels in certain geographic areas and is spreading to new geographic areas. Lyme disease is caused by the spirochete bacterium Borrelia burgdorferi and is vectored by Ixodes scapularis and I. pacificus (black-legged ticks) in the eastern and western United States, respectively [9,10]. The wildlife reservoirs for Borrelia include rodents in the genera Peromyscus, Sciurus and Tamias. The incidence and prevalence of Lyme disease depends on reservoir and other wildlife populations, environmental factors, tick seasonality and behavior, and landscape level habitat changes as well as other drivers [11,12]. Host response to infection can cause neurologic, cardiovascular, arthritic, and dermatologic issues throughout the stages of infection [13]. Early clinical signs include erythema migrans and early infections are frequently associated with neurological disease as well as arthralgia, fever, and headache [13,14]. Disseminated infection frequently includes arthritis as well as other complications. Most patients recover after treatment with antibiotics. Although approximately 30,000 cases of Lyme disease are reported to the CDC annually, estimated annual diagnosed cases are much higher, i.e., >450,000 [15,16], representing substantial economic and disease burden. Due to the expanding range of Borrelia in the United States, diagnosis should be considered based on clinical signs and history of exposure, especially in emerging areas [17].
Here, we demonstrate how Google Trends data can be used as a tool for the prediction of Lyme disease cases. We build on previous work from Kim et al., 2020 [18], who investigated the spatial-temporal associations of monthly Lyme disease incidence and Google Trend search data in the United States from 2011 to 2015 and found that there were similar patterns between the search patterns and incidence at the state-level and at the metro-level in Texas. However, the authors noted that validation of the method is needed due to the non-specific symptoms of Lyme that correspond to other conditions. In addition, the analysis was correlative rather than predictive. Therefore, we aimed to validate their findings by analyzing search terms for diseases with similar symptoms, including fibromyalgia, multiple sclerosis, and arthritis. In addition, we aimed to build predictive models for Lyme disease incidence by state to improve the utility of the models. The results of this paper serve as a case study for using Google Trends search data as a tool for the prediction of zoonotic disease incidences and to highlight benefits and potential barriers to using it as a tool.

2. Materials and Methods

2.1. Data Retrieval

The Lincoln Memorial University Institutional Review Board approved the study protocol (1075 V.0). Monthly state-level Lyme disease case count data from 2010 to 2021 were requested from multiple state public health departments or obtained from online repositories. Only states with 10 or more cases in 2019 were considered [19]. The final states included in the analysis were based on convenience, lack of missing or concerns regarding protection of individually identifiable health information, and data availability.
Google Trends search data were downloaded using the ‘gtrendsR’ package in R version 4.0.2. [20,21]. Google Trends reports data as “interest over time,” which ranges from 0 to 100 and represents the terms current interest level compared to its highest interest level (at 100). Search terms were selected by evaluating previous research [18] and through discussions of the primary literature and colloquial knowledge by the study team. The final list of search terms included terms for Lyme disease (“Lyme”, “Lyme disease”, and “Lymes”), tick (“seed tick”), symptoms of Lyme disease (“tick bite”, “bone pain”, “stiff neck”, “circular rash”, “brain fog”, tick fever”, “tick rash”, “bulls eye”, “droopy eye”, “muscle ache”, and “lethargy”), and diseases with similar symptoms to Lyme disease (“bells palsy”, “arthritis”, “fibromyalgia”, “multiple sclerosis”, “chronic fatigue”, “Summer Flu”, and “Rocky Mountain Spotted Fever”). The search terms for diseases with similar symptoms were used to test specificity of the search terms for Lyme disease and its symptoms for predicting Lyme disease case count.

2.2. Statistical Analysis

Twelve-month expanding window negative binomial regression models were built using the ‘rolling’ command in Stata version 17.0 [22] to predict the number of Lyme disease cases, after determining the data were over-dispersed. Separate models were built by search term, so in total, 22 models were tested. Predictors in the model included the current search term interest and a 12-month lag term to adjust for seasonal differences. Predictive ability was assessed in the test dataset via root mean squared error (RMSE) and through plots of the observed versus predicted counts. RMSE was calculated using the following equation for each observation (i), within state (j), within year (k), and within month (l) [23]:
RMSE = 1 n   i j k l = 1 n ( O i E i   ^ ) 2
where O is the observed Lyme disease case count and E is the expected, or predicted, case count. RMSE can be interpreted on the same scale as the outcome (Lyme disease case count) and is the average deviation of expected versus observed counts. Therefore, the lower the RMSE, the better the model is at predicting Lyme disease case count.

3. Results

The final sample included data from 16 states (Figure 1). Seven of the sixteen states are considered high-incidence states according to the CDC (https://www.cdc.gov/lyme/datasurveillance/lyme-disease-maps.html, accessed on 2 August 2023). All available data provided from 2010 to 2021 were used for the analysis, and states had variable levels of missing data (Table 1). Data notes and caveats supplied from health departments are listed in Supplementary File S1. Washington had the lowest amount of missing data, and Virginia had the highest amount of missing data. Descriptive statistics of the average monthly Lyme disease case counts stratified by state are summarized in Table 1.

Predictive Models

The strongest predictive terms were terms for Lyme disease, including “Lyme Disease”, “Lymes”, and “Lyme”, which had the lowest overall RMSE values (Table 2). The RMSE for “Lyme Disease” was 49.8, which can be interpreted as follows: on average, the model with search terms for “Lyme Disease” predicted within 49.8 cases of the actual case count. The interpretation and evaluation of RMSE depends on the scale of the outcome; therefore, RMSEs are expected to be smaller for low-incidence states. For example, for a high-incidence state where the model performed well (New Hampshire, range of 2–527 cases per month), the average RMSE was 73.1, meaning that on average the model predicted within 73.1 cases of the actual case count. For a low-incidence state where the model performed well (North Dakota, range of 0 to 21 cases a month), the RMSE was 3.8, meaning that on average, the model predicted within 3.8 cases of the actual case count. The worst performing terms were “tick bite”, “tick rash”, and “Rocky Mountain Spotted Fever”, indicating that Google Trends terms have some specificity for predicting Lyme disease. However, the average RMSEs ranged from 49.8 to 83.4 for all search terms, indicating that the search terms for Lyme disease were more predictive for Lyme disease, but not by a large margin.
We used a mean monthly Lyme disease case count as calculated from the data to define states into “very high incidence” (>78.6), “high incidence” (19.3–78.6) “low incidence” (3.9–19.2) and “very low incidence” (<3.9) categories for data presentation in Figure 2, Figure 3, Figure 4 and Figure 5. Prediction intervals are presented in Supplementary Table S1. Results for the best performing term “Lyme Disease” are presented. In some states and across all incidence levels, the predicted case counts closely follow the observed case counts, which indicated good predictive ability. The model appeared to perform best for North Dakota, Indiana, Michigan, Vermont, Connecticut, Maine, and New Hampshire. However, the model appeared to perform poorly for Kansas, Texas, Washington, California, Oregon, Rhode Island, and Virginia.

4. Discussion

Google Trends data are freely available and downloadable, which provides accessibility for researchers, epidemiologists, and health departments. Google Trends was used by the CDC for the prediction of yearly influenza cases, but eventually they discontinued use due to low predictive ability [1]. In this study, we assessed the predictive ability of Google search terms for monthly Lyme disease case count at the state level. We found that the models produced accurate predictions for many states, as demonstrated by the closeness of the predicted and observed case counts. In addition, the most predictive terms for Lyme disease case count were terms for Lyme disease, which indicates specificity for Google trends search term in predicting Lyme disease case counts. However, Google Trends underperformed in multiple states and there were no clear trends across incidence levels. We conclude that Google Trends data have the potential to be used as a tool for zoonotic disease incidence prediction in addition to other surveillance tools.
In this case study, we encountered numerous barriers to using Google Trends data to predict Lyme disease case count. First, Google Trends data may have underperformed in multiple states due to differences in how each health department tracks and reports Lyme disease case data. In the United States, each state health department tracks and reports Lyme disease data and there is not a centralized data system. The health departments then report yearly data to the CDC. Case definitions are not consistent across state or even across time, which may have also affected the performance of the models. Some state departments censor small cell sizes, so we were unable to include those states in the models. A barrier for data acquisition is that the system for requesting data in each state varies. Some states have data readily available for use on their official websites, whereas others require full Institutional Review Board review. Another challenge of using Google Trends data for disease prediction is the geographical units of the Google Trends search data. Google Trends data does not report at the county level, likely due to search volume and data privacy issues. The smallest geographical unit reported is at the metro-level, which is a geographical area that corresponds to a metropolitan area. Unfortunately, this does not correspond directly to county-level data, which is how most health departments report case data. Another challenge is selecting search terms. In the future, we recommend considering regional differences in terminology when selecting Google Trends search terms to potentially improve model performance in states with poor predictions, while also considering search volume. In less-populated states, some of our selected Google Trends search terms did not reach an adequate search volume to use in the models.
A nationwide, centralized data reporting system with standardized definitions for monthly Lyme disease cases would improve the feasibility of utilizing Google Trends for Lyme disease prediction. Currently, the CDC maintains a Lyme disease data dashboard, although the units reported are at the yearly level, which makes finer prediction not possible. Lyme disease cases are now reported at epidemic levels in some areas and there should be urgency in improving access to data [24]. Nonetheless, states could utilize Google Trends search data as a tool in their toolbox for disease surveillance in conjunction with other surveillance techniques, such as physician-reported cases, tick dragging, insurance claims, and wildlife or pet reports. This likely would require collaboration across state departments.
Future studies can determine if we can predict the spread of Lyme disease in new geographical locations and on a finer scale. In addition, future studies should investigate the inclusion of environmental, tick, and companion animal data for model refinement and to consider the full One Health triad. Future studies can also validate the findings of this case study in other zoonotic diseases and determine if the Lyme models could improve over time with additional data. There is a risk that with more media attention on Lyme disease, the models will be less predictive. However, using the expanding window approach where the model retrains every 12 months will help prevent media attention from affecting model performance.

5. Conclusions

In this study, we demonstrate the potential use of Google Trends search data for the prediction of monthly Lyme disease case counts at the state-level. Model performance varied by state. We outline challenges for Google Trends disease prediction, such as different case definitions and reporting procedures by state, data availability, and mismatch of Google Trends geographical units with county case counts. However, there are many opportunities for utilizing Google Trends data, as it is a free, publicly available resource and has not yet been tested for predictive ability for many zoonotic diseases. Integration of environmental, tick, and companion animal data is the next step to make it a true One Health model. Incorporating additional data sources may improve predictions for states where Lyme disease search terms did not accurately predict case counts.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/pathogens12111332/s1, File S1: Data notes and caveats; Table S1: Prediction intervals for Lyme disease case counts.

Author Contributions

Conceptualization, L.W.; methodology, L.W., K.G., V.F. and B.S.; formal analysis, L.W.; data curation, L.W.; writing—original draft preparation, L.W.; writing—review and editing, L.W., K.G., V.F. and B.S.; visualization, L.W. and K.G.; funding acquisition, L.W., K.G., V.F. and B.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board of Lincoln Memorial University (protocol code 1075 V.0. approved 15 February 2022).

Informed Consent Statement

Not applicable. Secondary analysis of de-identified case data was used for the analysis.

Data Availability Statement

Restrictions apply to the availability of Lyme disease case data. Supplementary File S1 includes links to data repositories for Connecticut, Indiana, Oregon, and Virginia. Data from the remaining states can be requested from individual health departments. Google Trends search volume data available on request from the corresponding author.

Acknowledgments

We would like to thank the following health departments for providing the Lyme disease case data for this project: California Department of Public Health; Kansas Department of Health and Environment; Maine Department of Health and Human Services; Michigan Department of Health and Human Services; New Hampshire Division of Public Health Services; North Dakota Department of Health; Rhode Island Department of Health, Division of Preparedness, Response, Infectious Disease and Emergency Medical Services, Center for Acute Infectious Disease Epidemiology; South Carolina Department of Health and Environmental Control; Texas Department of State Health Services; Vermont Department of Health; Washington State Department of Health; West Virginia Department of Health and Human Resources. We would also like to thank the following health departments with online data portals and reports that we used to obtain case data: Connecticut State Department of Public Health; Indiana Department of Health; Oregon Health Authority; Virginia Department of Health.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Lazer, D.; Kennedy, R.; King, G.; Vespignani, A. The parable of Google Flu: Traps in big data analysis. Science 2014, 343, 1203–1205. [Google Scholar] [CrossRef] [PubMed]
  2. Katsikopoulas, K.V.; Şimşek, Ö.; Buckmann, M.; Gigerenzer, G. Transparent modeling of influenza incidence: Big data or a single data point from psychological theory? Int. J. Forecast. 2022, 38, 613–619. [Google Scholar] [CrossRef]
  3. Carneiro, H.A.; Mylonakis, E. Google Trends: A web-based tool for real-time surveillance of disease outbreaks. Clin. Infect. Dis. 2009, 49, 1557–1564. [Google Scholar] [CrossRef] [PubMed]
  4. Zhang, Y.; Bambrick, H.; Mengersen, K.; Tong, S.; Hu, W. Using Google Trends and ambient temperature to predict seasonal influenza outbreaks. Environ. Int. 2018, 117, 284–291. [Google Scholar] [CrossRef] [PubMed]
  5. Morsy, S.; Dang, T.N.; Kamel, M.G.; Zayan, A.H.; Makram, O.M.; Elhady, M.; Hirayama, K.; Huy, N.T. Prediction of Zika-confirmed cases in Brazil and Colombia using Google Trends. Epidemiol. Infect. 2018, 146, 1625–1627. [Google Scholar] [CrossRef] [PubMed]
  6. Wang, M.-Y.; Tang, N.-J. The correlation between Google Trends and salmonellosis. BMI Public Health 2021, 21, 1575. [Google Scholar] [CrossRef] [PubMed]
  7. Sulyok, M.; Richter, H.; Sulyok, Z.; Kapitány-Fövény, M.; Walker, M.D. Predicting tick-borne encephalitis using Google Trends. Ticks Tick-Borne Dis. 2020, 11, e101306. [Google Scholar] [CrossRef] [PubMed]
  8. Kapitány-Fövény, M.; Ferenci, T.; Sulyok, Z.; Kegele, J.; Richter, H.; Vályi-Nagy, I.; Sulyok, M. Can Google Trends data improve forecasting of Lyme disease incidence? Zoonoses Public Health 2018, 66, 101–107. [Google Scholar] [CrossRef] [PubMed]
  9. Stanek, G.; Wormser, G.P.; Gray, J.; Strle, F. Lyme borreliosis. Lancet 2012, 379, 461–473. [Google Scholar] [CrossRef] [PubMed]
  10. Eisen, R.J.; Kugeler, K.J.; Eisen, L.; Beard, C.B.; Paddock, C.D. Tick-borne zoonoses in the United States: Persistent and emerging threats to human health. ILAR J. 2017, 58, 319–335. [Google Scholar] [CrossRef] [PubMed]
  11. Levi, T.; Kilpatrick, M.; Mangel, M.; Wilmers, C.C. Deer, predators, and the emergence of Lyme disease. Proc. Natl. Acad. Sci. USA 2012, 109, 10942–10947. [Google Scholar] [CrossRef]
  12. Kilpatrick, A.M.; Dobson, A.D.M.; Levi, T.; Salkeld, D.J.; Swei, A.; Ginsberg, H.S.; Kjemtrup, A.; Padgett, K.A.; Jensen, P.M.; Fish, D.; et al. Lyme disease ecology in a changing world: Consensus, uncertainty and critical gaps for improving control. Philos. Trans. R. Soc. B Biol. Sci. 2017, 372, e20160117. [Google Scholar] [CrossRef]
  13. Coburn, J.; Garcia, B.; Hu, L.T.; Jewett, M.W.; Kraiczy, P.; Norris, S.J.; Skare, J. Lyme disease pathogenesis. Curr. Issues Mol. Biol. 2021, 42, 473–517. [Google Scholar] [CrossRef] [PubMed]
  14. Kullberg, B.J.; Vrijmoeth, H.D.; van de Schoor, F. Lyme borreliosis: Diagnosis and management. BMJ 2020, 369, m0141. [Google Scholar] [CrossRef] [PubMed]
  15. Hook, S.A.; Jeon, S.; Niesobecki, S.A.; Hansen, A.P.; Meek, J.I.; Bjork, J.K.H.; Dorr, F.M.; Rutz, H.J.; Feldman, K.A.; White, J.L.; et al. Economic burden of reported Lyme disease in high-incidence areas, United States, 2014–2016. Emerg. Infect. Dis. 2022, 28, 1170–1179. [Google Scholar] [CrossRef] [PubMed]
  16. Kugeler, K.J.; Schwartz, A.M.; Delorey, M.J.; Mead, P.S.; Hinckley, A.F. Estimating the frequency of Lyme disease diagnosis, United States, 2010–2018. Emerg. Infect. Dis. 2021, 27, 616–619. [Google Scholar] [CrossRef] [PubMed]
  17. Schwartz, A.M.; Hinckley, A.F.; Mead, P.S.; Hook, S.A.; Kugeler, K.J. Surveillance for Lyme disease- United States, 2008–2015. MMWR 2017, 66, e22. [Google Scholar] [CrossRef] [PubMed]
  18. Kim, D.; Maxwell, S.; Le, Q. Spatial and temporal comparison of perceived risks and confirmed cases of Lyme Disease: An exploratory study of google trends. Front. Public Health 2020, 8, 395. [Google Scholar] [CrossRef] [PubMed]
  19. Surveillance Data. Available online: https://www.cdc.gov/lyme/datasurveillance/surveillance-data.html?CDC_AA_refVal=https%3A%2F%2Fwww.cdc.gov%2Flyme%2Fstats%2Fgraphs.html (accessed on 3 August 2023).
  20. R Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2021. [Google Scholar]
  21. Massicotte, P.; Eddelbuettel, D. GtrendsR: Perform and Display Google Trends Queries, R Package Version 1.5.1; 2022. Available online: https://CRAN.R-project.org/package=gtrendsR (accessed on 3 August 2023).
  22. StataCorp. Stata Statistical Software: Release 17; StataCorp LLC: College Station, TX, USA, 2021. [Google Scholar]
  23. Kuhn, M.; Johnson, K. Applied Predictive Modeling; Springer Nature: New York, NY, USA, 2013; pp. 95–100. [Google Scholar]
  24. Stricker, R.B.; Johnson, L. Lyme disease: Call for a “Manhattan Project” to combat the epidemic. PLOS Pathog. 2014, 10, e1003796. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Map displaying states included in analysis (dots) and by high (red) versus low (blue) incidence.
Figure 1. Map displaying states included in analysis (dots) and by high (red) versus low (blue) incidence.
Pathogens 12 01332 g001
Figure 2. Observed (blue line) versus predicted (red line) monthly Lyme disease case counts using the search term “Lyme disease” for very low-incidence states.
Figure 2. Observed (blue line) versus predicted (red line) monthly Lyme disease case counts using the search term “Lyme disease” for very low-incidence states.
Pathogens 12 01332 g002
Figure 3. Observed (blue line) versus predicted (red line) monthly Lyme disease case counts using the search term “Lyme disease” for low-incidence states.
Figure 3. Observed (blue line) versus predicted (red line) monthly Lyme disease case counts using the search term “Lyme disease” for low-incidence states.
Pathogens 12 01332 g003
Figure 4. Observed (blue line) versus predicted (red line) monthly Lyme disease case counts using the search term “Lyme disease” for high-incidence states.
Figure 4. Observed (blue line) versus predicted (red line) monthly Lyme disease case counts using the search term “Lyme disease” for high-incidence states.
Pathogens 12 01332 g004
Figure 5. Observed (blue line) versus predicted (red line) monthly Lyme disease case counts using the search term “Lyme disease” for very high-incidence states.
Figure 5. Observed (blue line) versus predicted (red line) monthly Lyme disease case counts using the search term “Lyme disease” for very high-incidence states.
Pathogens 12 01332 g005
Table 1. Descriptive statistics for monthly Lyme disease case count by state included in analysis (N = 1879 observations).
Table 1. Descriptive statistics for monthly Lyme disease case count by state included in analysis (N = 1879 observations).
StateNMeanSDMinimumMedianMaximum
California13210.07.01834
Connecticut108206.0170.311152.5860
Indiana8410.213.10451
Kansas1322.22.30210
Maine132113.3112.11271557
Michigan12019.225.209127
New Hampshire132106.1103.6264527
North Dakota1402.83.90121
Oregon9629.122.2125.589
Rhode Island10878.661.91456.5269
South Carolina1323.82.80315
Texas843.93.90316
Vermont13155.764.8128312
Virginia7287.663.8378261
Washington1442.53.20118
West Virginia13238.660.5017396
Table 2. Root mean squared error (RMSE) of predictions from model predicting monthly Lyme disease case count stratified by Google search term 1.
Table 2. Root mean squared error (RMSE) of predictions from model predicting monthly Lyme disease case count stratified by Google search term 1.
CACTINKSMEMINHNDORRISCTXVAVTWAWVAll
Symptoms
Bulls eye6.8136.613.32.3108.225.1106.2-21.9-2.73.363.3-3.3-63.4
Droopy eye6.8168.012.82.4114.325.2105.7-21.2-2.73.563.4-3.3 69.9
Stiff neck6.9152.213.22.3115.324.8101.23.621.659.12.73.757.267.53.352.962.5
Tick bite6.2249.238.92.0123.936.6130.99.122.145.32.53.661.677.24.958.083.4
Tick fever6.5159.217.82.1101.522.3101.5-22.161.52.72.658.865.73.361.864.9
Tick rash6.3197.313.02.0105.575.2105.1-22.162.82.73.362.168.72.946.673.7
Similar diseases
Arthritis7.0161.912.32.3110.824.1103.13.720.857.52.73.862.570.13.355.664.2
RMSF6.1172.6140.41.9107.024.8101.43.622.153.82.73.058.760.53.257.770.1
Summer flu5.9167.89.62.3114.319.9105.73.922.461.52.73.059.865.72.661.865.6
Lyme disease
Lyme6.2156.49.62.0114.424.469.94.922.232.12.63.760.653.53.342.157.3
Lyme disease6.596.89.72.0113.524.473.13.822.338.82.63.261.154.43.544.049.8
Lymes6.4120.98.02.1103.523.095.33.823.041.52.73.260.958.43.343.554.1
Other
Seed tick7.0168.012.62.4114.323.3103.3-22.661.52.83.455.265.73.361.867.9
1 Missing values in the table are due to low search volume. RMSF: Rocky Mountain Spotted Fever.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wisnieski, L.; Gruszynski, K.; Faulkner, V.; Shock, B. Challenges and Opportunities in One Health: Google Trends Search Data. Pathogens 2023, 12, 1332. https://doi.org/10.3390/pathogens12111332

AMA Style

Wisnieski L, Gruszynski K, Faulkner V, Shock B. Challenges and Opportunities in One Health: Google Trends Search Data. Pathogens. 2023; 12(11):1332. https://doi.org/10.3390/pathogens12111332

Chicago/Turabian Style

Wisnieski, Lauren, Karen Gruszynski, Vina Faulkner, and Barbara Shock. 2023. "Challenges and Opportunities in One Health: Google Trends Search Data" Pathogens 12, no. 11: 1332. https://doi.org/10.3390/pathogens12111332

APA Style

Wisnieski, L., Gruszynski, K., Faulkner, V., & Shock, B. (2023). Challenges and Opportunities in One Health: Google Trends Search Data. Pathogens, 12(11), 1332. https://doi.org/10.3390/pathogens12111332

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop