Hostname: page-component-788cddb947-nxk7g Total loading time: 0 Render date: 2024-10-14T03:29:12.693Z Has data issue: false hasContentIssue false

Understanding the leading indicators of hospital admissions from COVID-19 across successive waves in the UK

Published online by Cambridge University Press:  04 September 2023

Jonathon Mellor
Affiliation:
UK Health Security Agency, Data, Analytics and Surveillance, Nobel House, London, UK
Christopher E Overton
Affiliation:
UK Health Security Agency, Data, Analytics and Surveillance, Nobel House, London, UK Department of Mathematical Sciences, University of Liverpool, Liverpool, UK Department of Mathematics, University of Manchester, Manchester, UK
Martyn Fyles
Affiliation:
UK Health Security Agency, Data, Analytics and Surveillance, Nobel House, London, UK Department of Mathematics, University of Manchester, Manchester, UK
Liam Chawner
Affiliation:
UK Health Security Agency, Data, Analytics and Surveillance, Nobel House, London, UK
James Baxter
Affiliation:
UK Health Security Agency, Data, Analytics and Surveillance, Nobel House, London, UK
Tarrion Baird
Affiliation:
UK Health Security Agency, Data, Analytics and Surveillance, Nobel House, London, UK Department of Pathology, University of Cambridge, Cambridge, UK
Thomas Ward*
Affiliation:
UK Health Security Agency, Data, Analytics and Surveillance, Nobel House, London, UK
*
Corresponding author: Thomas Ward; Email: Tom.Ward@ukhsa.gov.uk
Rights & Permissions [Opens in a new window]

Abstract

Following the end of universal testing in the UK, hospital admissions are a key measure of COVID-19 pandemic pressure. Understanding leading indicators of admissions at the National Health Service (NHS) Trust, regional and national geographies help health services plan for ongoing pressures. We explored the spatio-temporal relationships of leading indicators of hospitalisations across SARS-CoV-2 waves in England. This analysis includes an evaluation of internet search volumes from Google Trends, NHS triage calls and online queries, the NHS COVID-19 app, lateral flow devices (LFDs), and the ZOE app. Data sources were analysed for their feasibility as leading indicators using Granger causality, cross-correlation, and dynamic time warping at fine spatial scales. Google Trends and NHS triages consistently temporally led admissions in most locations, with lead times ranging from 5 to 20 days, whereas an inconsistent relationship was found for the ZOE app, NHS COVID-19 app, and LFD testing, which diminished with spatial resolution, showing cross-correlation of leads between –7 and 7 days. The results indicate that novel surveillance sources can be used effectively to understand the expected healthcare burden within hospital administrative areas though the temporal and spatial heterogeneity of these relationships is a key determinant of their operational public health utility.

Type
Original Paper
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2023. Published by Cambridge University Press

Introduction

The cessation of mass community testing has hampered the ability to understand contemporary estimates of localised COVID-19 growth. Epidemiological prevalence surveillance studies such as the Office of National Statistics (ONS) COVID-19 Infection Survey (CIS) were produced with inconsistent spatial sampling and significant reporting lags for real-time public health policy. Therefore, the surveillance of hospital admissions and bed occupancy may more reliably capture the true growth of COVID-19 in the community in addition to the current pressures on health services. Understanding leading indicators of hospital admissions, with improved spatial precision, allows the National Health Service (NHS) to appropriately prepare, and policymakers to plan interventions. Hospital admissions are less influenced by testing ascertainment rates than by community testing; however, they are more impacted by the age composition of incidence due to the age severity gradient and are therefore less able to detect early growth in younger ages.

A variety of model structures have been employed to predict COVID-19 admission dynamics. For example, incorporating local testing has been shown to improve admission forecasting at fine spatial scales over autoregressive models [Reference Meakin, Abbott, Bosse, Munday, Gruson, Hellewell, Sherratt, Chapman, Prem, Klepac, Jombart, Knight, Jafari, Flasche, Waites, Jit, Eggo, Villabona-Arenas, Russell, Medley, Edmunds, Davies, Liu, Hué, Brady, Pung, Abbas, Gimma, Mee, Endo, Clifford, Sun, McCarthy, Quilty, Rosello, Sandmann, Barnard, Kucharski, Procter, Jarvis, Gibbs, Hodgson, Lowe, Atkins, Koltai, Pearson, Finch, Wong, Quaife, O’Reilly, Tully and Funk1]. More complex Bayesian structural time series modelling techniques have been developed across several countries to forecast dynamics nationally [Reference Feroze2]; however, these univariate time series models struggle at epidemic turning points. In addition, causal approaches aimed at capturing herd immunity effects have been employed [Reference Friston, Flandin and Razi3], though these methods rely on assumption-driven scenarios. Transmission modelling has been used throughout the pandemic; however, these models are fit to coarse spatial scales to reduce model uncertainty [Reference Birrell, Blake, van Leeuwen, Gent and de Angelis4], which limits their operational utility.

There has been work exploring indirect surveillance approaches for disease incidence. For example, perturbations in Google Trend’s search terms have been shown to precede cases and deaths at a national level [Reference Lampos, Majumder, Yom-Tov, Edelstein, Moura, Hamada, Rangaka, McKendry and Cox5] and incorporated in neural network architectures to forecast clinical risk in the UK [Reference Ward, Johnsen, Ng and Chollet6]. These Google Trend search queries also work with other epidemic metrics of interest by improving model predictive performance of case rates, hotspot detection [Reference McDonald, Bien, Green, Hu, DeFries, Hyun, Oliveira, Sharpnack, Tang, Tibshirani, Ventura, Wasserman and Tibshirani7], and deaths at a US state level [Reference Ma and Yang8]. However, the surveillance data sources are not limited to search engine records; during the COVID-19 pandemic mobility measurements, social media and wearable technology have been explored to forecast healthcare pressures [Reference Gerlee, Karlsson, Fritzell, Brezicka, Spreco, Timpka, Jöud and Lundh9Reference Conroy, Silva, Mehraei, Damiano, Gross, Salvati, Feng, Schneider, Olson, Rizzo, Curtin, Frassica and McFarlane12]. These applications are broader than just COVID-19 – digitised syndromic surveillance (including search engines, news reports, social media, clinician search queries, and crowdsourcing apps) has been used effectively to monitor other disease pressures including influenza and Zika [Reference Ginsberg, Mohebbi, Patel, Brammer, Smolinski and Brilliant13Reference Lampos and Cristianini17].

Indicator–admissions temporal relationships are not consistent due to changes in behaviour, population immunological response, testing coverage, and antigenic drift/shift. Novel variants of COVID-19 have had distinct epidemiological characteristics affecting this temporal relationship. The extent of a variant’s immunological evasion to prior immunity impacts the incidence growth rate and the rate of spatial dispersion [Reference Jassat, Abdool Karim, Mudara, Welch, Ozougwu, Groome, Govender, von Gottberg, Wolter, Wolmarans, Rousseau, Blumberg and Cohen18]. Novel variants have unique severity profiles [Reference Wrenn, Pakala, Vestal, Shilts, Brown, Bowen, Strickland, Williams, Mallal, Jones, Schmitz, Self and das19], for instance, relative to wild type, the Alpha variant was estimated to have a 62% (HR – 1.62 (95% CI: 1.48, 1.78)) increased risk of hospitalisation [Reference Grint, Wing, Houlihan, Gibbs, Evans, Williamson, McDonald, Bhaskaran, Evans, Walker, Hickman, Nightingale, Schultze, Rentsch, Bates, Cockburn, Curtis, Morton, Bacon, Davy, Wong, Mehrkar, Tomlinson, Douglas, Mathur, MacKenna, Ingelsby, Croker, Parry, Hester, Harper, DeVito, Hulme, Tazare, Smeeth, Goldacre and Eggo20]. This evolving relationship between infection and hospitalisation impacts the temporal relationship between indicators and admissions. The COVID-19 vaccination campaign began in the UK in December 2020, reaching 150 million total doses across the first, second, spring, and autumn booster doses [Reference Curtis, Inglesby, Morton, MacKenna, Green, Hulme, Walker, Morley, Mehrkar, Bacon, Hickman, Bates, Croker, Evans, Ward, Cockburn, Davy, Bhaskaran, Schultze, Rentsch, Williamson, Rowan, Fisher, McDonald, Tomlinson, Mathur, Drysdale, Eggo, Wing, Wong, Forbes, Parry, Hester, Harper, O’Hanlon, Eavis, Jarvis, Avramov, Griffiths, Fowles, Parkes, Douglas, Evans, Smeeth and Goldacre21, 22]. The vaccination, in combination with high population infection attack rates over successive waves of SARS-CoV-2 incidence, has led to an increasingly complex picture of immunity, at individual and population levels, against SARS-CoV-2 infection [Reference Dan, Mateus, Kato, Hastie, Yu, Faliti, Grifoni, Ramirez, Haupt, Frazier, Nakao, Rayaprolu, Rawlings, Peters, Krammer, Simon, Saphire, Smith, Weiskopf, Sette and Crotty23, Reference Khoury, Najjar-Debbiny, Elemy, Jabbour, Haj, Abu-Sini, Yasin, Amin, Hellou, Nasrallah, Saffouri and Hakim24], which impacts syndromic surveillance efforts to understand the spatio-temporal infection burden.

We have evaluated leading indicators of COVID-19 hospitalisations during the Omicron BA.1, BA.2, and BA.4/5 variant waves of 2021/2022. This analysis has been conducted at a National Health Service (NHS) Trust geographic scale (local groups of secondary care providers) [25]. We use a variety of methods to assess temporal relationships between the indicator and COVID-19 hospital admission at a high spatial resolution, including Granger causality, cross-correlation analysis, and dynamic time warping.

Methodology

The data assessed were available at different geographic designations and with varying quality. Hospital admission counts by date are provided by the NHS England (NHSE) daily COVID-19 hospital situational report [26]. This contains Trust-level hospital admissions stratified by age, with bed occupancy and staff absence counts. Google Trends [27] was curated to capture search query trends relevant to syndromic surveillance of COVID-19. NHS 111 calls and online pathways [28] were provided by the NHS, with COVID-19-relevant treatment pathways extracted. ZOE Health [29] provided counts of crowdsourced self-reported symptoms from the ZOE app. Lateral flow device (LFD) testing data were accessed from the UK COVID-19 dashboard [30]. Aggregated NHS COVID-19 app [31] metrics were extracted from data made available by the UKHSA.

Several other data sources were explored for feasibility but were excluded from this study as they were determined to be of limited utility in an operational context. Data sources with a transfer/access latency of greater than one week were excluded, as were data sources that lagged admissions. Primary care and general practitioner calls were excluded due to incompleteness of their national spatial coverage. School attendance reports were not evaluated due to the substantial reporting lag and data availability, which hampers utility in operational settings. PCR testing data were excluded due to changes in mass testing policies, which impacted eligibility and spatial coverage. Care home data were excluded due to highly heterogeneous spatial coverage and a lagging relationship with community transmission. The Office for National Statistics Covid Infection Survey of community positivity was explored, and we estimated a high correlation with hospital admission, see Supplementary Section A, Figure S1. However, further analysis of these data was not conducted as the quantity of sampled tests in this infection study was not informative at smaller geographic levels (like NHS Trust) nor is it released in a timeframe useful for real-time analysis. Data availability is given in Supplementary Table S1. We analysed the data sources in real time; however, for lateral flow device data, the analysis was conducted using specimen date, which is impacted by data correction over time.

Data

Admissions

English hospital admissions, provided by NHS England [29], are reported at a Trust level - a collection of secondary care providers. A COVID-19 admission is defined as a patient who had a positive test upon arrival to hospital, or within the past 24 hours while an inpatient. These counts therefore include admissions for COVID-19, incidental presentations, and hospital acquired infection. NHSE data are provided daily, with each individual Trust submitting the web form by 11 am for the preceding 24 hours. Due to the fast operational turnaround and considerable number of hospitals, there are some missing entries per day and occasional inaccurate values reported. The admissions data were reconstructed to provide one record per Trust per day, with resulting missing data being imputed using the last observation carried forward. Organisational mergers were coded manually ensuring records were accurate to the end of the study date. Hospitals with fewer than 10 admissions in 2022 or clear non-acute specialisations were removed from the analysis as they represented misreporting, non-COVID specialisations, or purely incidental admissions, leaving 121 acute Trusts. Admissions are presented from 01 October 2021 to the 29 August 2022, covering the end of the Delta plateau and the BA.1, BA.2, and BA.4/5 Omicron waves in 2021/2022.

Surveillance data

Google Trends

A large set of potentially predictive Google Trends search terms were analysed in previous research to determine which terms should be evaluated [Reference Ward, Johnsen, Ng and Chollet6]. Initially, over 1,000 terms were captured from the most common phrases used within NHS Pathway 111 telephonic COVID-19 triages, COVID-19 symptoms, over-the-counter medicine, and natural language variations on requests for tests. These terms were screened for relevance and relative occurrence at a national level within the Google Trends web interface. Analysis at a national level using generalised additive models with a negative binomial error structure and dynamic time warping were used to assess each term’s relevance for COVID-19 incidence.

From the wide range of potential terms, 84 COVID-19-relevant Google Trends search term volume scores were collected hourly and aggregated to daily scores per lower tier local authorities (LTLAs) across the UK [Reference Ward, Johnsen, Ng and Chollet6]. An LTLA is an administrative geography within the UK, capturing a district of local government, smaller than a county. The Google Trends scores relate to a ranked relative search volume in a city. We transform the city-level data to LTLA using a coordinate mapping. The LTLA granularity for the London region is limited as Google defines London as central London, and then outer London areas, reducing precision when mapping to London LTLAs. Due to the considerable number of Google Search Terms collected, similar terms were grouped together with the aim of increasing the signal-to-noise ratio. COVID-19 symptom search terms were grouped into ‘common’, ‘rare’, and ‘severe’ terms. Terms relating to general symptoms and tests were combined. General COVID-19 terms such as ‘coronavirus’ were grouped, and terms such as ‘tier system’ relating to policies no longer in effect were combined. The terms and processing logic used are outlined in Supplementary Section B.

NHS 111 pathways

NHS 111 provides non-emergency health advice based on an individual’s symptoms [30]. A user follows the triage process, inputs symptoms and receives healthcare advice (treatment), either through the online or telephone service. Data are presented as counts by day for given pathway outcomes, stratified by age, gender, and Lower Super Output Area (LSOA), an area more granular than LTLA. COVID-19 treatments are aggregated together into clinical assessment, ambulance, or self-care – stratified by age. The ages were binned into three groups to increase the counts per group. These age groups were 0–19, 20–59, and 60 and over.

LFD tests

Lateral flow devices (LFDs) are self-administered rapid tests allowing for real-time detection of COVID-19 by an individual. The tests were provided free universally until 01 April 2022, with the data obtained from the UK COVID-19 Dashboard [31]. An aggregation of positive and total test counts was made available daily at the LTLA level. There can be latency in this dataset due to upload delays for the specimen date of test; however, we did not have access to the historical real-time data, and therefore, analysis included the complete backfilled data. From the LFD data, a positivity rate for tests was calculated using positive and total test counts. A metric of test counts per capita was calculated using the associated population size of an NHS Trust.

NHS COVID-19 app

The NHS COVID-19 app [Reference Menni, Valdes, Freidin, Sudre, Nguyen, Drew, Ganesh, Varsavsky, Cardoso, el-Sayed Moustafa, Visconti, Hysi, Bowyer, Mangino, Falchi, Wolf, Ourselin, Chan, Steves and Spector32] produces daily aggregated metrics of app events for analysis. These events covered both app-specific metrics, such as downloads, app store ratings, and users, as well as contact tracing-relevant processes such as notifications of exposure. The data were provided at an LTLA level, with a weekly release schedule within the UKHSA, with epidemiologically relevant events – contact exposure notifications of users and reported positive tests via the app interface.

ZOE COVID-19 study app

The ZOE app allows users to self-report COVID-19 relevant symptoms daily with the aim of capturing an up-to-date picture of the pandemic [Reference Meakin, Abbott and Funk33]. The data are stratified by age, ethnicity, sex, healthcare worker status, and LTLA. Due to the sparsity of counts at this high resolution, the counts were aggregated as total symptom counts per LTLA per day. The different symptom counts were combined into categories, ‘common’, ‘severe’, ‘rare’, and ‘irrelevant’. Groupings for all variables are further described in Supplementary Section C, Table S2.

Processing

Spatial mapping

NHS 111, ZOE app, LFD tests, and NHS COVID app leading indicators are reported in LTLA geographies (or geographies that are strict subregions of LTLAs) and therefore cannot be aggregated directly to hospital admissions at a Trust level. A mapping is therefore required to relate LTLA-level data sources to NHS Trust admissions. Using the methodology based on the covid19.nhs.data R package [Reference Meakin, Abbott and Funk33], we produced a more contemporary probabilistic mapping using count data from the SUS APC (Secondary Use Service, All Patient Care) hospitals’ admitted patient database. The data extracted were from the 6 months preceding the study and contain test-confirmed admissions (at a Trust level) and discharge locations (the associated LTLA). From these records, we calculate a proportion of people from a local area (LTLA) who were admitted to a specified Trust. Using the proportion of people from an LTLA who attend a Trust, we use the residential population size of that LTLA to calculate the weighted population size for a Trust, referred to as population size in this study. The LTLA residential population counts were obtained from the 2019 mid-year population estimates of each LTLA from the Office for National Statistics. The same mapping between LTLA and Trust allows us to convert LTLA indicator data sources to the hospital Trust level. The distribution of populations across NHS Trusts is shown in Supplementary Figure S2, compared to counties and LTLAs, showing they are of a comparable scale. For the Google Trends data, a mapping was used, which combines London geographies. Leading indicator sources are first aggregated to the LTLA level and then transformed with the Trust–LTLA mapping with a weighted sum to determine the effective impact of an indicator on a Trust.

Scaling and smoothing

All indicators were smoothed using a locally estimated scatterplot smoothing (LOESS) method to reduce noise and scaled between 0 and 1 by Trust to allow comparison between hospitals. For dynamic time warping, the variables were z-score normalised.

Evaluation

Data sources were evaluated against present admissions and admissions in 14 days. We looked at present admissions to understand the use of the indicator as a real-time proxy for admissions. Following discussion with public health practitioners, 14 days was determined as a meaningful time window to act upon leading indicator insight. Consultation on the time length included those creating situational awareness products, senior public health leaders, and colleagues within the NHS. Leading relationships less than 14 days are still of interest for surveillance.

To understand how the leading relationships have changed over time, the data are analysed across the recent Omicron epidemic waves, with each wave being evaluated independently for the Granger causality and cross-correlation approaches. The timings defined for these waves are given in Supplementary Table S3.

Granger causality

The Granger causality test estimates if a time series (indicator) can linearly forecast another time series (admissions), not whether there is a causal relationship [Reference Granger34]. The test uses lagged time series and a combination of t-tests and f-tests to determine if the indicator meaningfully adds explanatory power for predicting admissions. Two regression models were constructed, one contains the explanatory time series (indicator) and the other does not. The comparison between the two models tells us whether the explanatory time series adds useful information in predicting the response. We have two time series $ {x}_t $ and $ {y}_t $ , the indicator and admissions, respectively, at time $ t $ . We then construct two regression equations, giving $ {y}_t $ explained firstly by lags of $ {y}_t $ and lags of $ {x}_t $ and secondly by lags of $ {y}_t $ alone

(1) $$ {y}_t={\alpha}_0+{\sum}_{j=1}^m{\alpha}_j{y}_{t-j}+{\sum}_{j=1}^m{\beta}_j{x}_{t-j}+{\epsilon}_t, $$
(2) $$ {y}_t={\alpha}_0+{\sum}_{j=1}^m{\alpha}_j{y}_{t-j}+{\epsilon}_t, $$

where $ {\alpha}_j $ and $ {\beta}_j $ are the regression coefficients for lag $ j $ . The null hypothesis is that

$$ {H}_0:{\beta}_1={\beta}_2=\dots ={\beta}_m=0, $$

and an f-test is performed on the two models (1) and (2) to determine the effect of $ x $

$$ F=\frac{\left(\frac{\left( RS{S}_1- RS{S}_2\right)}{p_2-{p}_1}\right)}{\left(\frac{RS{S}_2}{n-{p}_2}\right)}, $$

where $ n $ is the number of data points, $ {p}_1 $ and $ {p}_2 $ are the number of parameters in (1) and (2), and RSS is the residual sum of squares. We take the maximum as lag $ m=3 $ ; therefore, lags 1, 2, and 3 days are used, with larger numbers of lags reducing the power of the tests. Due to the spatial variation in trends and behaviours at the hospital level, Granger causality tests were performed per Trust, rather than at higher aggregations. A test is performed between an indicator and current hospital admissions, as well as between the indicator and hospital admissions in 14 days, to test the relationship within a practically useful temporal distance. Using both times, we can understand whether an indicator leads admissions and if the indicator leads admissions enough to be useful.

Cross-correlation

A linear time delay analysis using cross-correlation functions (CCFs) allows us to calculate the cross-correlation between indicators and admissions, producing scores over different lead times [Reference Vio and Wamsteker35].

Given two times $ {x}_t $ and $ {y}_t $ where $ t=0,1,2\dots N-1 $ and that $ {m}_x $ and $ {m}_y $ are the respective means, then the cross-correlation $ {R}_{xy} $ at delay $ d $ is

$$ {R}_{xy}=\frac{\sum \limits_i\left[\Big({x}_t-{m}_x\right)\times \left({y}_{t-d}-{m}_y\right)\Big]}{\sqrt{\sum \limits_i{\left({x}_t-{m}_x\right)}^2}\sqrt{\sum \limits_i{\left({y}_{t-d}-{m}_y\right)}^2\;}}. $$

We define an ‘optimal lead time’ as the lead day $ d $ , fewer than 30 days, with maximum CCF between indicator and admissions. Within 30 days was selected to avoid detecting periodic effects in the growth–peak–decline–plateau cycle of admissions.

Dynamic time warping

Dynamic time warping (DTW) calculates the non-linear alignment between two sequences of values. The algorithm creates a mapping (warping curve) between the sequences, which we analyse to understand how indicators and admissions relate over time. The algorithm aims to find the minimal path along the warping curve which aligns the two time series, applied using the R dtw package [Reference Giorgino36]. Further details on how the method works are provided in [Reference Senin37].

We aim to find the optimal warping curve $ \phi $ between $ {x}_i $ and $ {y}_j $ . For DTW, each index in time series $ {x}_i $ must match at least one time index $ {y}_j $ , and the subsequent matches from one index to the next must be monotonically increasing. The algorithm finds the best index matching, which minimises the size of $ \phi $ , the sum of absolute differences in value for each matched pair of indexes.

$$ D\left(x,y\right)=\underset{\phi }{\min }{d}_{\phi}\left(x,y\right). $$

The warping between time series can be calculated in a multivariate context (multiple pairs of time series) by minimising the joint distance across the indexes of $ x $ and $ y $ , $ i $ and $ j $ along the column $ C $ .

$$ d{\left(i,j\right)}^2={\sum}_{c=1}^C{\left({x}_{ic}-{x}_{jc}\right)}^2. $$

To produce sensible results, we place restrictions on the warping curve. Firstly, a 35-day window is used (specifically a Sakoechiba window) to avoid unrealistically long lead times. An asymmetric step pattern with a P2 slope constraint was chosen to allow for a leading alignment between the time series. The specifics of these parameters are addressed in further detail within the DTW literature [Reference Senin37, Reference Sakoe and Seibi38]. From the DTW, we can understand the alignment and lead/lag relationship between the time series in a non-linear fashion, identifying leads at different points in time and epidemic phases by analysing which index pairs are matched. Open start and end conditions were chosen to best capture the leading relationships as not all sequence points of the time series are available, which can cause a beginning or end ‘bunching’ effect, distorting calculated lead times [Reference Tormene, Giorgino, Quaglini and Stefanelli39]. We analyse the alignments by calculating the difference between the index of the indicator with its matched index in the admissions series, which tells us how far the indicator is ahead of the admissions – a lead time. Additionally, as part of the DTW algorithm, we can extract the cumulative warping $ \phi $ between the indicator and admissions, which is a measure of how much manipulation is needed (how different the time series are) to map between the sequences – the warping distance. As time series can be different lengths, or have partial matches, we use the normalised distance to compare how well the indicators match admissions – with greater distances corresponding to larger leads.

Data operations

To be a useful indicator, a data source must be available in near-real-time to capitalise on leading information detected and make decisions before the associated admissions occur. Data sources with substantial reporting latency are not viable candidates for operational indicators. Such a ‘leading indicator’ may have led the measurement in question, but if they are not available to analyse in real time, they cannot have an impact on real-time decisions and were therefore not assessed. Additionally, completion lags, or backfilling, were considered as a major limitation for some data sources as they are unreliable in the near term.

Results

The indicators in this study vary in their magnitude and timing relative to the successive Omicron admission waves – examples of this effect are shown in Figure 1.

Figure 1. A time series plot of example variables from each data source’s indicators, aggregated nationally, with reference to national admissions (red dashed line). Indicators that lead admissions well should appear shifted leftward of the admissions line. Indicators and admissions are scaled between 0 and 1 to allow for easy visual comparison of temporal offsets.

Granger causality

With the assumption of linearity, the Granger tests are used to examine the utility of an indicator in predicting future hospital admissions. A low centred distribution of p-values corresponds to an indicator with a strong leading relationship across many Trusts. Figure 2 describes the p-value of the leading relationship of an indicator. The NHS 111 and Google Trends indicators show consistent leading relationships across waves, for some variables. For other indicators such as LFD tests and the NHS COVID app, there is a high variation in relationship strength across waves. Each wave has unique spatio-temporal indicator relationships; for example, the LFD tests were estimated to have the most robust leading association with hospital admissions during the BA.2 period of growth, whereas the strength of association with NHS 111 variables reduced during this wave.

Figure 2. The distribution of p-values across Trusts from Granger causality tests. The tests are performed at the Trust level for each separate wave between indicator and admissions. Low p-values tell us the indicator leads admissions well, given the conditions of the test.

Figure 3 shows whether an indicator has a lead at 14 days or greater, which is indicative of whether actionable insight can be derived for real-world applications. Lower distributed p-values indicate a leading relationship at 14 days. Again, there are a majority of Trusts with low p-values for the Google and NHS 111 variables, though for the waves available, the ZOE symptom data does lead well. However, as the ZOE data are truncated for BA.2, the Granger causality is potentially just capturing the growth phase, which will be easier to linearly predict using Granger causality. Even the strongest indicators have multiple Trusts with high p-values, highlighting the spatial variation in relationship.

Figure 3. The distribution of p-values across Trusts from Granger causality tests. The tests are performed at the Trust level for each separate wave between indicators and admissions in 14 days. Low p-values tell us the indicator leads admissions in 14 days well given the conditions of the test.

Cross-correlation

We used cross-correlation analysis to assess each indicator at a range of temporal offsets with admissions. We extracted optimal lead times and CCF values across Trusts and waves. By calculating the cross-correlation between the indicator and admission under linear relationship assumptions at each time offset, we approximated the optimal lead time, Figure 4, and quantify how strong the relationship at 14 days is, Figure 5. It is important that we both look at ‘best lead time’ (optimal lead) and the strength at 14 days to understand the nuances of the relationship. Though we are interested operationally in the correlation at 14 days, a given indicator may, in fact, have a stronger relationship at greater than 14 days, which could be more useful. An indicator may have a large optimal lead time, but that large lead is not useful if the correlation for the indicator is small. Conversely, there may be a small optimal lead, with a highly correlated tracking time series but high CCF at 14 days, due to a wide peak in CCF values across temporal offsets.

Figure 4. The distribution, via boxplot, of optimal lead times between the indicator and admissions across all Trusts for each wave. Optimal lead is defined as the lead time with the maximum non-negative CCF value within a 30-day forward and backward window. Larger optimal leads will be most useful for forecasting, wide ranges in optimal lead correspond to high variation spatially, and smaller variation indicates a more consistent lead. The indicators with higher average (across waves) optimal leads are sorted to the left.

Figure 5. The distribution of CCF values for the indicator and admissions at 14 days lead across Trusts for each wave. High CCF values correspond to a high correlation between time series, and CCF values centred around zero would show that an indicator does not have a meaningful temporal lead against admissions. High variation in the CCF values show how consistent leading relationships are across the different Trusts. The indicators with higher average (across wave) CCF values are sorted to the left.

The calculation of optimal lead time is sensitive to its definition and to perturbation, if CCF peaks are wide across lead times. As some proportion of Trusts will not have a leading relationship for an indicator, we expect there to be optimal leads that are either negative (lags) or noisy across waves. An example cross-correlation function visualisation is provided in Supplementary Figure S3, showing the correlation between NHS 111 self-care for ages 0–19 years and hospital admissions.

An optimal lead time does not necessarily imply a strong leading relationship, it indicates that the specific time offset is the lead of highest correlation. For example, as the maximum window is 30 days, maximum correlations at 30 days may indicate a poor signal or out-of-phase relationship. From an analysis of the distribution of optimal lead times in Figure 4, we can see there is substantial variation across the waves for individual indicators. For Google Trends and NHS 111, there is a wide range in optimal lead times for the BA.1 wave, whereas for the LFD variables, there are instead lagged relationships, centred below zero. For a given indicator, the sign of the lead can vary substantially across Trusts. While there is a median lead of >10 days for the ZOE symptom variables, the lead reduces in the truncated BA.2 wave, as shown in Figure 1.

By looking at the cross-correlation offset at 14 days, we can see how related the indicators are with admissions 14 days in the future. A high CCF value indicates a strong relationship at a fixed temporal offset, though variables are likely to have stronger relationships (peak CCFs) at different offsets. We can see the CCF values for the indicators in Figure 5, where there is a high variation between Trusts and variant waves. Some of the indicators are much more highly correlated than others, though which are the strongest varies wave on wave. Most LFD indicators show stronger relationships in the later waves, for example. There are some negative CCF values, such as NHS 111 Clinical 0–19, which may mean the signal lags admissions – though it may also indicate the signals are out of phase at this specific time offset. The NHS COVID app has consistently strong cross-correlation across indicators and waves.

Dynamic time warping

We first explore an example dynamic time warping (DTW) for a single Trust with a single indicator to demonstrate the concept, Figure 6. There is a clear leading relationship between the indicator and admissions, shown by the diagonal index matches, though the effect decreases in the successive waves. For BA.2, there is a clearer tracking alignment displayed with vertical matches, with peaks at the same indexes for indicator and admission.

Figure 6. Dynamic time warping mapping between indicators and admissions used to generate lead times. The DTW is shown for a single indicator and Trust. The solid time series represents the variable being evaluated, the indicator, the dashed are admissions, and the lines between are the aligned sequence pairs. Vertical lines indicate no temporal offset between time series.

For this analysis, we focus on the multivariate case of the DTW algorithm, with all Trusts analysed simultaneously for a given indicator, with an example shown in Supplementary Figure S4. An initial period of the Delta plateau sequence is included to avoid boundary effects at the start of the series in the DTW algorithm; however, the sequence values are excluded from the reported results. Due to the study time period, an end period following the BA.4/5 wave was not available; therefore, we expect boundary effects to be present in this wave, particularly for the Google Trends, where the time series is further truncated.

In Figure 7, we show how the alignment lead times are distributed within a wave. Indicators such as Google Trends (Common, Rare, Entity, Symptom + Test) and NHS 111 (Clinical 0–19, Ambulance 0–19, 20–59) had leads of greater than 10 days for the BA.1 and BA.2 waves. Using the linear cross-correlation method however, the Google Rare indicator had a smaller lead in BA.2, highlighting how the results are sensitive to approach. For some LFD variables, the performance increased over successive waves. For variables with strong leads, the increase was not necessarily replicated for successive waves.

Figure 7. The distribution of lead times calculated from sequence index matching between indicators and admissions across the different epidemic waves of study. The lead times correspond to the optimal time warping between indicator and admissions, with a higher value indicating a larger temporal lead.

The normalised distance is a measure of misalignment between the indicator and admissions produced from the dynamic time warping. While the step pattern chosen does not strictly produce a lead alignment for all types of signals, other analysis in this manuscript has shown linear lead times >0; we therefore assume that the normalised distances produced can be treated as a proxy for lead time length. A smaller normalised distance corresponds to a small warp required to align the indicator and admission, which means the indicator tracks the admissions well. Poorly aligned indicators therefore have greater normalised distances and high corresponding lead. Using a normalised distance allows comparison of timeseries of different lengths, such as the ZOE app data and Google Trends, which are truncated. Within Figure 8, we see the NHS 111 and Google Trends have larger normalised distances than other indicators – they are not as aligned. The result may be because they have a large lead/lag, which requires a large warping, or due to having a relationship that is not consistent through time and location. The COVID app, LFD tests, and ZOE app have smaller normalised distances, indicating a smaller lead/lag or more consistent time shifts across geographies.

Figure 8. The normalised path distance produced by warping the space between the indicator and admissions across all three waves. Produced using multivariate DTW across Trusts. The normalised path distance indicates how much total warping is needed between indicators and admissions, a proxy for how big a lead time there is between time series.

Data operations

To be a useful indicator, data must be available in a timely and complete manner. While leading relationships can be evaluated statistically for historical data, how useful an indicator is in a practical setting is a result of how quickly analysis can be made available. If there are significant reporting/completeness lags in data being presented, then it will erode any actionable lead times. The different release frequencies and reporting/completeness lags are given in Table 1 for those indicators explored in this study.

Table 1. The operational considerations at time of investigation

Note: A 1-day reporting/completeness lag corresponds to yesterdays’ collected data being available today. More frequent releases allow for more flexibility in times of increased demand for analysis, particularly in a fast-changing epidemic. The reporting/completeness lag ‘eats into’ lead times, reducing how useful a leading indicator is in practice.

Google Trends is gathered in real time via an API, therefore available without latency, which allows for daily analysis. The NHS 111 data are provided via egress from the NHS, which introduces a day lag. The LFD testing and NHS COVID-19 app’s weekly release schedule reduce the operational effectiveness of their use as an analysis or model may be needed for policymaking at different points in the week relative to when the data are released. The one-off provision of ZOE data makes the operational effectiveness and limitations unknown without further exploration. How quickly data can be collected and transferred to analysts directly impacts its utility for decision making.

Discussion

Using a variety of statistical approaches, we highlight the potential utility of novel surveillance data, displaying their strengths and limitations for use as leading indicators of hospital admissions with COVID-19 at a high spatial resolution. The strength of the leading indicator relationship changes temporally and spatially across resurgent waves of COVID-19 incidence, and the implications of our results show a range of indicators should be used to accurately capture epidemic trends.

Consistent with the start of the pandemic [Reference Kogan, Clemente, Liautaud, Kaashoek, Link, Nguyen, Lu, Huybers, Resch, Havas, Petutschnig, Davis, Chinazzi, Mustafa, Hanage, Vespignani and Santillana11], we show that Google Trends is an effective leading indicator of the healthcare burden associated with SARS-CoV-2 transmission when extended to fine spatial resolutions. Existing work on COVID indicator analysis at fine spatial scales explores cases [Reference Meakin, Abbott, Bosse, Munday, Gruson, Hellewell, Sherratt, Chapman, Prem, Klepac, Jombart, Knight, Jafari, Flasche, Waites, Jit, Eggo, Villabona-Arenas, Russell, Medley, Edmunds, Davies, Liu, Hué, Brady, Pung, Abbas, Gimma, Mee, Endo, Clifford, Sun, McCarthy, Quilty, Rosello, Sandmann, Barnard, Kucharski, Procter, Jarvis, Gibbs, Hodgson, Lowe, Atkins, Koltai, Pearson, Finch, Wong, Quaife, O’Reilly, Tully and Funk1], which we expand upon to show that there are other novel, viable indicators at these scales. Building on the data collection approach used to determine clinical risk at low spatial geographies [Reference Ward, Johnsen, Ng and Chollet6], this analysis shows the Google Trends data’s utility can be extended to hospital admissions at a hospital Trust, rather than administrative geography. While also analysing leading indicators of hospital admissions, we extend the work of forecasting US hospital admissions in 2021 [Reference Wang, Ma, Baek and Yang40] to show effectiveness of Google Trends as a leading indicator within the Omicron waves.

The ZOE app was created to measure current COVID-19 symptomatic prevalence [Reference Menni, Valdes, Freidin, Sudre, Nguyen, Drew, Ganesh, Varsavsky, Cardoso, el-Sayed Moustafa, Visconti, Hysi, Bowyer, Mangino, Falchi, Wolf, Ourselin, Chan, Steves and Spector32]. We found that there was a temporal lead during the BA.1 wave that degrades by the BA.2 wave – likely due to declining usership. We introduce the NHS COVID-19 app, which as well as reducing COVID-19 incidence [Reference Wymant, Ferretti, Tsallis, Charalambides, Abeler-Dörner, Bonsall, Hinch, Kendall, Milsom, Ayres, Holmes, Briers and Fraser41] could have utility as a forecasting data source. Though the NHS COVID-19 app’s usefulness will depend on continued widespread usage and user notifications [42]. Previous literature shows that univariate epidemic signals have limited reliability across waves [Reference O’brien and Clements43], which corroborates our findings that there will be differences between waves, admissions, and indicators. This analysis shows that NHS 111 Pathways are a valuable leading indicator of COVID-19 hospital admissions, in agreement with work done within a single NHS Trust [Reference Rostami-Tabar and Rendon-Sanchez44]. We extend these findings across England, beyond the first COVID-19 wave, and further enrich the data using the 111 Online Pathway.

There is large spatial variation in indicators’ relationships with Trust-level admissions. At higher level geographical scales and in aggregate, there are trends, but these trends are not always reliable at a local level – with changing lead times in different waves and locations. The explanation for the change will depend on the indicator (such as testing policy for LFDs, media interest for Google Trends, and app use uptake for ZOE and NHS COVID-19 app), which implies the indicators’ useful in the most recent wave are not guaranteed to work in the next wave – requiring active monitoring. Even within wave, there is a high variation, as shown by the variety in DTW lead times, which implies that leading indicators may be more performant at different epidemic phases. As the public conscious and government policies shift away from COVID-19, and other respiratory pathogens resume circulation, we would expect a degradation in the strength of some of these signals.

These findings have implications for developing forecasting models using these indicators. The large spatial variation implies that relying on Trust level-indicators alone will perform poorly for a proportion of Trusts. By using pooling techniques via hierarchical modelling, the Trusts with stronger leading indicators could inform those with weaker relationships. The temporal variation shows that we cannot assume a single fixed temporal relationship between indicators and admissions. Therefore, to account for the variation, a range of temporal offsets should be considered by a model for use in prediction. The changing relationship between epidemic waves has multiple implications – primarily that no single indicator signal should be considered entirely temporally and spatially reliable for use in a forecasting model fit to historic data. As the relationships over time change substantially, the effect should either be corrected for, or ad hoc signal exclusions applied. By using a range of indicators, rather than single variables, there is more likely to be a leading relationship in the set of variables chosen. Since individual indicators can be impacted by external events (policy changes, data collection, etc.), multiple different sources of indicators should be utilised.

Limitations and further study

There are several limitations to this analysis which reflect either data availability or further research required. The Google Trends and ZOE app data were not available, or changed collection format from the start to the end of the period of study, which reduces the reliability of the conclusions drawn from these data and prevents direct comparison with other indicators. The analysis of leading relationships included techniques that assume linear relationships (Granger causality and cross-correlation). If non-linearity is present, this assumption could provide a partial picture as to whether there is a leading relationship. Further research should employ more complex methods such as kernel approaches to Granger causality [Reference Marinazzo, Pellicoro and Stramaglia45].

The addition of dynamic time warping to this analysis sheds light on the potential non-linear relationships between indicator and admissions. The dynamic time warping showed a positive lead for the indicators with complete data, which supports this study’s aim. The scaling and smoothing of the different indicators and admissions will impact the results of the different analysis substantially since the low-level indicator trends are noisy. The scaling of signals across the whole study period will decrease the accuracy of individual Trust-level analysis since the magnitude of the measured indicator changes over time. To further understand the relationships between indicators and admissions, the signals should be modelled directly and compared; however, modelling was considered beyond the scope of the study and will be explored in future work. In addition, further work should explore factors such as Trust catchment size, acuteness, specialisations, and demographics, which impact the strength of temporal and spatial relationship.

Work in this area should focus on how to forecast using spatially and temporally granular data streams for potentially large numbers of indicators. Further utilisation of time series methods would help support the understanding of indicator–outcome relationships and how to best harness them. Substantial amounts of work has looked at leading indicators of healthcare pressures at national or regional levels, but understanding how to best use indicators at a finer spatial scale increases the value of these indicators to inform public health systems.

Conclusion

This work shows that novel surveillance data sources can be used to reliably understand the expected hospital burden from SARS-CoV-2 transmission. Modelling at a hospital Trust level requires an understanding of the variation across space between the indicators and admissions, as well as the drift of the indicator–admissions temporal leads. We show that clinical, crowdsourced, and healthcare seeking behaviour data at a fine spatial level have positive correlations at lead times, but these relationships are not stable between epidemic waves or constant across England. The different variables selected for analysis were chosen to be as disaggregated as possible while maintaining sufficient counts, which allowed us to see indicators at low geographies with strong relationships, rather than aggregated signals at higher geographies. This approach was shown to be valuable in understanding the heterogeneity in fine spatial scales. While some data sources and indicators perform better overall, our analysis did not identify an indicator that had no substantial variability. Therefore, in practical public health operations, spatial and temporal heterogeneity should be considered in a modelling context, avoiding a reliance on single signals, which can diverge or degrade from the quantity that is being forecast.

Supplementary material

The supplementary material for this article can be found at http://doi.org/10.1017/S0950268823001449.

Data availability statement

The UKHSA operates a robust governance process for applying to access protected data that considers:

  • the benefits and risks of how the data will be used

  • compliance with policy, regulatory, and ethical obligations

  • data minimisation

  • how the confidentiality, integrity, and availability will be maintained

  • retention, archival, and disposal requirements

  • best practice for protecting data, including the application of ‘privacy by design and by default’, emerging privacy conserving technologies and contractual controls

Access to protected data is always strictly controlled using legally binding data sharing contracts.

The UKHSA welcomes data applications from organisations looking to use protected data for public health purposes.

To request an application pack or discuss a request for UKHSA data you would like to submit, contact .

Author contribution

Investigation: C.O., J.B., J.M., L.C., M.F., T.B., T.W.; Methodology: C.O., J.M., T.W.; Writing – review & editing: C.O., J.M., T.W.; Formal analysis: J.B., J.M., L.C., M.F., T.B.; Conceptualization: J.M., T.W.; Data curation: J.M.; Project administration: J.M.; Validation: J.M., T.W.; Visualization: J.M., T.B.; Writing – original draft: J.M., T.B., T.W.; Supervision: T.W.

Financial support

The authors were employed by the UKHSA but received no specific funding for this study.

Competing interest

The authors declare none.

References

Meakin, S, Abbott, S, Bosse, N, Munday, J, Gruson, H, Hellewell, J, Sherratt, K, CMMID COVID-19 Working Group, Chapman, LAC, Prem, K, Klepac, P, Jombart, T, Knight, GM, Jafari, Y, Flasche, S, Waites, W, Jit, M, Eggo, RM, Villabona-Arenas, CJ, Russell, TW, Medley, G, Edmunds, WJ, Davies, NG, Liu, Y, Hué, S, Brady, O, Pung, R, Abbas, K, Gimma, A, Mee, P, Endo, A, Clifford, S, Sun, FY, McCarthy, CV, Quilty, BJ, Rosello, A, Sandmann, G, Barnard, RC, Kucharski, AJ, Procter, SR, Jarvis, CI, Gibbs, HP, Hodgson, D, Lowe, R, Atkins, KE, Koltai, M, Pearson, CAB, Finch, E, Wong, KLM, Quaife, M, O’Reilly, K, Tully, DC and Funk, S (2022) Comparative assessment of methods for short-term forecasts of COVID-19 hospital admissions in England at the local level. BMC Medicine 20(1), 115.CrossRefGoogle ScholarPubMed
Feroze, N (2020) Forecasting the patterns of COVID-19 and causal impacts of lockdown in top five affected countries using Bayesian structural time series models. Chaos, Solitons & Fractals 140, 110196.CrossRefGoogle ScholarPubMed
Friston, KJ, Flandin, G and Razi, A (2020) Dynamic causal modelling of COVID-19. Nature Scientific Reports 12, 12419.CrossRefGoogle Scholar
Birrell, P, Blake, J, van Leeuwen, E, Gent, N and de Angelis, D (2021) Real-time nowcasting and forecasting of COVID-19 dynamics in England: The first wave. Philosophical Transactions of the Royal Society B 376(1829), 20200279.CrossRefGoogle ScholarPubMed
Lampos, V, Majumder, MS, Yom-Tov, E, Edelstein, M, Moura, S, Hamada, Y, Rangaka, MX, McKendry, RA and Cox, IJ (2021) Tracking COVID-19 using online search. NPJ Digital Medicine 4(1), 111.CrossRefGoogle ScholarPubMed
Ward, T, Johnsen, A, Ng, S and Chollet, F (2022) Forecasting SARS-CoV-2 transmission and clinical risk at small spatial scales by the application of machine learning architectures to syndromic surveillance data. Nature Machine Intelligence 4, 814827.CrossRefGoogle Scholar
McDonald, DJ, Bien, J, Green, A, Hu, AJ, DeFries, N, Hyun, S, Oliveira, NL, Sharpnack, J, Tang, J, Tibshirani, R, Ventura, V, Wasserman, L and Tibshirani, RJ (2021) Can auxiliary indicators improve COVID-19 forecasting and hotspot prediction? Proceedings of the National Academy of Sciences 118(51), e2111453118.CrossRefGoogle ScholarPubMed
Ma, S and Yang, S (2022) Covid-19 forecasts using internet search information in the United States. Nature Scientific Reports 12(1), 116.Google ScholarPubMed
Gerlee, P, Karlsson, J, Fritzell, I, Brezicka, T, Spreco, A, Timpka, T, Jöud, A and Lundh, T (2021) Predicting regional COVID-19 hospital admissions in Sweden using mobility data. Nature Scientific Reports 11(1), 18.Google ScholarPubMed
Fox, SJ, Lachmann, M, Tec, M, Pasco, R, Woody, S, du, Z, Wang, X, Ingle, TA, Javan, E, Dahan, M, Gaither, K, Escott, ME, Adler, SI, Johnston, SC, Scott, JG and Meyers, LA (2022) Real-time pandemic surveillance using hospital admissions and mobility data. Proceedings of the National Academy of Sciences 119(7), e2111870119.CrossRefGoogle ScholarPubMed
Kogan, NE, Clemente, L, Liautaud, P, Kaashoek, J, Link, NB, Nguyen, AT, Lu, FS, Huybers, P, Resch, B, Havas, C, Petutschnig, A, Davis, J, Chinazzi, M, Mustafa, B, Hanage, WP, Vespignani, A and Santillana, M (2021) An early warning approach to monitor COVID-19 activity with multiple digital traces in near real time. Science Advances 7(10), eabd6989.CrossRefGoogle ScholarPubMed
Conroy, B, Silva, I, Mehraei, G, Damiano, R, Gross, B, Salvati, E, Feng, T, Schneider, J, Olson, N, Rizzo, AG, Curtin, CM, Frassica, J and McFarlane, DC (2022) Real-time infection prediction with wearable physiological monitoring and AI to aid military workforce readiness during COVID-19. Nature Scientific Reports 12(1), 112.Google ScholarPubMed
Ginsberg, J, Mohebbi, MH, Patel, RS, Brammer, L, Smolinski, MS and Brilliant, L (2009) Detecting influenza epidemics using search engine query data. Nature 457(7232), 10121014.CrossRefGoogle ScholarPubMed
McGough, SF, Brownstein, JS, Hawkins, JB and Santillana, M (2017) Forecasting Zika incidence in the 2016 Latin America outbreak combining traditional disease surveillance with search, social media, and news report data. PLoS Neglected Tropical Diseases 11(1), e0005295.CrossRefGoogle ScholarPubMed
Santillana, M, Nsoesie, EO, Mekaru, SR, Scales, D and Brownstein, JS (2014) Using clinicians’ search query data to monitor influenza epidemics. Clinical Infectious Diseases 59(10), 14461450.CrossRefGoogle ScholarPubMed
Smolinski, MS, Crawley, AW, Baltrusaitis, K, Chunara, R, Olsen, JM, Wójcik, O, Santillana, M, Nguyen, A and Brownstein, JS (2015) Flu near you: Crowdsourced symptom reporting spanning 2 influenza seasons. American Journal of Public Health 105(10), 21242130.CrossRefGoogle ScholarPubMed
Lampos, V and Cristianini, N (2010) Tracking the flu pandemic by monitoring the social web. In 2010 2nd International Workshop on Cognitive Information Processing, Elba, Italy. https://ieeexplore.ieee.org/document/5604088.Google Scholar
Jassat, W, Abdool Karim, SS, Mudara, C, Welch, R, Ozougwu, L, Groome, MJ, Govender, N, von Gottberg, A, Wolter, N, Wolmarans, M, Rousseau, P, DATCOV author group, Blumberg, L and Cohen, C (2022) Clinical severity of COVID-19 in patients admitted to hospital during the omicron wave in South Africa: A retrospective observational study. The Lancet Global Health 10(7), E961E969.CrossRefGoogle ScholarPubMed
Wrenn, JO, Pakala, SB, Vestal, G, Shilts, MH, Brown, HM, Bowen, SM, Strickland, BA, Williams, T, Mallal, SA, Jones, ID, Schmitz, JE, Self, WH and das, SR (2022) COVID-19 severity from omicron and Delta SARS-CoV-2 variants. Influenza and Other Respiratory Viruses 16(5), 832836.CrossRefGoogle ScholarPubMed
Grint, DJ, Wing, K, Houlihan, C, Gibbs, HP, Evans, SJW, Williamson, E, McDonald, HI, Bhaskaran, K, Evans, D, Walker, AJ, Hickman, G, Nightingale, E, Schultze, A, Rentsch, CT, Bates, C, Cockburn, J, Curtis, HJ, Morton, CE, Bacon, S, Davy, S, Wong, AYS, Mehrkar, A, Tomlinson, L, Douglas, IJ, Mathur, R, MacKenna, B, Ingelsby, P, Croker, R, Parry, J, Hester, F, Harper, S, DeVito, NJ, Hulme, W, Tazare, J, Smeeth, L, Goldacre, B and Eggo, RM (2022) Severity of severe acute respiratory system coronavirus 2 (SARS-CoV-2) alpha variant (B.1.1.7) in England. Clinical Infectious Diseases 75(1), e1120e1127.CrossRefGoogle ScholarPubMed
Curtis, HJ, Inglesby, P, Morton, CE, MacKenna, B, Green, A, Hulme, W, Walker, AJ, Morley, J, Mehrkar, A, Bacon, S, Hickman, G, Bates, C, Croker, R, Evans, D, Ward, T, Cockburn, J, Davy, S, Bhaskaran, K, Schultze, A, Rentsch, CT, Williamson, EJ, Rowan, A, Fisher, L, McDonald, HI, Tomlinson, L, Mathur, R, Drysdale, H, Eggo, RM, Wing, K, Wong, AY, Forbes, H, Parry, J, Hester, F, Harper, S, O’Hanlon, S, Eavis, A, Jarvis, R, Avramov, D, Griffiths, P, Fowles, A, Parkes, N, Douglas, IJ, Evans, SJ, Smeeth, L, Goldacre, B and The OpenSAFELY Collaborative (2021) Trends, regional variation, and clinical characteristics of COVID-19 vaccine recipients: A retrospective cohort study in 23.4 million patients using OpenSAFELY. British Journal of General Practice 72(714), e51e62.CrossRefGoogle Scholar
UK Government (2022) Covid-19 Pandemic in the United Kingdom. GOV.UK Coronavirus (COVID-19) in the UK - Vaccinations in the UK. Available at https://coronavirus.data.gov.uk/details/vaccinations (accessed 1 September 2022).Google Scholar
Dan, JM, Mateus, J, Kato, Y, Hastie, KM, Yu, ED, Faliti, CE, Grifoni, A, Ramirez, SI, Haupt, S, Frazier, A, Nakao, C, Rayaprolu, V, Rawlings, SA, Peters, B, Krammer, F, Simon, V, Saphire, EO, Smith, DM, Weiskopf, D, Sette, A and Crotty, S (2021) Immunological memory to SARS-CoV-2 assessed for up to 8 months after infection. Science 371(6529), eabf4063.CrossRefGoogle ScholarPubMed
Khoury, J, Najjar-Debbiny, R, Elemy, A, Jabbour, A, Haj, J, Abu-Sini, M, Yasin, R, Amin, M, Hellou, E, Nasrallah, N, Saffouri, A and Hakim, F (2022) Immunity waning after COVID vaccine booster vs. infection—Better than expected. Infectious Diseases 54(11), 828831.CrossRefGoogle ScholarPubMed
National Health Service. NHS Trust. NHS Data Model and Dictionary. Available at https://www.datadictionary.nhs.uk/nhs_business_definitions/nhs_trust.html?hl=nhs%2Ctrust (accessed 10 July 2023).Google Scholar
NHS England. COVID-19 Hospital Activity. NHS England – Statistical Work Areas. Available at https://www.england.nhs.uk/statistics/statistical-work-areas/covid-19-hospital-activity/ (accessed 8 September 2022).Google Scholar
Google. FAQ about Google Trends Data – Trends Help. Trends Help Available at https://support.google.com/trends/answer/4365533 (accessed 15 September 2022).Google Scholar
NHS Digital. NHS Pathways. NHS Digital Services Available at https://digital.nhs.uk/services/nhs-pathways (accessed 8 September 2022).Google Scholar
NHS England. NHS England. Available at https://www.gov.uk/government/organisations/nhs-england (accessed 8 September 2022).Google Scholar
UK Government. Testing in England. GOV.UK Coronavirus (COVID-19) in the UK. Available at https://coronavirus.data.gov.uk/details/testing?areaType=nation&areaName=England (accessed 1 September 2022).Google Scholar
UK Health Security Agency. NHS Covid-19 App. GOV.UK Available at https://www.gov.uk/government/collections/nhs-covid-19-app (accessed 13 June 2022).Google Scholar
Menni, C, Valdes, AM, Freidin, MB, Sudre, CH, Nguyen, LH, Drew, DA, Ganesh, S, Varsavsky, T, Cardoso, MJ, el-Sayed Moustafa, JS, Visconti, A, Hysi, P, Bowyer, RCE, Mangino, M, Falchi, M, Wolf, J, Ourselin, S, Chan, AT, Steves, CJ and Spector, TD (2020) Real-time tracking of self-reported symptoms to predict potential COVID-19. Nature Medicine 26(7), 10371040.CrossRefGoogle ScholarPubMed
Meakin, S, Abbott, S and Funk, S (2020) NHS Trust level Covid-19 data aggregated to a range of spatial scales. https://doi.org/10.5281/zenodo.4447465CrossRefGoogle Scholar
Granger, CWJ (1969) Investigating causal relations by econometric models and cross-spectral methods. Econometrica: Journal of the Econometric Society 37, 424438.CrossRefGoogle Scholar
Vio, R and Wamsteker, W (2001) Limits of the cross-correlation function in the analysis of short time series. Publications of the Astronomical Society of the Pacific 113(779), 86.CrossRefGoogle Scholar
Giorgino, T (2009) Computing and visualizing dynamic time warping alignments in R: The dtw package. Journal of Statistical Software 31, 124.CrossRefGoogle Scholar
Senin, P (2008) Dynamic time warping algorithm review. Information and Computer Science Department University of Hawaii at Manoa Honolulu, USA 855, 123.Google Scholar
Sakoe, H and Seibi, C (1978) Dynamic programming algorithm optimization for spoken word recognition. IEEE Transactions on Acoustics, Speech, and Signal Processing 26(1), 4349.CrossRefGoogle Scholar
Tormene, P, Giorgino, T, Quaglini, S and Stefanelli, M (2009) Matching incomplete time series with dynamic time warping: An algorithm and an application to post-stroke rehabilitation. Artificial Intelligence in Medicine 45(1), 1134.CrossRefGoogle Scholar
Wang, T, Ma, S, Baek, S and Yang, S (2022) COVID-19 hospitalizations forecasts using internet search data. Nature Scientific Reports 12, 9661.CrossRefGoogle ScholarPubMed
Wymant, C, Ferretti, L, Tsallis, D, Charalambides, M, Abeler-Dörner, L, Bonsall, D, Hinch, R, Kendall, M, Milsom, L, Ayres, M, Holmes, C, Briers, M and Fraser, C (2021) The epidemiological impact of the NHS COVID-19 app. Nature 594(7863), 408412.CrossRefGoogle ScholarPubMed
UK Health Security Agency. NHS COVID-19 App Statistics. NHS COVID App Support. Available at https://stats.app.covid19.nhs.uk/#contact-tracing-alerts (accessed 18 September 2022).Google Scholar
O’brien, DA and Clements, CF (2021) Early warning signal reliability varies with COVID-19 waves. Biology Letters 17(12), 20210487.CrossRefGoogle ScholarPubMed
Rostami-Tabar, B and Rendon-Sanchez, JF (2021) Forecasting COVID-19 daily cases using phone call data. Applied Soft Computing 100, 106932.CrossRefGoogle ScholarPubMed
Marinazzo, D, Pellicoro, M and Stramaglia, S (2008) Kernel-Granger causality and the analysis of dynamical networks. Physical Review E, Statistical, Nonlinear, and Soft Matter Physics 77, 056215.CrossRefGoogle ScholarPubMed
Figure 0

Figure 1. A time series plot of example variables from each data source’s indicators, aggregated nationally, with reference to national admissions (red dashed line). Indicators that lead admissions well should appear shifted leftward of the admissions line. Indicators and admissions are scaled between 0 and 1 to allow for easy visual comparison of temporal offsets.

Figure 1

Figure 2. The distribution of p-values across Trusts from Granger causality tests. The tests are performed at the Trust level for each separate wave between indicator and admissions. Low p-values tell us the indicator leads admissions well, given the conditions of the test.

Figure 2

Figure 3. The distribution of p-values across Trusts from Granger causality tests. The tests are performed at the Trust level for each separate wave between indicators and admissions in 14 days. Low p-values tell us the indicator leads admissions in 14 days well given the conditions of the test.

Figure 3

Figure 4. The distribution, via boxplot, of optimal lead times between the indicator and admissions across all Trusts for each wave. Optimal lead is defined as the lead time with the maximum non-negative CCF value within a 30-day forward and backward window. Larger optimal leads will be most useful for forecasting, wide ranges in optimal lead correspond to high variation spatially, and smaller variation indicates a more consistent lead. The indicators with higher average (across waves) optimal leads are sorted to the left.

Figure 4

Figure 5. The distribution of CCF values for the indicator and admissions at 14 days lead across Trusts for each wave. High CCF values correspond to a high correlation between time series, and CCF values centred around zero would show that an indicator does not have a meaningful temporal lead against admissions. High variation in the CCF values show how consistent leading relationships are across the different Trusts. The indicators with higher average (across wave) CCF values are sorted to the left.

Figure 5

Figure 6. Dynamic time warping mapping between indicators and admissions used to generate lead times. The DTW is shown for a single indicator and Trust. The solid time series represents the variable being evaluated, the indicator, the dashed are admissions, and the lines between are the aligned sequence pairs. Vertical lines indicate no temporal offset between time series.

Figure 6

Figure 7. The distribution of lead times calculated from sequence index matching between indicators and admissions across the different epidemic waves of study. The lead times correspond to the optimal time warping between indicator and admissions, with a higher value indicating a larger temporal lead.

Figure 7

Figure 8. The normalised path distance produced by warping the space between the indicator and admissions across all three waves. Produced using multivariate DTW across Trusts. The normalised path distance indicates how much total warping is needed between indicators and admissions, a proxy for how big a lead time there is between time series.

Figure 8

Table 1. The operational considerations at time of investigation

Supplementary material: File

Mellor et al. supplementary material
Download undefined(File)
File 802 KB