Difference between revisions of "Gomes"
Line 3: | Line 3: | ||
Note on [https://www.medrxiv.org/content/10.1101/2020.04.27.20081893v2 "Individual variation in susceptibility or exposure to SARS-CoV-2 lowers the herd immunity threshold"] by M. Gabriella M. Gomes et al. | Note on [https://www.medrxiv.org/content/10.1101/2020.04.27.20081893v2 "Individual variation in susceptibility or exposure to SARS-CoV-2 lowers the herd immunity threshold"] by M. Gabriella M. Gomes et al. | ||
− | = | + | This paper made a media splash early in May 2020 with |
+ | [https://www.spectator.co.uk/article/herd-immunity-may-only-need-a-10-per-cent-infection-rate headlines] such as "Herd immunity may only need 10-20 per cent of people to be infected". The proposed mechanism is that people who are more susceptible to the infection are those more involved in spreading it, but will also become immune at a faster rate than others, so the disease naturally tends to make the key players immune more so than others. | ||
+ | |||
+ | = Discussion = | ||
+ | |||
+ | * This paper uses two variants of a standard SEIR model. The first is the "susceptibility model", where the population is divided into classes according to a parameter that governs to what extent they are likely to become infected. Passing on the infection is still assumed to occur uniformly. The second is the "connectivity model" which takes this further by assuming that infectivity is proportional to susceptibility. I suppose the latter model is called "connectivity" because it is equivalent to changing the connectivity of the underlying graph - making it random subject to a degree distribution. | ||
+ | |||
+ | * Varying the infectivity alone would just reduce to a standard SEIR model because everyone (all infectivity classes) would be being infected at the same rate. By contrast, in the two above models the susceptibility classes are being depleted at different rates from each other. | ||
+ | |||
+ | * As the authors conclude, there is an effect where the herd immunity threshold (HIT) in the two models above is lower than that predicted by simple homogeneous modelling ($$1-1/R_0$$). This is a valuable point to make given that many people seem to be taking as read that the herd immunity threshold must be 60-70%, and we obviously very much need to know when herd immunity arises, or at least to what extent we should expect to feel the effects of partial herd immunity. | ||
− | + | * However, I doubt that a HIT as low as 10-20% is at all likely. It arises in this paper's model from an extreme distribution of susceptibilities where there is a big concentration at the low end. I think the evidence that is presented in the paper for this particular distribution (Gamma) of susceptibilities comes from evidence about the distribution of infectivity, which is a different thing. (More on this below.) | |
− | |||
− | + | * Contrary to the practice in the paper, I believe the Coefficient of Variation (CV), or other dispersion parameter that is basically a function of the mean and variance, is not, for these purposes, a suitable parameter to use to measure how much the susceptibility distribution varies from a point value (constant). To demonstrate this we can choose different distributions with the same CV and get a huge range of HITs. | |
− | |||
− | |||
− | |||
− | * Contrary to the practice in the paper, I believe the Coefficient of Variation (CV) is not a suitable parameter to use to measure how much the susceptibility distribution varies from a point value (constant) | ||
− | |||
− | + | * There is actually a nice clean way to evaluate the HIT directly from the susceptibility distribution. | |
− | + | * This paper shows what happens if you "artificially" add in a variation in susceptibility: treat everyone uniformly except for a susceptibility parameter controlled by a given distribution. An alternative approach would be to use a real life contact graph, or some version of it. Age- and location-based contact information have been studied for this purpose, e.g., [https://journals.plos.org/ploscompbiol/article?id=10.1371/journal.pcbi.1005697 Prem et al] and [https://www.sciencedirect.com/science/article/pii/S1755436518300306 Klepac et al]. Such a calculation is carried out in [https://arxiv.org/abs/2005.03085 Bitton et al] using Age- and Activity-based mixing matrices. They find, inter alia, that if $$R_0=3$$ then the HIT is reduced from $$66.7%$$ to $$49.1%$$. This empirically-derived distribution has the merit of being justifiable, but, speculating, I wonder if it understates the full heterogeneity, in which case there might be an argument for trying a hybrid method where you artificially introduce a little more variability, maybe in the manner of the present paper, on top of the empirically-found mixing matrices. | |
− | + | = In more detail = | |
− | + | The very low HIT estimates in this paper arise from using a Gamma distribution with $$CV=3$$ for the susceptibility distribution, which corresponds to shape parameter $$k=1/9$$. This has a large chunk of its probability at the very low end: it is saying that 63% of the population has susceptibility less than 0.09 (relative to a mean of 1) and 50% has susceptibility less than 0.01. With this in mind, it's not surprising that we end up with low HIT estimates because, roughly speaking, in this case you only need to induce herd immunity amongst the minority susceptible population. | |
− | + | Note that a similar-sounding argument to this is not correct: if the infectivities, but not susceptibilities, were clustered near 0 then it would still be true that you only need to induce herd immunity amongst the minority infectious population, but that wouldn't occur naturally because the infectious portion of the population wouldn't be getting infected any more than the non-infectious portion. | |
− | + | There is some evidence that infectivity is strongly clustered like this (and has a long tail - i.e., superspreaders). The present paper cites [https://wellcomeopenresearch.org/articles/5-67 Endo et al] that suggests a dispersion of something like $$k=1/10$$ in SARS-CoV-2. This relies (slightly optimistically in my opinion) on knowledge of the seed infections in different countries to derive its result, but perhaps more importantly this cited paper is making a statement about infectivity not susceptibility. Also cited is the classic 2005 paper [https://www.nature.com/articles/nature04153 Lloyd-Smith et al] on superspreading for the original SARS outbreak. This estimates a dispersion parameter of $$k=0.16$$ and also gives evidence that the Gamma family is a good one to use (because a Poisson with parameter Gamma is a Negative Binomial, for which they find evidence), but again this is talking about infectivity not susceptibility so isn't directly applicable to the situation of the present paper. | |
− | + | Despite the evidence being shakey, is it yet possible that the real susceptibility distribution looks like a Gamma with shape $$1/9$$? I suspect this is unlikely. As far as I am aware anyone can catch the disease, and there aren't any known huge differences in susceptibility amongst subpopulations. The best candidate might be age, since the disease is so strongly age-dependent in terms of severity, but while young people may be a bit less susceptible to catching it, it's clearly not the case that the younger 50% of the population (median age in the UK is 40.5 years) are less than $$1/100$$ as susceptible compared with the average. | |
− | To | + | To illustrate how the HIT doesn't properly depend on $$CV$$, I tried using a "two-point" distribution parameterized by $$x, y$$ and $$p$$, where $$P(X=x)=p$$, and $$P(X=y)=1-p$$. Fixing the mean to be 1 and the variance to be $$CV^2$$ leaves a free parameter that may as well be $$x$$. $$x=0.99$$ corresponds to some very rare extreme-superspreaders, while $$x=0$$ corresponds to a lot of (for want of a better term) "superhermits" - i.e., like Gamma at $$CV=3$$, lots of the distribution is concentrated at or near 0. Of course these are unrealistic for real-world distributions. They are just there to make a point. |
− | To | + | (To help my understanding, and to check the paper, I [https://github.com/alex1770/Covid-19/blob/master/variation.py reimplemented their models in Python]. This gives a good match to the output of version 1 of their paper in "susceptibility" mode, though not in "connectivity" mode. Possibly the discrepancy in the latter mode is due to a difference in our initial conditions. In any case there is only a discrepancy in the progress of the infection, not in the HITs, which are anyway separately calculable - see below.) |
− | Version 2 of the paper came out after I conducted | + | Version 2 of the paper came out after I conducted this test, and in it the authors also try out a different family of distributions (lognormal) to test robustness and dependence of their results on a particular family (Gamma). Using lognormal they do in fact get much bigger answers for the HIT than they do for Gamma in the $$CV=3$$ case, but as far as I can see they do not mention this point in the main body of their paper, or discuss how it may be a problem for their use of $$CV$$. It's also relevant that the lognormal family doesn't have a free parameter to vary like the two-point distribution has, so you can't make them particularly extreme like you can with the two-point family. |
− | + | As an illustration, the HIT values for the four distributions, in "susceptibility" mode at $$R_0=2.7$$ (to coincide with what was used in v1 of the paper), all with $$CV=3$$, are: Two-point at x=0: 6.3%, Gamma: 9.5%, Lognormal: 20.4%, Two-point at x=0.99: 62.6%. We see there is a large variation in HIT for the same $$CV$$, so it doesn't make sense to think of HIT as a function of $$CV$$. | |
= Formulae = | = Formulae = |
Revision as of 13:38, 21 May 2020
Article
Note on "Individual variation in susceptibility or exposure to SARS-CoV-2 lowers the herd immunity threshold" by M. Gabriella M. Gomes et al.
This paper made a media splash early in May 2020 with headlines such as "Herd immunity may only need 10-20 per cent of people to be infected". The proposed mechanism is that people who are more susceptible to the infection are those more involved in spreading it, but will also become immune at a faster rate than others, so the disease naturally tends to make the key players immune more so than others.
Discussion
- This paper uses two variants of a standard SEIR model. The first is the "susceptibility model", where the population is divided into classes according to a parameter that governs to what extent they are likely to become infected. Passing on the infection is still assumed to occur uniformly. The second is the "connectivity model" which takes this further by assuming that infectivity is proportional to susceptibility. I suppose the latter model is called "connectivity" because it is equivalent to changing the connectivity of the underlying graph - making it random subject to a degree distribution.
- Varying the infectivity alone would just reduce to a standard SEIR model because everyone (all infectivity classes) would be being infected at the same rate. By contrast, in the two above models the susceptibility classes are being depleted at different rates from each other.
- As the authors conclude, there is an effect where the herd immunity threshold (HIT) in the two models above is lower than that predicted by simple homogeneous modelling ($$1-1/R_0$$). This is a valuable point to make given that many people seem to be taking as read that the herd immunity threshold must be 60-70%, and we obviously very much need to know when herd immunity arises, or at least to what extent we should expect to feel the effects of partial herd immunity.
- However, I doubt that a HIT as low as 10-20% is at all likely. It arises in this paper's model from an extreme distribution of susceptibilities where there is a big concentration at the low end. I think the evidence that is presented in the paper for this particular distribution (Gamma) of susceptibilities comes from evidence about the distribution of infectivity, which is a different thing. (More on this below.)
- Contrary to the practice in the paper, I believe the Coefficient of Variation (CV), or other dispersion parameter that is basically a function of the mean and variance, is not, for these purposes, a suitable parameter to use to measure how much the susceptibility distribution varies from a point value (constant). To demonstrate this we can choose different distributions with the same CV and get a huge range of HITs.
- There is actually a nice clean way to evaluate the HIT directly from the susceptibility distribution.
- This paper shows what happens if you "artificially" add in a variation in susceptibility: treat everyone uniformly except for a susceptibility parameter controlled by a given distribution. An alternative approach would be to use a real life contact graph, or some version of it. Age- and location-based contact information have been studied for this purpose, e.g., Prem et al and Klepac et al. Such a calculation is carried out in Bitton et al using Age- and Activity-based mixing matrices. They find, inter alia, that if $$R_0=3$$ then the HIT is reduced from $$66.7%$$ to $$49.1%$$. This empirically-derived distribution has the merit of being justifiable, but, speculating, I wonder if it understates the full heterogeneity, in which case there might be an argument for trying a hybrid method where you artificially introduce a little more variability, maybe in the manner of the present paper, on top of the empirically-found mixing matrices.
In more detail
The very low HIT estimates in this paper arise from using a Gamma distribution with $$CV=3$$ for the susceptibility distribution, which corresponds to shape parameter $$k=1/9$$. This has a large chunk of its probability at the very low end: it is saying that 63% of the population has susceptibility less than 0.09 (relative to a mean of 1) and 50% has susceptibility less than 0.01. With this in mind, it's not surprising that we end up with low HIT estimates because, roughly speaking, in this case you only need to induce herd immunity amongst the minority susceptible population.
Note that a similar-sounding argument to this is not correct: if the infectivities, but not susceptibilities, were clustered near 0 then it would still be true that you only need to induce herd immunity amongst the minority infectious population, but that wouldn't occur naturally because the infectious portion of the population wouldn't be getting infected any more than the non-infectious portion.
There is some evidence that infectivity is strongly clustered like this (and has a long tail - i.e., superspreaders). The present paper cites Endo et al that suggests a dispersion of something like $$k=1/10$$ in SARS-CoV-2. This relies (slightly optimistically in my opinion) on knowledge of the seed infections in different countries to derive its result, but perhaps more importantly this cited paper is making a statement about infectivity not susceptibility. Also cited is the classic 2005 paper Lloyd-Smith et al on superspreading for the original SARS outbreak. This estimates a dispersion parameter of $$k=0.16$$ and also gives evidence that the Gamma family is a good one to use (because a Poisson with parameter Gamma is a Negative Binomial, for which they find evidence), but again this is talking about infectivity not susceptibility so isn't directly applicable to the situation of the present paper.
Despite the evidence being shakey, is it yet possible that the real susceptibility distribution looks like a Gamma with shape $$1/9$$? I suspect this is unlikely. As far as I am aware anyone can catch the disease, and there aren't any known huge differences in susceptibility amongst subpopulations. The best candidate might be age, since the disease is so strongly age-dependent in terms of severity, but while young people may be a bit less susceptible to catching it, it's clearly not the case that the younger 50% of the population (median age in the UK is 40.5 years) are less than $$1/100$$ as susceptible compared with the average.
To illustrate how the HIT doesn't properly depend on $$CV$$, I tried using a "two-point" distribution parameterized by $$x, y$$ and $$p$$, where $$P(X=x)=p$$, and $$P(X=y)=1-p$$. Fixing the mean to be 1 and the variance to be $$CV^2$$ leaves a free parameter that may as well be $$x$$. $$x=0.99$$ corresponds to some very rare extreme-superspreaders, while $$x=0$$ corresponds to a lot of (for want of a better term) "superhermits" - i.e., like Gamma at $$CV=3$$, lots of the distribution is concentrated at or near 0. Of course these are unrealistic for real-world distributions. They are just there to make a point.
(To help my understanding, and to check the paper, I reimplemented their models in Python. This gives a good match to the output of version 1 of their paper in "susceptibility" mode, though not in "connectivity" mode. Possibly the discrepancy in the latter mode is due to a difference in our initial conditions. In any case there is only a discrepancy in the progress of the infection, not in the HITs, which are anyway separately calculable - see below.)
Version 2 of the paper came out after I conducted this test, and in it the authors also try out a different family of distributions (lognormal) to test robustness and dependence of their results on a particular family (Gamma). Using lognormal they do in fact get much bigger answers for the HIT than they do for Gamma in the $$CV=3$$ case, but as far as I can see they do not mention this point in the main body of their paper, or discuss how it may be a problem for their use of $$CV$$. It's also relevant that the lognormal family doesn't have a free parameter to vary like the two-point distribution has, so you can't make them particularly extreme like you can with the two-point family.
As an illustration, the HIT values for the four distributions, in "susceptibility" mode at $$R_0=2.7$$ (to coincide with what was used in v1 of the paper), all with $$CV=3$$, are: Two-point at x=0: 6.3%, Gamma: 9.5%, Lognormal: 20.4%, Two-point at x=0.99: 62.6%. We see there is a large variation in HIT for the same $$CV$$, so it doesn't make sense to think of HIT as a function of $$CV$$.
Formulae
As it happens, there is a nice clean procedure to directly calculate the HIT under the authors' model, so there is no need for simulation, and there is no dependency of this on the particular characteristics such as the timing of the switching on and off of social distancing.
For the Gamma family there happens to be a closed-form formula. Using the "susceptibility" model it's
\[\text{HIT} = 1-R_0^{-(1+CV^2)^{-1}},\] and with the "connectivity" model it's
\[\text{HIT} = 1-R_0^{-(1+2CV^2)^{-1}}.\]
In general the procedure for any distribution of initial susceptibilities/connectivities, written as the random variable $$X$$ with $$E[X]=1$$, is as follows:
- Susceptibility: choose $$\Lambda\ge0$$ such that $$R_0 E[Xe^{-\Lambda X}]=1$$, then $$\text{HIT}=1-E[e^{-\Lambda X}]$$.
- Connectivity: choose $$\Lambda\ge0$$ such that $$R_0 E[X^2e^{-\Lambda X}]=E[X^2]$$, then $$\text{HIT}=1-E[e^{-\Lambda X}]$$.
(If you need a susceptibility distribution with $$E[X]\neq1$$, and also want to keep the notation of this paper, and want $$R_0$$ to retain its conventional meaning of the initial branching factor, then you need to absorb the factor of $$E[X]$$ into $$R_0$$ and rescale $$X$$ to have mean 1.)
We may now compare the solid curves from the paper of fig.3 (HIT as a function of $$CV$$ assuming Gamma distribution), and those of fig. S22 (HIT as a function of $$CV$$ assuming lognormal distribution), with those given by the above procedure. We can also add in the two-point $$x=0$$ and two-point $$x=0.99$$ distributions for comparison.