Skip to content
Publicly Available Published by De Gruyter November 22, 2016

Estimating intrinsic and extrinsic noise from single-cell gene expression measurements

  • Audrey Qiuyan Fu EMAIL logo and Lior Pachter EMAIL logo

Abstract

Gene expression is stochastic and displays variation (“noise”) both within and between cells. Intracellular (intrinsic) variance can be distinguished from extracellular (extrinsic) variance by applying the law of total variance to data from two-reporter assays that probe expression of identically regulated gene pairs in single cells. We examine established formulas [Elowitz, M. B., A. J. Levine, E. D. Siggia and P. S. Swain (2002): “Stochastic gene expression in a single cell,” Science, 297, 1183–1186.] for the estimation of intrinsic and extrinsic noise and provide interpretations of them in terms of a hierarchical model. This allows us to derive alternative estimators that minimize bias or mean squared error. We provide a geometric interpretation of these results that clarifies the interpretation in [Elowitz, M. B., A. J. Levine, E. D. Siggia and P. S. Swain (2002): “Stochastic gene expression in a single cell,” Science, 297, 1183–1186.]. We also demonstrate through simulation and re-analysis of published data that the distribution assumptions underlying the hierarchical model have to be satisfied for the estimators to produce sensible results, which highlights the importance of normalization.

1 Introduction

A gene can have different expression levels in living cells that have the same genetic material and are subject to the same environment (Stegle et al., 2015). During early development of an organism, distinct expression profiles eventually lead to formation of different tissues. Moreover, complex tissues such as brain have many different subtypes of cells with different gene expression profiles. However, variation in expression between cells is reflective not only of distinct biological state, but also of stochasticity underlying many of the processes fundamental to the molecular biology of cell.

In a classic paper on the stochasticity of gene expression in single cells, Elowitz et al. (2002) introduced a clever two-reporter expression assay designed to tease apart “intrinsic” and “extrinsic” variation (also called “noise”) from the overall variability in gene expression: the intrinsic noise is the variation in the expression of the same gene in identical environment, whereas the extrinsic noise is the variation in gene expression due to cellular environment that impacts all the genes at once. The idea is as follows: two identically regulated reporter genes (cyan fluorescent protein and yellow fluorescent protein) are inserted into individual E. coli. cells, allowing for comparable expression measurements within and between cells. If n cells are assayed, this leads to expression measurements c1, … cn and y1, … yn, where the pair (ci, yi) represent the expression measurements for the cyan and yellow reporters in the ith cell. The goal of the experiment is to measure the variance in gene expression from the pairs (ci, yi) (denoted by ηtot2) and to ascribe it to two different sources: first, variability due to the different states of cells (“extrinsic noise,” denoted by ηext2), and second, inherent variability that exists even when the state of cells is fixed (“intrinsic noise,” denoted by ηint2). In Elowitz et al. (2002), these noise terms are defined as squared coefficients of variation and specific formulas are provided for estimating ηext2,ηint2 and ηtot2 (hereafter referred to as the ELSS estimates):

(1)ηint2=1n(i=1n12(ciyi)2)c¯y¯,
(2)ηext2=1ni=1nciyic¯y¯c¯y¯,
(3)ηtot2=1ni=1n12(ci2+yi2)c¯y¯c¯y¯,

where c¯=1ni=1nci and y¯=1ni=1nyi.

Hilfinger and Paulsson (2011) later interpreted these estimates in terms of the “law of total variance” (explained in the next section), which sheds light on the statistical basis of the ELSS estimators but does not address questions about their statistical properties. In this paper, we derive the bias and mean squared error of the ELSS estimators and examine their optimality. We also examine the geometric and biological interpretation of the estimators.

The processes that lead to the expression of the reporters (or genes in general) are much more complex than described here, e.g. the models described in the paper ignore the effects of translation. Many studies (e.g. Rausenberger and Kollmann 2008 and Komorowski et al. 2013) have developed detailed mathematical models for these processes. While some of our results may generalize and be relevant in more general settings, we restrict our analysis to the intrinsic and extrinsic noise as examined by Elowitz et al. (2002) and accessible via static reporter expression experiments. Analyses are implemented in the R package noise available on CRAN.

2 A hierarchical model

We begin by introducing a hierarchical model that provides a formal model for the experiments of Elowitz et al. (2002) and that provides insight into the numerators of (1,2,3). They are the key components of the Elowitz et al. (2002) formulas and can be viewed as estimators of true variances. We note that lower case letters such as ci and yi denote observations not only in the ELSS formulas but throughout our paper; we reserve uppercase letters for random variables.

A hierarchical model for expression of the two reporters in a cell emerges naturally from the assumption that reporter expression, conditioned on the same cellular environment, is represented by independent and identically distributed random variables. To allow each cell to be different from the others, we introduce independent identically distributed random variables Zi, for i = 1, …, n that represent the environments of cells [as in Hilfinger and Paulsson (2011)]. Consistent with Elowitz et al. (2002), we posit that the cellular conditional random variables associated to the two reporters have the same distribution F with mean Mi and variance Σi2, both parameters being unique to the i th cell:

(4)Ci|ZiF(Mi,Σi2)

and

(5)Yi|ZiF(Mi,Σi2).

Thinking of a two reporter experiment as “random,” in the sense that the states of cells Z1, … Zn are random, across cells we have

MiG(μ,σμ2)

and

Σi2H(σ2,ϵ),

where G is the distribution of all the Mis, with mean μ and variance σμ2, and H that of all the Σi2s, with mean σ2 and variance ϵ. In other words, both the mean and variance of reporter expression level is cell specific and the random variable Σi2 and its mean σ2 represent the “within-cell” variation as distinguished from the parameter σμ2 which represents the “between-cell” variability in the ANOVA setting.

For any i, the mean of Ci or Yi is μ, according to the following calculation:

(6)E[Ci]=EZi[E[Ci|Zi]]=E[Mi]=μ.

The total variance in Ci (or Yi) can be calculated using the “law of total variance”:

(7)Var[Ci]=EZi[Var[Ci|Zi]]+VarZi[E[Ci|Zi]].

Using the notation of the hierarchical model described above, and dropping the subscripts for expectation because they are clear by context, we have, for any i,

(8)E[Var[Ci|Zi]]=σ2(within-cell variability; intrinsic noise),
(9)Var[E[Ci|Zi]]=σμ2(between-cell variability; extrinsic noise).

With this notation equation (7) becomes

(10)Var[Ci]=E[Var[Ci|Zi]]+Var[E[Ci|Zi]]=σ2+σμ2   (total noise).

This means that the marginal (unconditional) distributions of Ci and Yi are identical:

CiF(μ,σ2+σμ2);YiF(μ,σ2+σμ2),

where the marginal distribution F′ may or may not be the same as the conditional distribution F.

In the next sections, we will derive the estimators for extrinsic and intrinsic noise, and examine the bias and MSE of each estimator. Specifically, for any estimator S, the MSE of S with respect to the true parameter τ is calculated as follows:

E[(Sτ)2]=E[SE[S]+E[S]τ]2=E[(SE[S])2+(E[S]τ)2+2(SE[S])(E[S]τ)]=E[SE[S]]2+E[E[S]τ]2=Var[S]+(E[S]τ)2,

where E[S] − τ is the bias of S.

3 Extrinsic noise

To examine estimators for extrinsic noise, we start with the law of total variance, noting that the within-cell variability Var[E[Ci|Zi]] can be written as:

(11)Var[E[Ci|Zi]]=E[E[Ci|Zi]2](E[E[Ci]Zi])2=E[E[Ci|Zi]E[Yi|Zi]](E[E[Ci]Zi])2=E[E[CiYi|Zi]]E[E[Ci|Zi]E[E[Yi|Zi]]=E[CiYi]E[Ci]E[Yi]=Cov[Ci,Yi].

This connection between the extrinsic noise, the law of total variance and the covariance of Ci and Yi was noted in Hilfinger and Paulsson (2011).

Formula (11) leads to the following unbiased estimator for the extrinsic noise, as it is an unbiased estimator estimator for the covariance:

Sext*=1n1(i=1nCiYinC¯Y¯).

We note that the ELSS estimator (2) uses the scalar 1/n, which unlike the case of the intrinsic noise estimator (1) leads to a biased estimator in this case.

In order to find the estimator that minimizes the MSE, we consider the following general estimator:

Sext=1a(i=1nCiYinC¯Y¯).

We assume that Mi is normal and that μ = 0 and ϵ = 0. The MSE of Sext is

E[Sextσμ2]2=n1a2(σ2+σμ2)2+(n1)2na2σμ4+(n1aσμ2σμ2)2=(n1)(σ2+σμ2)21a2+(n1)2(1+1n)σμ41a22(n1)σμ41a+σμ4=((n1)(σ2+σμ2)2+(n1)2(1+1n)σμ4)1a22(n1)σμ41a+σμ4,

which is minimized when

1a=σμ4(σ2+σμ2)2+(n1)(1+1n)σμ4, or equivalently
(12)a=(n1)(1+1n)+(σ2+σμ2σμ2)2=(n1)(1+1n)+1ρ2.

The last step in (12) is due to Equations (9), (10) and (11):

(13)σμ2σ2+σμ2=Cov[Ci,Yi]Var[Ci]=Cov[Ci,Yi]Var[Ci]Var[Yi]=ρ.

It is interesting to note that (12) comprises two parts: the first, (n1)(1+1n) converges to n − 1 as n → ∞, while the second, (σ2+σμ2σμ2)2 is equal to 1ρ2 where ρ is the correlation between the two reporter expression vectors C and Y. See Appendices A and B for more details.

4 Intrinsic noise

Also starting with the law of total variance, the within-cell variability E[Var[Ci|Zi]] for cell i can be written as:

(14)E[Var[Ci|Zi]]=Var[Ci]Var[E[Ci|Zi]]=12[Var[Ci]+Var[Yi]]Cov[Ci,Yi]=12[Var[Ci]2Cov[Ci,Yi]+Var[Yi]]=12Var[CiYi]=12(E[CiYi]2(E[CiYi])2).

This leads to the following unbiased estimator for the intrinsic noise:

Sint*=12(n1)i=1n[(CiYi)(C¯Y¯)]2=12(n1)i=1n(CiYi)2n2(n1)(C¯Y¯)2.

To find the estimator that minimizes the MSE, we consider estimators of the following general form

(15)Sint=12a(1n(CiYi)2n(C¯Y¯)2).

Assuming normality of the distribution G (i.e. cell-specific means Mi follow a normal distribution), as well as μ = 0 and ϵ = 0, the MSE is given by

E[Sintσ2]2=Var[Sint]+(E[Sint]σ2)2=12a2[(2n2+6n7)σ4+2(2n1)σ2σμ2+1nσμ4]2(n1)σ41a+σ4.

The value of a that minimizes this expression is

a=(2n37n+6)σ4+2(2n)σ2σμ2+σμ42(n2n)σ4=2n37n+62(n2n)+2nn2nσμ2σ2+12(n2n)(σμ2σ2)2=2n37n+62(n2n)+2nn2nρ1ρ+12(n2n)(ρ1ρ)2.

See Appendices A and C for the complete derivation.

The analysis above can be simplified with an additional assumption, namely that C¯ = Ȳ. In some experiments this may be a natural assumption to make, whereas in others the condition is likely to be violated; we comment on this in more detail in the discussion. Here we proceed to note that assuming that C¯ = Ȳ, the estimator (15) simplifies to

S~int=12ai=1n(CiYi)2.

The unbiased estimator with this form is easily derived by observing that

E[S~int]=12ai=1nE[CiYi]2=12ai=1nVar[CiYi]=n2a(2σ2+2σμ22σμ2)=naσ2.

Thus, in order for S~int to be unbiased the parameter a must be equal to n. The resulting formula is the ELSS formula in (1). This makes clear that the assumption C¯ = Ȳ underlies the derivation of the ELSS intrinsic noise estimator.

In order to study the mean squared error and derive an estimator that minimizes it, we again assume normality of G. The MSE of Sint is then given by

E[S~intσ2]2=Var[S~int]+(E[S~int]σ2)2=na2(3ϵ+2σ4)+(naσ2σ2)2.

Assuming again that μ = 0 and ϵ = 0, the MSE simplifies to

E[S~intσ2]2=2na2σ4+σ4((na)22na+1)=nσ4(n+2)a22nσ4a+σ4,

which is minimized when a = n + 2 (see Appendices A and D for the complete derivation).

5 Geometric interpretation

Figure 3A of Elowitz et al. (2002) shows a scatterplot of data (ci, yi) for an experiment and suggests thinking of intrinsic and extrinsic noise geometrically in terms of projection of the points onto a pair of orthogonal lines. While this geometric interpretation of noise agrees exactly with the ELSS intrinsic noise formula, the interpretation of extrinsic noise is more subtle. Here we complete the picture.

Figure 1 Geometric interpretation of intrinsic and extrinsic noise. The intrinsic noise, or the within-cell variability, is the variance of the points projected to the line y = −c, which is perpendicular to y = c. In other words, it is the average of the squared lengths 12(yi−ci)2$\frac{1}{2}(y_{i}-c_{i})^{2}$. The red point is the projection of point (ci, yi) onto the line y = c. The green point is the centroid (c¯$\bar{c}$, ȳ) (or ((c¯+y¯)/2$(\bar{c}+\bar{y})/\sqrt{2}$, 0) after projection) under the assumption that the two means are equal. See the main text for detail. The extrinsic noise, or the between-cell variability, is the sample covariance between ci and yi. The colored triangles around the blue point ( a randomly selected data point) illustrate the geometric interpretation of the sample covariance: it is the average (signed) area of triangles formed by pairs of data points: green triangles in Q1 and Q3 (some not shown) represent a positive contribution to the covariance, whereas the magenta triangles in Q2 and Q4 a negative contribution. Since most data points lie in the 1st (Q1) and 3rd (Q3) quadrants relative to the blue point, most of the contribution involving the blue point is positive. Similarly, since most pairs of data points can be connected by a positively signed line, their positive contribution will result in a positive covariance. In Elowitz et al. (2002) the direction along the line y = c is labeled extrinsic, which makes sense in terms of the intuition for positive sample covariance. However we have placed that label “extrinsic” in quotes because the extrinsic noise estimator corresponding directly to the sample variance for points projected onto the line y = c (in analogy with intrinsic noise) is heavily biased and not usable in practice.
Figure 1

Geometric interpretation of intrinsic and extrinsic noise. The intrinsic noise, or the within-cell variability, is the variance of the points projected to the line y = −c, which is perpendicular to y = c. In other words, it is the average of the squared lengths 12(yici)2. The red point is the projection of point (ci, yi) onto the line y = c. The green point is the centroid (c¯, ȳ) (or ((c¯+y¯)/2, 0) after projection) under the assumption that the two means are equal. See the main text for detail. The extrinsic noise, or the between-cell variability, is the sample covariance between ci and yi. The colored triangles around the blue point ( a randomly selected data point) illustrate the geometric interpretation of the sample covariance: it is the average (signed) area of triangles formed by pairs of data points: green triangles in Q1 and Q3 (some not shown) represent a positive contribution to the covariance, whereas the magenta triangles in Q2 and Q4 a negative contribution. Since most data points lie in the 1st (Q1) and 3rd (Q3) quadrants relative to the blue point, most of the contribution involving the blue point is positive. Similarly, since most pairs of data points can be connected by a positively signed line, their positive contribution will result in a positive covariance. In Elowitz et al. (2002) the direction along the line y = c is labeled extrinsic, which makes sense in terms of the intuition for positive sample covariance. However we have placed that label “extrinsic” in quotes because the extrinsic noise estimator corresponding directly to the sample variance for points projected onto the line y = c (in analogy with intrinsic noise) is heavily biased and not usable in practice.

To understand the intuition behind Figure 3A in Elowitz et al. (2002), we have redrawn it in a format that highlights the math (Figure 1). The projection of a point (ci, yi) onto the line y = c is the point ((yi+ci)/2,(yici)/2), shown as the red point in Figure 1. Assuming equal means (c¯ = ȳ), the intrinsic noise, as estimated by the unbiased estimator (1), is then the mean squared distance of the points from the line y = c.

The ELSS estimate for the extrinsic noise is the sample covariance. Intuitively, it indicates how the measurements of one reporter track that of the other across cells. The geometric meaning of the sample covariance in Figure 1 is based on an alternative formulation of sample covariance (Hayes, 2011):

(16)Cov(c,y)=2n(n1)i=1n1j>in12(cicj)(yiyj).

This formulation of the sample covariance has the interpretation of being an average of the signed area of triangles associated to pairs of points. Figure 1 illustrates these signed triangles using a randomly selected point (the blue point). This formulation is very different from what might be considered at first glance an appropriate analogy to intrinsic noise, namely the sample variance along the line y = c.

An alternative estimate for the extrinsic noise based on the sample variance of the projected points along the line y = c (using the projected centroid as the mean, which is shown as the green point in Figure 1) turns out to be biased by an amount equal to the total noise. This sample variance averages the squared distances of the data points from the centroid (green point) after projection onto the line y = c; see the distance between the red and green points in Figure 1. Since

S~ext*=1n1i=1n(12(YiY¯+CiC¯))2=12(n1)i=1n((Ci+Yi)2(C¯+Y¯)2)

the bias is

E[S~ext*]σμ2=12Var[Ci+Yi]σμ2=12(Var[Ci]+Var[Yi]+2Cov[Ci,Yi])σμ2=12(2(σ2+σμ2)+2σμ2)σμ2=σ2+σμ2

which is the true total noise.

The above calculation also shows that if the intrinsic and extrinsic noise are both estimated as variances along the projections to the lines y = −c and y = c respectively, then the total noise will be overestimated by a factor of two.

In summary, the caption to Figure 3A in Elowitz et al. (2002) is completely accurate in stating that “Spread of points perpendicular to the diagonal line on which CFP and YFP intensities are equal corresponds to intrinsic noise, whereas spread parallel to this line is increased by extrinsic noise.” However the geometric interpretation of covariance makes it precise how an increase in extrinsic noise relates to the spread of points in the direction of the line y = c.

6 Practical considerations

6.1 Optimal estimators for intrinsic and extrinsic noise

We have derived the estimators that are optimal for minimizing bias or the MSE (summarized in Table 1). The ELSS estimator in (1) is in fact a special case of the general estimator under the assumption that C¯ = Ȳ, and is appropriate for data that are normalized to have the same sample mean (i.e. c¯ = ȳ). In Elowitz et al. (2002), the intensities of the two reporters were normalized to have mean 1. In the case where the assumption of equal reporter means does not hold, the general estimator is more suitable.

Similar to the estimators for the intrinsic noise, we derived two estimators for extrinsic noise, optimized for bias and for MSE respectively (Table 1).

Table 1

Estimators for intrinsic and extrinsic noise. ρ is the correlation between the two reporters, and can be estimated by the sample correlation.

Exact estimator for small nLarge n
Minimizing bias

(Unbiased)
Minimizing MSE
Intrinsic noise
General12(n1)[1n(CiYi)2n(C¯Y¯)2]/(C¯Y¯)12a[1n(CiYi)2n(C¯Y¯)2]/(C¯Y¯), where a=2n37n+62(n2n)+2nn2nρ1ρ+12(n2n)(ρ1ρ)212n[1n(CiYi)2n(C¯Y¯)2]/(C¯Y¯)
Assuming C¯ = Ȳ12ni=1n(CiYi)2/(C¯Y¯)

(ELSS estimator)
12(n+2)i=1n(CiYi)2/(C¯Y¯)12ni=1n(CiYi)2/(C¯Y¯)

(ELSS estimator)
Extrinsic noise
General1n1(i=1nCiYinC¯Y¯)/(C¯Y¯)1a(i=1nCiYinC¯Y¯)/(C¯Y¯),

where a=1/ρ2+(n1)(1+1/n)
1n(i=1nCiYinC¯Y¯)/(C¯Y¯)

(ELSS estimator)

The sample size n is the leading term in the denominator of all the optimal (in either the bias or MSE sense) intrinsic and extrinsic noise estimators. As a result, the unbiased estimator has the same form as the min-MSE estimator for large n (Table 1). For extrinsic noise, the general estimators converge to the ELSS estimate (Table 1). The mean and variance of the estimators are summarized in Table 6 in Appendix E. For intrinsic noise, assuming c¯ = ȳ, the ELSS estimator is optimal for bias and MSE at large n and optimal for bias at small n. Indeed, in Elowitz et al. (2002), typical values for n are greater than 100, making the ELSS formulas suitable for the analyses performed (with the assumption of equal mean satisfied). However, our derivations indicate that the two types of noise can be estimated using fewer cells.

As a general rule we recommend computing the inverse squared correlation between the ci and yi values and applying the min-MSE estimators when the sample size is small (e.g. much less than 50).

It is worth pointing out that the correction factor 1/a in the min-MSE estimators tends to be smaller than that in the unbiased estimators (1/(n − 1)) and the asymptotic estimators (1/n; Table 1). This smaller correction 1/a makes the min-MSE estimators “shrinkage” estimators, such that they achieve better MSE despite being biased, just like the Jame-Stein estimator (James and Stein, 1961). Our simulation results confirm this point (Table 2). However, using the sample correlation, instead of the true one, in our min-MSE estimators leads to increased MSE, although the estimates with the sample correlation do not differ much on average from that with the true correlation.

Table 2

Estimates of extrinsic noise in simulated data. Data were simulated under the hierarchical model, where the conditional distributions of the two reporters are identical. Two min-MSE estimators are applied, one using the true correlation, and the other the sample correlation. Mean estimates (standard deviation in parentheses) of intrinsic and extrinsic noise are summarized. Note that in order to compare the estimates with the true parameters, the estimates are unscaled (i.e. not divided by c¯ȳ).

Simulation parameters
Sample size (n)50
Intrinsic noise (σ2)0.7
Extrinsic noise (σμ2)0.8
Distribution of means (G)N(1, 0.8)
Distribution of vars (H)Constant: Σi2 = 0.7
Distribution of Ci|ZiN(Mi, 0.7)
Distribution of Yi|ZiN(Mi, 0.7)
No. of data sets500
Extrinsic noise estimate
Unbiased0.80 (0.25; 0.0604)
minMSE (true corr)0.73 (0.23; 0.0552)
minMSE (sample corr)0.73 (0.24; 0.0634)
Asymptotic/ELSS0.78 (0.06; 0.0582)

6.2 Data normalization

Our hierarchical model, as well as the ANOVA interpretation, is consistent with the model in Elowitz et al. (2002); both models assume that within each cell there are two distributions for the expression of the two reporter genes and that they have the same true mean and true variance. With the normality assumption, this means that the two reporters have identical distributions. Elowitz et al. measured the single-color distributions of strains that contained lac-repressible promoter pairs, which verified that this was a reasonable assumption in the case of cyan fluorescent protein (CFP) and yellow fluorescent protein (YFP) in their experiment. We also performed simulations under the hierarchical model, with and without identical distribution for the two reporters, and summarized the results in Table 3. Estimates of intrinsic and extrinsic noise are the same as the truth when the identical distribution assumption applies. When this assumption is not satisfied, the theory breaks down and it is unclear what the estimates mean.

Table 3

Estimates of intrinsic and extrinsic noise in simulated data. Data were simulated under two schemes. The first scheme is consistent with the hierarchical model, where the conditional distributions of the two reporters are identical. Under the second scheme, the conditional distributions are different. Intrinsic and extrinsic noise are in fact not defined under the second scheme. Mean estimates (standard deviation in parentheses) of intrinsic and extrinsic noise are summarized. Note that in order to compare the estimates with the true parameters, the estimates are unscaled (i.e. not divided by c¯ȳ).

Identical distributionDifferent distributions
Simulation parameters
Sample size (n)10001000
Intrinsic noise (σ2)0.70.7
Extrinsic noise (σμ2)0.80.8
Distribution of means (G)N(1, 0.8)N(1, 0.8)
Distribution of vars (H)Constant: Σi2 = 0.7Constant: Σi2 = 0.7
Distribution of Ci|ZiN(Mi, 0.7)N(Mi, 0.7)
Distribution of Yi|ZiN(Mi, 0.7)N(2Mi, 1.5 × 0.7)
No. of data sets500500
Sample correlation0.53 (0.02)0.60 (0.02)
Intrinsic noise (σ2^)
General
Unbiased0.70 (0.03)1.54 (0.07)
minMSE0.70 (0.03)1.54 (0.07)
Asymptotic0.70 (0.03)1.54 (0.07)
Equal mean
Unbiased/ELSS0.70 (0.03)2.04 (0.08)
minMSE0.70 (0.03)2.04 (0.08)
Asymptotic/ELSS0.70 (0.03)2.04 (0.08)
Extrinsic noise (σμ2^)
Unbiased0.80 (0.06)1.60 (0.10)
minMSE0.80 (0.06)1.59 (0.10)
Asymptotic/ELSS0.80 (0.06)1.60 (0.10)
σμ2^/(σμ2^+σ2^)
General0.530.51
Equal mean0.530.44

Other studies have adapted this system and used other reporter combinations that may have markedly different distributions. For example, Yang et al. (2014) used CFP and mCherry with vastly different ranges of intensity values: whereas CFP varied from 0 to 6000 (arbitrary units; i.e. a.u.), mCherry could vary from 0 to 9000 (a.u.); see Figure 3A from their paper. In contrast, Schmiedel et al. (2015) normalized the two reporters used in their experiment (ZsGreen and mCherry) to have the same mean. However, the variances, or more generally, the two distributions, also need to be the same. Since the decomposition of the total noise depends on the assumption that both reporters in the same cellular environment have similar variance (see equations 4 and 5), we recommend that in general a quantile normalization which normalizes the reporter measurements to identical distributions be performed before the calculations of noise components. Such a normalization procedure is standard in many settings requiring similar assumptions.

6.3 Assessing the ratio of extrinsic to intrinsic noise from sample correlation

We have seen from (13) that the proportion of the between-cell variability to total variability is the correlation ρ(C, Y). This leads to a simple approach for estimating the relative magnitude of the two types of noise: one can compute the sample correlation of the expression of the two reporters, ρ(c,y), and the ratio of extrinsic to intrinsic noise is then estimated by ρ(c, y)/[1 − ρ(c,y)].

7 Re-analysis of published two-reporter experiment data

Michael Elowitz and Peter Swain have kindly shared with us their data published in Elowitz et al. (2002). Here we focus on the data in Figure 3A of their paper, which contain the unnormalized fluorescence intensities of CFP and YFP in the E. coli. strain D22 and in strain M22. We normalized the data as follows such that the resulting scatterplots are close to Figure 3A:

D22:ci=(ci*c*¯)/(8sc*)+1;yi=(yi*y*¯)/(8sy*)+1;M22:ci=(ci*c*¯)/(12sc*)+1;yi=(yi*y*¯)/(12sy*)+1,

where ci* and yi* are the unnormalized intensity of the CFP and YFP, respectively, in the ith cell, c*¯ the sample mean, and sy* the sample standard deviation. The normalized intensities are close to normal distributions, and all four distributions have mean 1. At a sample size of over 200, the different estimators in Table 4 give essentially the same result. Additionally, the ratio of the estimated extrinsic and total noise is close to the sample correlation, verifying our theoretical result.

Table 4

Re-analysis of published two-reporter experiment data. Summary statistics and estimates (×10−2) of intrinsic and extrinsic noise are listed, using the estimators from Table 1.

Elowitz et al. dataYang et al. data
D22M22Figure 3ANormalized on log2
Sample meansCFP: 1CFP: 1CFP: 2660CFP: 11
YFP: 1YFP: 1mCherry: 3986mCherry: 11
Sample correlation0.500.490.860.86
Intrinsic noise
General
 Unbiased0.790.365.440.11
 minMSE0.780.355.440.11
 Asymptotic0.780.355.440.11
Equal mean
 Unbiased/ELSS0.780.3513.720.11
 minMSE0.780.3513.720.11
 Asymptotic/ELSS0.780.3513.720.11
Extrinsic noise
 Unbiased0.780.3430.290.68
 minMSE0.760.3330.290.68
 Asymptotic/ELSS0.770.3430.290.68
σμ2^/(σμ2^+σ2^)
General0.500.490.850.86
Equal mean0.500.490.690.86

Nam Ki Lee and Sora Yang have also kindly shared with us their data published in Yang et al. (2014). Here we analyze the data in Figure 3A of their paper, which are the expression levels (intensities) of two reporters, CFP and mCherry (also see Sec. 6.2). The shared, unnormalized intensities have very different sample means (Table 4). Application of the estimators in Table 1 to these data gives two different estimates of the intrinsic noise, with the ELSS estimate being nearly three times the estimates under the equal mean assumption. To normalize the data, we removed the few negative values, log2 transformed the data, and quantile normalized between the two reporters (see summary statistics in Table 4). Applying our estimators to the normalized data, all estimates are consistent with one another. This analysis illustrates the importance of the equal mean assumption: when this assumption is not satisfied, the ELSS estimator leads to overestimation of the intrinsic noise.

Additionally, we subsampled from these data sets and assessed the performance of the estimators as the sample size decreased. At each sample size, we repeated the subsampling 1000 times and computed the mean and standard deviation of the noise estimates (Table 5). Whereas the means of the estimates do not differ from those obtained using the entire data sets, the variation (measured by the standard deviation) increases quickly with decreasing sample sizes. For the Elowitz et al. data, the standard deviation in the estimates roughly doubles for both types of noise as the sample size halves. Comparing the standard deviation to the mean suggests that 200 is indeed a reasonable sample size for estimates with small variation (compare with their actual sample sizes of 284 and 250 for the two strains). For the Yang et al. data, the increase in the standard deviation is much less drastic, and 200 also appears a decent sample size for reasonably small variation in the estimates.

Table 5

Noise estimates (×10−2) based on subsets of published data. Similar to Table 4, we used the estimators from Table 1.

Elowitz et al. dataYang et al. data
D22M22Normalized on log2
Original sample size28425040658
n = 200
Intrinsic noise
GeneralUnbiased0.79 (0.06)0.36 (0.02)0.11 (0.02)
minMSE0.78 (0.06)0.35 (0.02)0.11 (0.02)
Asymptotic0.78 (0.06)0.35 (0.02)0.11 (0.02)
Equal meanUnbiased/ELSS0.78 (0.06)0.35 (0.02)0.11 (0.02)
minMSE0.78 (0.06)0.35 (0.02)0.11 (0.02)
Asymptotic/ELSS0.78 (0.06)0.35 (0.02)0.11 (0.02)
Extrinsic noiseUnbiased0.78 (0.07)0.34 (0.02)0.68 (0.09)
minMSE0.76 (0.07)0.33 (0.02)0.67 (0.08)
Asymptotic/ELSS0.78 (0.07)0.34 (0.02)0.68 (0.08)
n = 100
Intrinsic noise
GeneralUnbiased0.79 (0.13)0.36 (0.04)0.11 (0.03)
minMSE0.77 (0.12)0.35 (0.04)0.11 (0.03)
Asymptotic0.78 (0.12)0.35 (0.04)0.11 (0.03)
Equal meanUnbiased/ELSS0.78 (0.12)0.35 (0.04)0.11 (0.03)
minMSE0.77 (0.12)0.35 (0.04)0.11 (0.03)
Asymptotic/ELSS0.78 (0.12)0.35 (0.04)0.11 (0.03)
Extrinsic noiseUnbiased0.77 (0.14)0.34 (0.05)0.69 (0.12)
minMSE0.73 (0.14)0.32 (0.05)0.67 (0.12)
Asymptotic/ELSS0.76 (0.14)0.34 (0.05)0.68 (0.12)
n = 50
Intrinsic noise
GeneralUnbiased0.78 (0.21)0.36 (0.07)0.11 (0.04)
minMSE0.75 (0.20)0.35 (0.07)0.11 (0.04)
Asymptotic0.77 (0.20)0.35 (0.07)0.11 (0.04)
Equal meanUnbiased/ELSS0.78 (0.21)0.36 (0.07)0.11 (0.04)
minMSE0.75 (0.20)0.34 (0.07)0.11 (0.04)
Asymptotic/ELSS0.78 (0.21)0.36 (0.07)0.11 (0.04)
Extrinsic noiseUnbiased0.78 (0.24)0.34 (0.09)0.68 (0.16)
minMSE0.70 (0.24)0.30 (0.09)0.65 (0.15)
Asymptotic/ELSS0.76 (0.23)0.33 (0.09)0.66 (0.16)

8 Conclusions and discussion

Our hierarchical model for Elowitz et al. (2002) provides statistically interpretable parameters representing intrinsic and extrinsic noise, and allows for the derivation of estimators with optimality guarantees. Furthermore, the model highlights experimental assumptions that need to be satisfied for the estimators to be valid, specifically that the two reporters need to have the same distribution (within a cell) and hence normalization may be necessary. Whereas similar hierarchical models have been proposed before to study heterogeneity among single cells (see, e.g. Finkenstädt et al., 2013, and Koeppl et al., 2012), our hierarchical model explicitly parameterize the two types of noise, and reveals their equivalence to other quantities, as indicated by (11) and (14), which enable derivation of closed-form estimators of these parameters (summarized in Table 1). We use bias and MSE to explicitly evaluate the performance of different estimators, and recognize the asymptotic equivalence of multiple estimators.

Other experiments have been set up to explore and assess intrinsic and extrinsic noise, and some of our results may be useful in those settings. For example, Volfson et al. (2006) used a single reporter but two Saccharomyces cerevisiae strains, with one strain containing only one copy of the reporter, and the other strain two copies. Assuming no strain effect, which may be thought of as batch effect, the authors applied the following estimators for (unscaled) intrinsic and extrinsic noise (consistent with their notation, and without the denominator of C¯Ȳ as used in the ELSS estimators in Table 1):

(17)Vi=2V1V2/2;
(18)Ve=V2/2V1,

where V1 and V2 are the variance in the 1-copy and 2-copy strains, respectively, and Vi and Ve are intrinsic and extrinsic noise, respectively. These estimators are in fact consistent with (11) and (14) under our hierarchical model:

(19)V1=Var[C1]=Vi+Ve=Var[C2];
(20)V2=Var[C1+C2]=Var[C1]+Var[C2]+2Cov[C1,C2]=2(Vi+Ve)+2Ve.

Together, (19) and (20) give rise to (17) and (18). Note that (19) and (20) imply that the extrinsic noise is also the covariance here, except that the covariance is between the 1-copy and 2-copy strains with the same reporter; this is also pointed out by Sherman et al. (2015). Additionally, the total (marginal) noise of the reporter is the sum of intrinsic and extrinsic noise (19). However, consistent with our analysis of the assumptions of the hierarchical model, these estimators hold only when the variance for each single copy in the 2-copy strain is identical to that in the 1-copy strain. This is equivalent to assuming no strain (batch) effect, which can be a rather strong assumption.

We note that during the preparation of this manuscript, Erik van Nimwegen independently examined the Elowitz et al. (2002) paper form a Bayesian point of view (van Nimwegen, 2016).

Acknowledgement

This project began as a result of discussion during a journal club meeting of Prof. Jonathan Pritchard’s group that A.F. was attending. We thank Michael Elowitz, Peter Swain, Nam Ki Lee and Sora Yang for sharing their data from Elowitz et al. (2002) and from Yang et al. (2014), respectively. We also thank helpful comments we have received since posting the manuscript online. In particular, we thank Arjun Raj for bringing up the 1- vs 2-copy experiment, and Erik van Nimwegen for helpful discussions. We also thank Editor in Chief Prof. Michael Stumpf and two anonymous reviewers for insightful comments that led to a significantly enriched version. A.F. was partially supported by K99 HG007368 and R00 HG007368 (NIH/NHGRI). L.P. was partially supported by NIH grants R01 HG006129 and R01 DK094699.

A Moments of Mi and Ci under normality

Assuming that MiN(μ,σμ2), we have

E[Miμ]3=0;E[Miμ]4=3σμ4.

We can compute the third and fourth moments of Mi as follows:

E[Miμ]3=E[Mi2+μ22Miμ)(Miμ]=E[Mi32Mi2μ+Miμ2Mi2μμ3+2Miμ2]=E[Mi33Mi2μ+3Miμ2μ3]=E[Mi3]3μ(σμ2+μ2)+3μ3μ3=E[Mi3]3μσμ2μ3,

which gives

E[Mi3]=3μσμ2+μ3.
E[Miμ]4=E[Mi22Miμ+μ2]2=E[Mi4+μ4+4Mi2μ2+2Mi2μ24Mi3μ4Miμ3]=E[Mi4+μ4+6Mi2μ24Mi3μ4Miμ3]=E[Mi4]+μ4+6μ2(σμ2+μ2)4μ(3μσμ2+μ3)4μ4=E[Mi4]+μ4+6μ2σμ2+6μ412μ2σμ24μ44μ4=E[Mi4]6μ2σμ2μ4,

which gives

E[Mi4]=3σμ4+6μ2σμ2+μ4.

For the random variable Ci, since Σi2H(σ2,ϵ), such that

E[Σi2]=σ2;Var[Σi2]=ϵ,

we have

E[Ci4]=E[E[Ci4|Zi]]=E[3Σi4+6Mi2Σi2+Mi4]=3(ϵ+σ4)+6(σμ2+μ2)σ2+3σμ4+6μ2σμ2+μ4=3ϵ+3σ4+6σμ2σ2+6μ2σ2+3σμ4+6μ2σμ2+μ4.

Further assuming that μ = 0, i.e. the means are all 0, and that ϵ = 0, which means that the variability is the same across cells, we have

E[Mi3]=0E[Mi4]=3σμ4;

and

E[Ci3]=0E[Ci4]=3(σ2+σμ2)2.

B Calculating Var[Sext]

Var[Sext]=Var[1a(i=1nCiYinC¯Y¯)]=1a2Var[i=1nCiYinC¯Y¯]=1a2(Var[i=1nCiYi]+Var[nC¯Y¯]2Cov[i=1nCiYi,nC¯Y¯]).

B.1 Calculating Var[i=1nCiYi]

Var[i=1nCiYi]=i=1nVar[CiYi]=i=1n(E[Ci2Yi2](E[CiYi])2)

where

E[CiYi]2=E[E[Ci2Yi2|Zi]]=E[E[Ci2|Zi)E(Yi2|Zi]]=E[Σi2+Mi2]2=E[Σi4+Mi4+2Σi2Mi2]=Var[Σi2]+(E[Σi2])2+E[Mi4]+2E[Σi2]E[Mi2]=ϵ+σ4+E[Mi4]+2σ2(σμ2+μ2);

and

E[CiYi]=Cov[Ci,Yi]+E[Ci]E[Yi]=σμ2+μ2.

Therefore,

Var[i=1nCiYi]=i=1n(ϵ+σ4+EMi4+2σ2(σμ2+μ2)(σμ2+μ2)2).

B.2 Calculating Var[nC¯Y¯]

Var[nC¯Y¯]=n2Var[C1++CnnY1++Ynn]=n2n4Var[kCkYk+ijCiYj]=1n2(Var[kCkYk]+Var[ijCiYj]+2Cov[kCkYk,ijCiYj]).

Assuming normality on Mi and assuming that μ = 0 and ϵ = 0 (constant variance across cells), we have

Var[kCkYk]=n(σ4+3σμ4+2σ2σμ2σμ4)=n(σ2+σμ2)2+nσμ4.

Also,

Var[ijCiYj]=ijVar[CiYj]+2i=korj=lCov[CiYj,CkYl]+2ikandjlCov[CiYj,CkYl].

Under the assumptions made above, we have

Var[CiYj]=E[Ci2Yj2](E[CiYj])2=E[Ci2]E[Yj2](E[Ci]E[Yj])2=(σ2+σμ2)2.

If i = k,

Cov[CiYj,CkYl]=E[CiYjCkYl]E[CiYj]E[CkYl]=E[Ci2]E[Yj]E[Yl](E[Ci])2E[Yj]E[Yl]=0.

Similarly, we can derive that the covariance is 0 for other cases where j = l or where ik and jl. Hence,

Var[ijCiYj]=n(n1)(σ2+σμ2)2.

Additionally, under the normality assumption and with μ = 0 and ϵ = 0,

Cov[kCkYk,ijCiYj]=0.

Therefore,

Var[nC¯Y¯]=1n2(n(σ2+σμ2)2+nσμ4+n(n1)(σ2+σμ2)2)=1n2(n2(σ2+σμ2)2+nσμ4)=(σ2+σμ2)2+σμ4n.

B.3 Calculating Cov[i=1nCiYi,nC¯Y¯]

Cov[i=1nCiYi,nC¯Y¯]=1nCov[i=1nCiYi,kCkYk+ijCiYj]=1n(Cov[i=1nCiYi,kCkYk]+Cov[i=1nCiYi,ijCiYj])=1n(Var[i=1nCiYi])=(σ2+σμ2)2+σμ4.

Putting the terms above together, we have

Var[Sext]=1a2(n(σ2+σμ2)2+nσμ4+(σ2+σμ2)2+σμ4n2(σ2+σμ2)22σμ4)=n1a2(σ2+σμ2)2+(n1)2na2σμ4.

C MSE of the general intrinsic noise estimator

The general form of the estimator for intrinsic noise is

S=12a(1n(CiYi)2n(C¯Y¯)2).

C.1 Calculating Var[S]

Thus

Var[S]=14a2(Var[(CiYi)2]+n2Var[(C¯Y¯)2]2nCov[(CiYi)2,(C¯Y¯)2]).

Below we will assume normality, as well as μ = 0 and ϵ = 0, to facilitate the derivation. Note that Var[(CiYi)2] is derived in Appendix D.

C.1.1 Calculating Var[(C¯Y¯)2]

First, we note that

Var[(C¯Y¯)2]=Var[C¯22C¯Y¯+Y¯2]=Var[C¯2]+4Var[C¯Y¯]+Var[Y¯2]4Cov[C¯2,C¯Y¯]4Cov[Y¯2,C¯Y¯]+2Cov[C¯2,Y¯2].
Var[C¯2]=Var[C1++CnnC1++Cnn]=1n4Var[Ck2+ijCiCj]=1n4(VarCk2+Var[ijCiCj]+2Cov[Ck2,ijCiCj])=1n4(2n(σ2+σμ2)2+n(n1)(σ2+σμ2)2+0)=n+1n3(σ2+σμ2)2.

This is because

Var[ijCiCj]=ijVar[CiCj]=ij(ECi2Cj2(ECiCj)2)=ij((σ2+σμ2)20)=n(n1)(σ2+σμ2)2.

Additionally, from Appendix B, we have

Var[C¯Y¯]=1n2Var[nC¯Y¯]=1n2((σ2+σμ2)2+σμ4n)=1n2(σ2+σμ2)2+σμ4n3.
Cov[C¯2,C¯Y¯]=1n4Cov[Ck2+ijCiCj,ClYl+mrCmCr]=1n4(Cov[Ck2,ClYl]+Cov[Ck2,mrCmCr]+Cov[ijCiCj,ClYl]+Cov[ijCiCj,mrCmCr]).
Cov[Ck2,ClYl]=Cov[Ck2,CkYk]=(E[Ck3Yk]E[Ck2]E[CkYk])=[3σ2σμ2+3σμ4(σ2+σμ2)σμ2]=2nσμ2(σ2+σμ2).

For Cov[Ck2,mrCmCr], since

Cov[Ci2,CiYj]=E[Ci3Yj]E[Ci2]E[CiYj]=0

and

Cov[Ci2,CjYk]=E[Ci2CjYk]E[Ci2]E[CjYk]=0,

we have

Cov[Ck2,mrCmCr]=0.

For Cov[ijCiCj,ClYl], since

Cov[CiCj,CiYi]=E[Ci2YiCj]E[CiCj]E[CiYi]=0

and

Cov[CkCl,CiYi]=E[CkClCiYi]E[CkCl]E[CiYi]=0,

we have

Cov[ijCiCj,ClYl]=0.

Additionally,

Cov[ijCiCj,mrCmCr]=i,j,m,rCov[CiCj,CmCr]=ijCov[CiCj,CiCj]=ijVar[CiCj]=n(n1)(σ2+σμ2)2.

Therefore,

Cov[C¯2,C¯Y¯]=2n3σμ2(σ2+σμ2)+n1n3(σ2+σμ2)2.

Furthermore,

Cov[C¯2,Y¯2]=1n4Cov[Ck2+ijCiCj,Yl2+mrYmYr]=1n4(Cov[Ck2,Yl2]+Cov[Ck2,mrYmYr]+Cov[Yl2,ijCiCj]+Cov[ijCiCj,mrYmYr]).

In the expression above,

Cov[Ck2,Yl2]=2nσμ4;
Cov[Ck2,mrYmYr]=Cov[Yl2,ijCiCj]=0;
Cov[ijCiCj,mrYmYr]=ijCov[CiCj,YiYj]=ij(E[CiCjYiYj]E[CiCj]E[YiYj])=ij(E[CiYi]E[CjYj]0)=n(n1)σμ4.

Then we have

Cov[C¯2,Y¯2]=1n4(2nσμ4+n(n1)σμ4)=n+1n3σμ4.

Putting the terms together, we have

Var[C¯Y¯]2=Var[C¯2]+4Var[C¯Y¯]+Var[Y¯2]4Cov[C¯2,C¯Y¯]4Cov[Y¯2,C¯Y¯]+2Cov[C¯2,Y¯2]=2(n+1)n3(σ2+σμ2)2+4n2(σ2+σμ2)2+4σμ4n316n3σμ2(σ2+σμ2)8(n1)n3(σ2+σμ2)2+2(n+1)n3σμ4=2n3((6n)(σ2+σμ2)28σμ2(σ2+σμ2)+(n+3)σμ4)=2n3((6n)σ4+(42n)σ2σμ2+σμ4).

C.1.2 Calculating Cov[(CiYi)2,(C¯Y¯)2]

Next, we note that

Cov[(CiYi)2,(C¯Y¯)2]=Cov[(CiYi)2,(C¯Y¯)2]=(E[(Ci22CiYi+Yi2)(C¯22C¯Y¯+Y¯2)]E[(Ci22CiYi+Yi2)]E[(C¯22C¯Y¯+Y¯2)]),

where

E[(Ci22CiYi+Yi2)(C¯22C¯Y¯+Y¯2)]=E[Ci2C¯22CiYiC¯2+Yi2C¯22Ci2C¯Y¯+4CiYiC¯Y¯2Yi2C¯Y¯+Ci2Y¯22CiYiY¯2+Yi2Y¯2],

and

E[Ci22CiYi+Yi2]=2(σ2+σμ2)2σμ2=2σ2,
E[C¯22C¯Y¯+Y¯2]=2n(σ2+σμ2)2nσμ2=2σ2n.
E[Ci2C¯2]=1n2E[Ci2(Ck2+ijCiCj)]=1n2(E[Ci4]+kiE[Ci2]E[Ck2]+ijE[Ck2CiCj])=1n2[3(σ2+σμ2)2+(n1)(σ2+σμ2)2+0]=n+2n2(σ2+σμ2)2.
E[CiYiC¯2]=E[CiYiCj2+klCkCln2]=1n2(E[CiYiCi2]+jiE[CiYiCj2]+klE[CiYiCkCl])=1n2(3(σ2σμ2+σμ4)+(n1)(σ2σμ2+σμ4)+0)=n+2n2σμ2(σ2+σμ2).
E[Yi2C¯2]=E[Yi2Cj2+klCkCln2]=1n2(E[Yi2Ci2]+jiE[Yi2Cj2]+klE[Yi2CkCl])=1n2((σ2+σμ2)2+2σμ4+(n1)(σ2+σμ2)2+0)=1n2(n(σ2+σμ2)2+2σμ4).
E[Ci2C¯Y¯]=1n2(E[Ci2CiYi]+jiE[Ci2CjYj]+klE[Ci2CkYl])=1n2(3(σ2σμ2+σμ4)+(n1)(σ2σμ2+σμ4)+0)=n+2n2σμ2(σ2+σμ2).
E[CiYiC¯Y¯]=1n2(E[Ci2Yi2]+jiE[CiYiCjYj]+klE[CiYiCkYl])=1n2((σ2+σμ2)2+2σμ4+(n1)σμ4+0)=1n2((σ2+σμ2)2+(n+1)σμ4).

Additionally,

E[Yi2C¯Y¯]=E[Ci2C¯Y¯]=n+2n2σμ2(σ2+σμ2);E[Ci2Y¯2]=E[Yi2C¯2]=1n2(n(σ2+σμ2)2+2σμ4);E[CiYiY¯2]=E[CiYiC¯2]=n+2n2σμ2(σ2+σμ2);E[Yi2Y¯2]=E[Ci2C¯2]=n+2n2(σ2+σμ2)2.

Therefore,

E[(Ci22CiYi+Yi2)(C¯22C¯Y¯+Y¯2)]=E[Ci2C¯22CiYiC¯2+Yi2C¯22Ci2C¯Y¯+4CiYiC¯Y¯2Yi2C¯Y¯+Ci2Y¯22CiYiY¯2+Yi2Y¯2]=2(n+2)n2(σ2+σμ2)24(n+2)n2σμ2(σ2+σμ2)+2n2(n(σ2+σμ2)2+2σμ4)4(n+2)n2σμ2(σ2+σμ2)+4n2((σ2+σμ2)2+(n+1)σμ4)=4(n+2)σ4n2.

So we have

Cov[(CiYi)2,(C¯Y¯)2]=Cov[(CiYi)2,(C¯Y¯)2]=(E(Ci22CiYi+Yi2)(C¯22C¯Y¯+Y¯2)E(Ci22CiYi+Yi2)E(C¯22C¯Y¯+Y¯2))=n(4(n+2)σ4n22σ22σ2n)=8σ4n.

The variance of the estimator is then

Var[S]=14a2(Var[(CiYi)2]+n2Var[C¯Y¯]22nCov[(CiYi)2,(C¯Y¯)2])=14a2(8nσ4+2n((6n)σ4+(42n)σ2σμ2+σμ4)16σ4)=12a2(4nσ4+1n((6n)σ4+(42n)σ2σμ2+σμ4)8σ4).

C.2 Calculating E[S]

The expectation of the estimator is

E[S]=12a(E[CiYi]2nE[C¯Y¯]2),

where

E[(CiYi)2]=Var[CiYi]=Var[Ci]+Var[Yi]2Cov[Ci,Yi]=2(σ2+σμ2)2σμ2=2σ2,

and

E[(C¯Y¯)2]=Var[C¯Y¯]=Var[C¯]+Var[Y¯]2Cov[C¯,Y¯]=2n(σ2+σμ2)2nσμ2=2σ2n.

Hence,

E[S]=12a(2nσ22σ2)=n1aσ2.

C.3 Calculating the MSE

The MSE of the estimator is then

E[(Sσ2)2]=Var[S]+(E[S]σ2)2=12a2(4nσ4+1n((6n)σ4+(42n)σ2σμ2+σμ4)8σ4)+(n1a1)2σ4=12a2(4nσ4+1n((6n)σ4+(42n)σ2σμ2+σμ4)8σ4+2(n1)2σ4)2(n1)σ41a+σ4=12a2((2n2+6n7)σ4+2(2n1)σ2σμ2+1nσμ4)2(n1)σ41a+σ4.

The value of a that minimizes this MSE is

a=(2n37n+6)σ4+2(2n)σ2σμ2+σμ42(n2n)σ4=2n37n+62(n2n)+2nn2nσμ2σ2+12(n2n)(σμ2σ2)2.

D Calculating Var[S~int]

Var[S~int]=14a2Var[i=1n(CiYi)2]=14a2Var[i=1n(Ci2+Yi22CiYi)]=14a2Var[i=1nCi2+i=1nYi22i=1nCi,Yi]=14a2(Var[i=1nCi2]+Var[i=1nYi2]+4Var[i=1nCiYi]+2Cov[i=1nCi2,i=1nYi2]4Cov[i=1nCi2,i=1nCiYi]4Cov[i=1nYi2,i=1nCiYi]).

The individual terms can be computed as follows:

Var[i=1nCi2]=i=1nVar[Ci2]=i=1n(E[Ci4](E[Ci2])2)=i=1n(E[Ci4](Var[Ci]+(E[Ci])2)2)=i=1n(E[Ci4](σ2+σμ2+μ2)2)=nEC14n(σ2+σμ2+μ2)2.

Assuming normality, we have

Var[i=1nCi2]=n(3ϵ+3σ4+6σμ2σ2+6μ2σ2+3σμ4+6μ2σμ2+μ4(σ2+σμ2+μ2)2)=n(3ϵ+2σ4+2σμ4+4σ2σμ2+4μ2σ2+4μ2σμ2).

Assuming additionally that μ = 0 and ϵ = 0, we have

Var[i=1nCi2]=2n(σ2+σμ2)2.

Since Ci and Yi are symmetrically defined, we have

Var[i=1nYi2]=Var[i=1nCi2].

Next, from Appendix B,

Var[i=1nCiYi]=i=1n(ϵ+σ4+EMi4+2σ2(σμ2+μ2)(σμ2+μ2)2).

Assuming normality, we have

E[CiYi]2=ϵ+σ4+3σμ4+6μ2σμ2+μ4+2σ2σμ2+2σ2μ2;E[CiYi]=σμ2+μ2;Var[i=1nCiYi]=n(ϵ+σ4+2σμ4+2σ2σμ2+2μ2σ2+4μ2σμ2).

Assuming additionally that μ = 0 and ϵ = 0, we have

E[CiYi]2=(σ2+σμ2)2+2σμ4;E[CiYi]=σμ2;Var[i=1nCiYi]=n[(σ2+σμ2)2+σμ4].

The covariance terms are computed as follows:

Cov[i=1nCi2,i=1nYi2]=i=1nCov[Ci2,Yi2]=i=1n(E[Ci2Yi2]E[Ci2]E[Yi2]).

Assuming normality, we have

Cov[i=1nCi2,i=1nYi2]=n(ϵ+σ4+3σμ4+6μ2σμ2+μ4+2σ2σμ2+2σ2μ2(σ2+σμ2+μ2)2)=n(ϵ+2σμ4+4μ2σμ2).

Assuming additionally that μ = 0 and ϵ = 0, we have

Cov[i=1nCi2,i=1nYi2]=2nσμ4.

Finally, since Ci and Yi are symmetrically defined, we have

Cov[i=1nCi2,i=1nCiYi]=Cov[i=1nYi2,i=1nCiYi]=i=1nCov[Ci2,CiYi]=i=1n(E[Ci3Yi]E[Ci2]E[CiYi]),

where

E[Ci3Yi]=E[E[Ci3Yi|Zi]]=E[E[Ci3|Zi]E[Yi|Zi]].

Assuming normality, we have

E[Ci3Yi]=E[(3MiΣi2+Mi3)Mi]=E[3Mi2Σi2+Mi4]=3E[Mi2]E[Σi2]+E[Mi4]=3(σμ2+μ2)σ2+3σμ4+6μ2σμ2+μ4=μ4+3σμ4+3σ2σμ2+3μ2σ2+6μ2σμ2;E[Ci2]=σ2+σμ2+μ2;E[CiYi]=σμ2+μ2;

and therefore,

Cov[i=1nCi2,i=1nCiYi]=n(μ4+3σμ4+3σ2σμ2+3μ2σ2+6μ2σμ2(σ2+σμ2+μ2)(σμ2+μ2))=n(μ4+3σμ4+3σ2σμ2+3μ2σ2+6μ2σμ2(μ4+σμ4+σ2σμ2+μ2σ2+2μ2σμ2))=2n(σμ4+σ2σμ2+μ2σ2+2μ2σμ2).

Assuming additionally that μ = 0 and ϵ = 0, we have

E[Ci3Yi]=3σ2σμ2+3σμ4;E[Ci2]=σ2+σμ2;E[CiYi]=σμ2;Cov[i=1nCi2,i=1nCiYi]=2nσμ2(σ2+σμ2).

Putting the terms together, we derive the variance as follows, assuming that Mi follows a normal distribution,

Var[S~int]=14a2{2n(3ϵ+2σ4+2σμ4+4σ2σμ2+4μ2σ2+4μ2σμ2)+4n(ϵ+σ4+2σμ4+2σ2σμ2+2μ2σ2+4μ2σμ2)+2n(ϵ+2σμ4+4μ2σμ2)16n(σμ4+σ2σμ2+μ2σ2+2μ2σμ2)}=na2(3ϵ+2σ4).

Assuming additionally that μ = 0 and ϵ = 0, we have

Var[S~int]=2na2σ4.

E Summary of mean and variance of the estimators

We summarize the mean and variance of the estimators in Table 6.

Table 6

Mean and variance of the estimators in Table 1. Note that only the numerators of the estimators in the general forms are considered here; that is, scalar a can take different values depending on which specific estimator is of interest. Values of a can be found in Table 1. As in the main text, we assume normality of all distributions, and that μ = 0 and ϵ = 0, when deriving the mean and variance.

EstimatorMeanVariance
Intrinsic noise
General
12a(1n(CiYi)2n(C¯Y¯)2)n1aσ212a2(4nσ4+1n((6n)σ4+(42n)σ2σμ2+σμ4)8σ4)
Equal mean
12ai=1n(CiYi)2naσ22na2σ4
Extrinsic noise
1a(i=1nCiYinC¯Y¯)n1aσμ2n1a2(σ2+σμ2)2+(n1)2na2σμ4

References

Elowitz, M. B., A. J. Levine, E. D. Siggia and P. S. Swain (2002): “Stochastic gene expression in a single cell,” Science, 297, 1183–1186.10.1126/science.1070919Search in Google Scholar PubMed

Finkenstädt, B., D. J. Woodcock, M. Komorowski, C. V. Harper, J. R. Davis, M. R. White and D. A. Rand (2013): “Quantifying intrinsic and extrinsic noise in gene transcription using the linear noise approximation: an application to single cell data,” Ann. Appl. Stat., 7, 1960–1982.10.1214/13-AOAS669Search in Google Scholar

Hayes, K. (2011): “A geometrical interpretation of an alternative formula for the sample covariance,” Am. Stat., 65, 110–112.10.1198/tast.2011.09067Search in Google Scholar

Hilfinger, A. and J. Paulsson (2011): “Separating intrinsic from extrinsic fluctuations in dynamic biological systems,” Proc. Natl. Acad. Sci. USA, 108, 12167–12172.10.1073/pnas.1018832108Search in Google Scholar PubMed PubMed Central

James, W. and C. Stein (1961): “Estimation with quadratic loss,” Proc. Fourth Berkeley Symp. Math. Stat. Prob., 1, 361–379.10.1007/978-1-4612-0919-5_30Search in Google Scholar

Koeppl, H., C. Zechner, A. Ganguly, S. Pelet and M. Peter (2012): “Accounting for extrinsic variability in the estimation of stochastic rate constants,” Int. J. Robust Nonlin., 22, 1103–1119.10.1002/rnc.2804Search in Google Scholar

Komorowski, M., J. Mie¸kisz and M. P. Stumpf (2013): “Decomposing noise in biochemical signaling systems highlights the role of protein degradation,” Biophys. J., 104, 1783–1793.10.1016/j.bpj.2013.02.027Search in Google Scholar PubMed PubMed Central

Rausenberger, J. and M. Kollmann (2008): “Quantifying origins of cell-to-cell variations in gene expression,” Biophys. J., 95, 4523–4528.10.1529/biophysj.107.127035Search in Google Scholar PubMed PubMed Central

Schmiedel, J. M., S. L. Klemm, Y. Zheng, A. Sahay, N. Blüthgen, D. S. Marks and A. van Oudenaarden (2015): “MicroRNA control of protein expression noise,” Science, 348, 128–232.10.1126/science.aaa1738Search in Google Scholar PubMed

Sherman, M. S., K. Lorenz, M. H. Lanier and B. A. Cohen (2015): “Cell-to-cell variability in the propensity to transcribe explains correlated fluctuations in gene expression,” Cell Syst., 1, 315–325.10.1016/j.cels.2015.10.011Search in Google Scholar PubMed PubMed Central

Stegle, O., S. A. Teichmann and J. C. Marioni (2015): “Computational and analytical challenges in single-cell transcriptomics,” Nat. Rev. Genet., 16, 133–145.10.1038/nrg3833Search in Google Scholar PubMed

van Nimwegen, E. (2016): “Inferring intrinsic and extrinsic noise from a dual fluorescent reporter,” bioRxiv 049486; doi: http://dx.doi.org/10.1101/049486.Search in Google Scholar

Volfson, D., J. Marciniak, W. J. Blake, N. Ostroff, L. S. Tsimring and J. Hasty (2006): “Origins of extrinsic variability in eukaryotic gene expression,” Nature, 439, 861–864.10.1038/nature04281Search in Google Scholar PubMed

Yang, S., S. Kim, Y. R. Lim, C. Kim, H. J. An, J.-H. Kim, J. Sung and N. K. Lee (2014): “Contribution of RNA polymerase concentration variation to protein expression noise,” Nat. Commun., 5, 4761.10.1038/ncomms5761Search in Google Scholar PubMed

Published Online: 2016-11-22
Published in Print: 2016-12-1

©2016 Walter de Gruyter GmbH, Berlin/Boston

Downloaded on 1.5.2024 from https://www.degruyter.com/document/doi/10.1515/sagmb-2016-0002/html
Scroll to top button