generated from rstudio/bookdown-demo
-
Notifications
You must be signed in to change notification settings - Fork 0
/
98-appendix.Rmd
308 lines (224 loc) · 20.9 KB
/
98-appendix.Rmd
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
# Appendix {-}
<!-- ## Gaussian Process{#gpbasics} -->
<!-- Let assume ti have cloud of points represented by two variables $X_1$ and $X_2$, figure \@ref(fig:gpcloud). The cloud of points are observation taken from a realization of two variables e.g. height and weight, What it might be observed is: -->
<!-- ![(#fig:gpclouds)Left panel: GP cloud of points fitted with a _circle_, Right panel: GP cloud of points fitted with a _ellipse_, source @YT:freitas1](appendix_images/gp_base_1.jpg) -->
<!-- Each circle in both panels figure \@ref(fig:gpclouds) represents a measurements of the observed phenomenon. Let assume to fit a multivariate gaussian distribution to the left panel data. The process of learning is to fit a Gaussian to data, the ultimate goal is to describe the data, the most perfectly evident Gaussian is the one interpolating the points and centering in $\left(0,0\right)$, a circle might be a good guess. Instead for the right one in \@ref(fig:gpclouds), a smart guess could be still centering the mean in, indeed now it is an ellipse describing the variability. -->
<!-- At this point it might be interesting to vectorized what is has been measured, the centers are then compressed into a vector $\mu_{i}$, i.e. it has two components $X_1$ and $X_2$ whose corresponding mean is 0. -->
<!-- $$ -->
<!-- \boldsymbol{\mu}=\left[\begin{array}{ll} -->
<!-- \mu_{x_1} \\ -->
<!-- \mu_{x_2} -->
<!-- \end{array}\right] -->
<!-- $$ -->
<!-- This is true for all the observations which have two coordinates too $x_1$ and $x_2$. for each of the points, e.g. for point 1: -->
<!-- $$ -->
<!-- \mathbf{x_1}=\left[\begin{array}{ll} -->
<!-- x_1 \\ -->
<!-- x_2 -->
<!-- \end{array}\right] -->
<!-- $$ -->
<!-- the can be neagtive positive, the Real numbers, usually we have $\mathbb{R}^{2}$ extending from - infinity to + infinity, to the power of two because we have 2 dimensions, a Real plane. -->
<!-- any point is gaussian distributed when with mean .. an variance. -->
<!-- how we explain covariance, thorugh _correlation._ -->
<!-- we do it by correlation with its noraml forms. the covariance is the term that goes insisde the matrices in the upper right of the matrxi we have the expectation of $x_1$ times $x_2$, like $\mathbb{E}(x_1 \cdot x_2)$, where the extactation in the gaussian case is the mean which is 0, so the corresponding values is 0. -->
<!-- the covariance essentially is the dot product [ref dot product](https://mathinsight.org/dot_product_matrix_notation) of $x_1$ and $x_2$ variable, so what happens when you take the dot product of vectors, -->
<!-- if for example you take a vector that looks like 1 and 0 and you take the dot product of one other vecto 1 and 0, so that: -->
<!-- $$ -->
<!-- \left[\begin{array}{ll} -->
<!-- 1 \\ -->
<!-- 0 -->
<!-- \end{array}\right]\left[\begin{array}{ll} -->
<!-- 1 & 0 \\ -->
<!-- \end{array}\right] = 1 -->
<!-- $$ -->
<!-- You will end up with 1, recall dot productm first element first vcetor times first element second vectro and second element first vector times second element secon vector. So identical vector will get a high dot product value leading to a high similarity measure. Dot product can be indeded as a similarity measure. -->
<!-- ... But if you take two different vector as 1 0 and 0 1 then: -->
<!-- $$ -->
<!-- \left[\begin{array}{ll} -->
<!-- 1 \\ -->
<!-- 0 -->
<!-- \end{array}\right]\left[\begin{array}{ll} -->
<!-- 0 & 1 \\ -->
<!-- \end{array}\right] = 0 -->
<!-- $$ -->
<!-- This time the multiplication leads to 0 value, as a matter of fact they are different. They are no similar. -->
<!-- IF two points are closed the dot product will be high in 2D. What the covariance should be? if variances are assumed to be 1 then in this case i qould expect to be 0, i.e. covariance matrix is: -->
<!-- $$ -->
<!-- \left[\begin{array}{ll} -->
<!-- 1 & 0 \\ -->
<!-- 0 & 1 -->
<!-- \end{array}\right] = \mathbf{cov_{plot1}}(x_1,x_2) -->
<!-- $$ -->
<!-- because I can picka poin tin two pointa in this cloud. Suppose i increase x1 then my chance of getting a x2 point that is positive or negative is the samee, knowing somthin about x1 give nothign about x2. no information is proivede. On the other hand i the second plot knowing a positive value of 1 can suggest with a certain probability that x2 will be positive (great proabibility. So some information is provided), e.g. -->
<!-- $$ -->
<!-- \left[\begin{array}{ll} -->
<!-- 1 & 0.5 \\ -->
<!-- 0.5 & 1 -->
<!-- \end{array}\right] = \mathbf{cov_{plot2}}(x_1,x_2) -->
<!-- $$ -->
<!-- Some positive number idicates that i expect a positive inc rease iwhen boht of the two are increasing singularly. thsi is what the correlation, the basis to do linear regresssion and non linear- thei is a bivariate gaussian. If the entri3es are means that they are uncoorellated, if they are non-zero then they are correlated, theby can be both positive or negative (correlatiob) -->
<!-- now lets generate a gassiian distrivution so x_1 and x_2 in 2D and then a third dimension hwere we express probability, this is said joint distribution. So i am going to cu this gaussian at certain point for x_1 and cut a plain rigght thgouth this gauissan imagine to ahava cake and then taka kkniw and cut it.(see the image) -->
<!-- form the man perspective you are goin to see a gaussian distribution, you will be lookong at x_ and you will be seeing a gaussain plot in green. this is the probability of x_1 gievn x_1. also said "conditioned" probabolity. This gaussian has a mean like the one alreasdy seen and this is the center of the gaussian, we can rewrite the mena and variance of the multivariate gaussian describing the cloud of points. sigma are the covaraince martix sigma. -->
<!-- ... -->
<!-- sigma 1 and signa 2 if you have 1 d varibale the widjth has to be postive, for mulitvariate gaussian equl so here positive definitness: covariance mateix symetric. -->
<!-- ... -->
<!-- any artibitray variable transposed x time the covarince matrix nedds to be positive. -->
<!-- what is the mean of this gaussian i might want to know what is the widht of this gaussian would it be great if there is a formula that guven the cloud of point and likelihood estimation. we coilf obtain the red bell in figure. -->
<!-- Compute the green curve how it is done? this requires some work and it is said matrix *invesrion lemma*, this is foudamental for machine learnign. let's assume it. The theorem says that the the mean fof the gaussian is the mean of x_1 and then some other operation with sigma, see below from paci (miss ref) -->
<!-- ![inverse lemma, Paci source](appendix_images/inverse_lemma.jpg) -->
<!-- the theorem says toi ocnsider a multivariate gaussian a vector 1 and a vector 2 each vecto compinent has a mena and a covarianc matrix, this by lemma gives us the expression and the math behind is no tremendous, but it is long. What it is important is to undestand fto go from a joont to a conditional distribution in our case. thats i the value od the theorem. -->
<!-- One background further thing: assume that we have a gaussian variable distribution that we want to sample fromm, we had now ewe are going to do the opposite, before we had poitns and we tried to figure out the curve, now we have the curve and we are gointg to try to rpoduce data. I need to be able to draw sample froma gaussian distribution. i will assume that i have a meachnism that produces a uniform samples, so you have a random numebr generatior with equal probabolity from 0 to 1, I assume a also the cumulative aof a gaussian. -->
<!-- the cumulative of a gaussian is what you get if you syrta summing the area under the curve of the gaussian as you move from the left. value after valure you can plot the cumulative ahead (see figure) the point where there is a flex point is the mean beacuse tha gassias is symmetric. The asymptot is 1 becuase the are under the curve sumes to 1. -->
<!-- If i can draw a random number form Uniform and the project it to thre cumulative and then finally projct it back to the gaussian distribution. Inverse cumulative mapping. If oyu do this multiple times you are going to have many sample palced next to the mean and as sparse as the variance. in this process of sampling try to sample a point i froma gaussian that has mean 0 a variance 1, now letes try to draw a point from a gaussian with mean mu and variance sigma. ... -->
<!-- In the multivariate case suppose that we have evctor with two variables how do i draw a vector from a multivariate gasussian with 0 means and plot 1 covarianc ematrix. the theormeem also says that the marginal distribution can be seen by civariance matrix , fist take the men_1 and take upper left element from the covariance matrix obtaining the marginal rpobabiloty for x_1, i.e. -->
<!-- <!-- $$ -->
<!-- <!-- \pi(x_1) = \mathbf{N}(\mu_1, \Sigma_{11}) \\ -->
<!-- <!-- \pi(x_2) = \mathbf{N}(\mu_2, \Sigma_{22}) -->
<!-- <!-- $$ -->
<!-- <!-- then in our problem: -->
<!-- <!-- $$ -->
<!-- <!-- \pi(x_1) = \mathbf{N}(0, 1) \\ -->
<!-- <!-- \pi(x_2) = \mathbf{N}(0, 1) -->
<!-- <!-- $$ -->
<!-- Then for simplicity we can simplyfy by groupign vector into: -->
<!-- (vectore exoression multivariate) -->
<!-- I need a wau to take square trotto of matrices, if x come sfroma MVG -->
<!-- 35:01-- -->
## SPDE and Triangulation{#triangular}
In order to fit LGM type of models with a spatial component INLA uses SPDE (Stochastic Partial Differential Equations) approach.
Suppose that is it given have a continuous Gaussian random process (a general continuous surface), what SPDE does is to approximate the continuous process by a discrete Gaussian process using a triangulation of the region of the study. Let assume to have a set of points defined by a CRS system (Latitude and Longitude, Easthings and Northings), and let assume that the object of the analysis is to estimate a spatially continuous process. Instead of exploiting this property as a whole, it is estimated the process only at the vertices of this triangulation. This requires to put a Triangulation of the region of the study on top the the process and the software will predict the value of the process at its vertices, then it interpolates the rest of the points obtaining a "scattered" surface.
![(#fig:triang)Triangulariation weights and associated process value, @YT:paumoraga source](appendix_images/appendix_triangularization.jpg)
Imagine to have a point a location X laying inside a triangle whose vertices are $S_1, S_2 and S_3$. SPDE operates by setting the values of the process at location x equal to the value of the process at their vertices with some weights, and the weights are given by the _Baricentric coordinates (BC)_. BC are proportional to the area at the point and the vertices. Let assume to have a piece of Triangulation $\boldsymbol{A}$ and let assume that the goal is to compute the value at location X. X, as in formula above $\boldsymbol{A}$, would be equal to $S_1$, multiplied by the area $A_1$ dived by the whole triangle area ($\boldsymbol{A}$) + $S_2$ multiplied by the area $A_2$ divided by $\boldsymbol{A}$ + $S_3$, multiplied by the area $A_3$ dived by ($\boldsymbol{A}$. This would be the value of the process at location X given the triangulations (number of vertices). SPDE is actually approximating the value of the process using a weighted average of the value of the process at the triangle vertices which ir proportional to the area of the below triangle.
In order to do this within INLA \@ref(inla) it is needed also a _Projection Matrix_ , figure \@ref(fig:projmat). The Projection matrix maps the continuous GRMF (when it is assumed a GP) from the observation to the triangulation. It essentially assigns the height of the triangle for each vertex of the Triangulation to the process. Matrix $\mathcal{A}$, whose dimensions are $\mathcal{A_{ng}}$. It has $n$ rows a $g$ columns, where $n$ is the number of observations and $g$ is the number of vertices of the triangulation. Each row has possibly three non-0 values, right matrix in figure \@ref(fig:projmat), and the columns represent the vertices of the triangles that contains the point. Assume to have an observation that coincides with a vertex $S_1$ of the triangle in \@ref(fig:triang), since the point is on top of the vertex (not inside), there are no weights ($A_1 = \mathcal{A}$) and 1 would be the value at $A_{(1,1)}$ and 0 would be the rest f the values in the row. Now let assume to have an observation coinciding with $S_3$ (vertex in position 3), then the result for $A_{(2,3)}$ would be 1 and the rest 0. Indeed when tha value is X that lies within one of the triangles all the elements of the rows will be 0, but three elements in the row corresponding of the p osition of the vertices $1 = .2, 2 = .2 and g = .6$, as a result X will be weighted down for the areas.
![(#fig:projmat)Projection Matrix to map values from tringulation back to the GP, @YT:paumoraga surce](appendix_images/appendix_proj.jpg)
## Laplace Approximation{#laplaceapprox}
@Blangiardo-Cameletti offers an INLA focused intuiton on how the Laplace approximation works for integrals. Let assume that the interest is to compute the follwing integral, assuming the notation followed throughout the analysis:
$$
\int \pi(x) \mathrm{d} x=\int \exp (\log f(x)) \mathrm{d} x
$$
Where $X$ is a random variable for which it is specified a distribution function $\pi$. Then by the Taylor series expansions [@taylorseries] it is possible to represent the $\log \pi(x)$ evaluated in an exact point $x = x_0$, so that:
\begin{equation}
\log \pi(x) \approx \log \pi\left(x_{0}\right)+\left.\left(x-x_{0}\right) \frac{\partial \log \pi(x)}{\partial x}\right|_{x=x_{0}}+\left.\frac{\left(x-x_{0}\right)^{2}}{2} \frac{\partial^{2} \log \pi(x)}{\partial x^{2}}\right|_{x=x_{0}}
(\#eq:secondordexpan)
\end{equation}
Then if it is assumed that $x_0$ is set equal to the mode $x_*$ of the distribution (the highest point), for which $x_{*}=\operatorname{argmax}_{x} \log \pi(x)$, then the first order derivative with respect to $x$ is 0, i.e. $\left.\frac{\partial \log f(x)}{\partial x}\right|_{x=x_{*}}=0$. That comes natural since once the function reaches its peak, i.e. the max then the derivative in that point is 0. Then by leaving out the first derivative element in eq. \@ref(eq:secondordexpan) it is obtained:
$$
\log \pi(x) \approx \log \pi\left(x_{*}\right)+\left.\frac{\left(x-x_{*}\right)^{2}}{2} \frac{\partial^{2} \log \pi(x)}{\partial x^{2}}\right|_{x=x_{*}}
$$
Then by integrating what remained, exponantiating and taking out non integrable terms,
<!-- $$ -->
<!-- \int \pi(x) \mathrm{d} x \approx \int \exp \left(\log \pi\left(x^{*}\right)+\left.\frac{\left(x-x^{*}\right)^{2}}{2} \frac{\partial^{2} \log \pi(x)}{\partial x^{2}}\right|_{x=x^{*}}\right) \mathrm{d} x -->
<!-- $$ -->
\begin{equation}
\int \pi(x) \mathrm{d} x \approx \exp \left(\log \pi\left(x_{*}\right)\right) \int \exp \left(\left.\frac{\left(x-x_{*}\right)^{2}}{2} \frac{\partial^{2} \log \pi(x)}{\partial x^{2}}\right|_{x=x_{*}}\right) \mathrm{d} x
(\#eq:integraltay)
\end{equation}
At this point it might be already intuited the expression \@ref(eq:integraltay) is actually the density of a Normal. As a matter of fact, by imposing $\sigma^{2}_{*}=-1 /\left.\frac{\partial^{2} \log f(x)}{\partial x^{2}}\right|_{x=x_{*}}$, then expression \@ref(eq:integraltay) can be rewritten as:
\begin{equation}
\int \pi(x) \mathrm{d} x \approx \exp \left(\log \pi\left(x_{*}\right)\right) \int \exp \left(-\frac{\left(x-x_{*}\right)^{2}}{2 \sigma^{2}_{*}}\right) \mathrm{d} x
(\#eq:rearrangeintr)
\end{equation}
Furthermore the integrand is the _Kernel_ of the Normal distribution having mean equal to the mode $x_*$ and variance specified as $\sigma^{2}_{*}$. By computing the finite integral of \@ref(eq:rearrangeintr) on the closed neighbor $[ \alpha, \beta]$ the approximation becomes:
$$
\int_{\alpha}^{\beta} f(x) \mathrm{d} x \approx \pi\left(x_{*}\right) \sqrt{2 \pi \sigma^{2^{*}}}(\Phi(\beta)-\Phi(\alpha))
$$
where $\Phi(\cdot)$ is the cumulative distribution function corresponding value of the Normal disttribution in eq. \@ref(eq:rearrangeintr).
<!-- further example: let assume to have a Beta distribution $\operatorname{Beta}(\alpha = 2,8)$ whose parameter are $\alpha$ and $\beta$ and whose density is: -->
<!-- FORSE L'ESEMPIO -->
For example consider the Chi-squared $\chi^{2}$ density function (since it is easily differentiable and Gamma $\Gamma$ term in denominator is constant). The following quantities of interest are the $\log \pi(x)$, then the $\frac{\partial \log \pi(x)}{\partial x}$, which has to be set equal to 0 to find the integrating point, and finally $\log \pi(x)$ $\frac{\partial^2 \log \pi(x)}{\partial x^2}$. The $\chi^{2}$ pdf, whose support is $x \in \mathbb{R}^{+}$ and whose degrees of freedom are $k$:
$$
\chi^{2}(x,k)=\frac{x^{(k / 2-1)} e^{-x / 2}}{2^{k / 2} \,\, \Gamma\left(\frac{k}{2}\right)}, x \geq 0
$$
for which are computed:
$$
\log f(x)=(k / 2-1) \log x-x / 2
$$
Ths single variable _Score Function_ and the $x_{*}=\operatorname{argmax}_{x} \log \pi(x)$,
\begin{equation}
\begin{aligned}
&\frac{\partial \log \pi(x)}{\partial x}= \frac{(k/2 -1)}{x} - \frac{1}{2} =0 \quad \text {solving for 0 }\\
&(k/2-1) = x/2\quad \text {moving addends}\\
&x_* = (k-2)
\end{aligned}
\end{equation}
And the _Fisher Information_ in the $x_*$ (for which it is known $\sigma^{2}_*=-1 / f^{\prime \prime}(x)$)
\begin{equation}
\begin{split}
&\frac{\partial^2 \log \pi(x)}{\partial x^2}=-\frac{(k / 2-1) }{x^{2}} \quad \text {substituting} \,\,x_* \\
&= -\frac{(k / 2-1) }{(k-2)^{2}} = - \frac{2(k/2-1)}{2(k-2)^{2}} \quad \text {multiply by 2 } \\
&=- \frac{(k-2)}{2(k-2)^{2}} = 2(k-2) = \sigma^{2}_*\quad \text {change of sign and inverse }
\end{split}
\end{equation}
finally,
$$
\chi^{2} \stackrel{L A}{\sim} N\left(x_*=k-2, \sigma^{2}_*=2(k-2)\right)
$$
Assuming $k = 8, 16$ degrees of freedom $\chi^{2}$ densities against their Laplace approximation, the figure \@ref(fig:laplacevisual) displays how the approximations fits the real density. Integrals are computed in the case of $k = 8$ in the interval $(5, 7)$, leading to a very good Normal approximation that slightly differ from the orginal CHisquared. The same has been done for the $k = 16$ case, whose interval is $(12, 17)$ showing other very good approximations. Note that the more are the degrees of freedom the more the chi-squared approximates the Normal leading to better approximations.
```{r laplacevisual, fig.cap="Chisquared density function with parameter $k = 8$ (top) and $k = 16$ (down) solid line. The point line refers to the corresponding Normal approximation obtained using the Laplace method"}
set.seed(2021)
df1 <- 8
df2 <- 16
x_star1 <- df1 - 2
x_star2 <- df2 - 2
sigma2_star1 <- 2 * (df1 - 2)
sigma2_star2 <- 2 * (df2 - 2)
## integral extremes
alpha1 <- 5
beta1 <- 7
alpha2 <- 12
beta2 <- 17
dat <- tibble(
x = seq(0, 30, l = 100)
)
chisq8 <- dat %>%
ggplot(aes(x = x)) +
stat_function(fun = dchisq, args = list(df = df1)) +
stat_function(fun = dnorm, args = list(mean = x_star1, sd = sqrt(sigma2_star1)), geom = "point") +
xlim(0, 30) +
xlab(TeX("$k \\,=\\, 8$")) +
ylab("Density") +
geom_vline(xintercept = alpha1, alpha = 0.6, linetype = "dotdash", color = "red") +
geom_vline(xintercept = beta1, alpha = 0.6, linetype = "dotdash", color = "red") +
annotate("text", x = 20, y = 0.05, label = TeX("$\\int_{5}^{7} \\chi^{2}(\\, 8\\,) \\mathrm{d} x\\, = \\, 0.2209435 $"), size = 4) +
annotate("text", x = 20, y = 0.09, label = TeX("$\\int_{5}^{7} \\mathit{N}(\\, 6,\\, 12) \\mathrm{d} x\\, = \\, 0.2209686 $"), size = 4) +
theme_nicco()
chisq16 <- dat %>%
ggplot(aes(x = x)) +
stat_function(fun = dchisq, args = list(df = df2)) +
stat_function(fun = dnorm, args = list(mean = x_star2, sd = sqrt(sigma2_star2)), geom = "point") +
xlim(0, 50) +
xlab(TeX("$k \\,=\\, 16$")) +
geom_vline(xintercept = alpha2, alpha = 0.6, linetype = "dotdash", color = "red") +
geom_vline(xintercept = beta2, alpha = 0.6, linetype = "dotdash", color = "red") +
annotate("text", x = 35, y = 0.03, label = TeX("$\\int_{12}^{17} \\chi^{2}(\\, 16\\,) \\mathrm{d} x\\, = \\, 0.3583827 $"), size = 4) +
annotate("text", x = 35, y = 0.062, label = TeX("$\\int_{12}^{17} \\mathit{N}(\\, 14,\\, 28) \\mathrm{d} x\\, = \\, 0.3576164 $"), size = 4) +
ylab("Density") +
theme_nicco()
(chisq8 + chisq16)
#
## integral extremes
# alpha = 10
# beta = 12
#
# ## exact values
# pchisq(beta1, df = df1) -pchisq(alpha1, df = df1)
# pchisq(beta2, df = df2) -pchisq(alpha2, df = df2)
# ## approx values
# dchisq(x_star1, df = df1)*
# sqrt(2*pi*sigma2_star1)*
# (pnorm(beta1,mean=x_star1,sd=sqrt(sigma2_star1)) -
# pnorm(alpha1,mean=x_star1,sd=sqrt(sigma2_star1)))
#
# dchisq(x_star2, df = df2)*
# sqrt(2*pi*sigma2_star2)*
# (pnorm(beta2,mean=x_star2,sd=sqrt(sigma2_star2)) -
# pnorm(alpha2,mean=x_star2,sd=sqrt(sigma2_star2)))
#
# xx=seq(0,20,l=100)
# plot(xx,dchisq(xx, df = df1),t="l",ylab="Density",xlab="x")
# par(new=TRUE)
# plot(xx,dnorm(xx,mean=x_star1,sd=sqrt(sigma2_star1)),lty=2)
```
$$
\pi(\boldsymbol{\psi}) \quad \boldsymbol{\psi}= \boldsymbol{\psi_1}, \boldsymbol{\psi_2}
$$