The package varycoef
contains methods to model and estimate varying coefficients. In its current version 0.2.9 it supports:
only spatially varying coefficient (SVC)
different MLE approaches to model SVC and to give predictions.
We define a full SVC model as
\[y(s) = x^{(1)}(s)\beta_1(s) + ... + x^{(p)}(s)\beta_p(s) + \epsilon(s)\]
with the coefficients represented by Gaussian random fields (GRF) \(\beta_j(\cdot) \sim \mathcal N (\mu_j \textbf 1_n, C_j(\cdot, \cdot))\). That is, every coefficient \(j = 1, ..., p\) is distinctly defined by a mean \(\mu_j\) and a covariance matrix defined by an underlying covariance function \(C_j(s_k, s_l) = \sigma_j^2 \phi_{\rho_j}(s_k, s_l)\), where \(\sigma_j^2\) is the variance and \(\rho_j\) is the scale of the GRF. Further, \(\epsilon\) is a nugget effect with variance \(\tau^2\).
However, there are some cases, where the assumption of a full SVC model is not applicable. We want to give options for covariates w.r.t.:
That is why we generalize the model above. First, note that we can write it in matrix form as
\[\textbf y(\textbf s) = \textbf X \beta(\textbf s) + \epsilon( \textbf s)\]
where \(\textbf X\) is our model matrix. Then we can write down the model divided into an fixed effect part and a random effect part:
\[\textbf y(\textbf s) = \textbf X \mu + \textbf W \eta(\textbf s) + \epsilon( \textbf s)\]
where \(\eta\) is the joint mean-zero GRF. Note that both model are the same if \(\textbf X = \textbf W\). Thus, we can specify options 1 to 3 from above by in- or excluding columns of \(\textbf X\) or \(\textbf W\), respectively.
To give a simple example, we start by sampling artificial data. So define an full SVC model as given above and sample from a regular grid using a help function:
# number of SVC
p <- 3
(pars <- data.frame(mu = rep(0, p),
var = c(0.1, 0.2, 0.3),
scale = c(0.3, 0.1, 0.2)))
## mu var scale
## 1 0 0.1 0.3
## 2 0 0.2 0.1
## 3 0 0.3 0.2
nugget.var <- 0.05
# sqrt of total number of observations
m <- 20
# function to sample SVCs
sp.SVC <- fullSVC_reggrid(m = m, p = p,
cov_pars = pars,
nugget = nugget.var)
library(sp)
# visualization
spplot(sp.SVC, colorkey = TRUE)
We further need some covariates which we sample from a standard normal. In order to model an intercept, we set \(x^{(1)} = 1\):
## [,1] [,2] [,3]
## [1,] 1 -0.2890233 -0.5116037
## [2,] 1 0.6565134 0.2369379
## [3,] 1 -0.4539977 -0.5415892
## [4,] 1 -0.5938646 1.2192276
## [5,] 1 -1.7103797 0.1741359
## [6,] 1 -0.2094484 -0.6152683
We compute the response \(y\):
varycoef
The main function of this package is SVC_mle
. Its function call starts the MLE but it requires some preparation and settings of control parameters. We go through each argument of the SVC_mle
function and its control parameter SVC_mle.control
.
SVC_mle
As one might see in the help file of the SVC_mle
function, it has 3 mandatory arguments: y
, the response; X
, the data matrix and locs
, the locations. If we do not change W
, i.e. W = NULL
, then we use W = X
and are in the case of a full SVC. We will give examples for different kinds of models.
As for the control parameters for SVC_mle
, we go through them as they are implemented in the current version of varycoef
. By calling SVC_mle_control
, we create an list with all needed arguments to start a simple SVC_mle
.
## List of 10
## $ cov.name : chr "exp"
## $ tapering : NULL
## $ cl : NULL
## $ init : NULL
## $ lower : NULL
## $ upper : NULL
## $ save.fitted: logi TRUE
## $ profileLik : logi FALSE
## $ mean.est : chr "GLS"
## $ pc.prior : NULL
Here we define the covariance function \(C_j(s_k, s_l) = \sigma_j^2 \phi_{\rho_j}(s_k, s_l)\). In its current version 0.2.9, varycoef
supports only exponential covariance functions, i.e. \(\phi_{\rho}(s_k, s_l) = \exp\left(\frac{\|s_k - s_l\|}{\rho}\right)\).
Covariance tapering goes back to (Furrer, Genton, and Nychka 2006) and is a technique to challenge the “big \(n\) problem” when dealing with spatial data. When working with \(n\) observations, the covariance matrix has dimension \(n \times n\). The likelihood function of a single GRF or, in particular, the SVC models as described above, includes the inverse as well as the determinant of a covariance matrix. The computation of both is based on the Cholesky decomposition which has run-time \(\mathcal O(n^3)\). There are several ways on how to deal with this computational burden.
With covariance tapering, we introduce a sparsity structure on the covariance matrix. This sparsity structure can then be used to our advantage with the Cholesky decomposition implemented in the package spam
, (Furrer and Sain 2010). In a nutshell, this decomposition becomes faster as the sparsity increases. However, this comes with a trade-off in the precision of the MLE of the covariance parameters with finitely many observation. Asymptotically, the procedure is consistent.
By default, the tapering
entry is NULL
, i.e. no tapering is applied. If a scalar is provided, then this is taper range is applies. In other words, every spatial dependency is cut for distances larger than tapering
. We illustrate the difference between both the untapered and tapered covariance matrix of the SVC on the regular grid example from above. The function fullSVC_reggrid
is used to sample SVCs for a full SVC model.
Finally, we show the time differences of evaluating the likelihood function between the different taper ranges. We use the microbenchmark
package:
library(microbenchmark)
(mb <- microbenchmark(no_tapering = out[[1]][[4]](),
tapering_0.5 = out[[2]][[4]](),
tapering_0.3 = out[[3]][[4]](),
tapering_0.1 = out[[4]][[4]]())
)
## Unit: milliseconds
## expr min lq mean median uq max neval
## no_tapering 55.63454 60.04109 71.44503 60.40445 63.06325 172.5575 100
## tapering_0.5 54.01774 56.48711 62.46798 58.41941 58.94872 143.2631 100
## tapering_0.3 35.81527 36.23613 43.73090 38.18897 40.20032 122.3703 100
## tapering_0.1 21.25382 23.55916 27.89986 24.32287 25.97414 108.4425 100
## cld
## d
## c
## b
## a
To be implemented.
To run an MLE, we need initial values for all variances and the scales as well as the means. A good starting point for the means are the estimated coefficients of an OLS, i.e.:
## X.1 X.2 X.3
## 0.04638084 0.02147242 0.17528015
For the variance, we take the estimated variance of the OLS method, i.e
## [1] 0.3999115
For the scale, we remind ourselves that the effective range of an exponential GRF is approximately 3 times its range. Since in our case we are in an restricted domain of the unit square, any dependency structure on a distance above the maximum distance in the domain, we cannot model. Therefore we suggest to take an initial scale of
The vector of initial values is then:
init <- c(
# GRFs scales and variances
rep(c(scale.init, var.init), p),
# nugget variance
var.init,
# means
mu.init)
One can either overwrite the default initial value (control$init
) or call SVC_mle_control
with corresponding arguments:
## NULL
The default values gives initial values of \(0.3\) for the variances and ranges and \(0\) for the means, which might not be the best option in general.
The optimization function optim
allows with its chosen method "L-BFGS-B"
the usage of lower and upper bounds. The lower and upper bounds for each parameter can be given in correspondence to the order in the initial values.
The argument save.fitted
takes logical values and decides whether the fitted values of the SVC model after the MLE should be computed. This takes some extra time, but can be useful since an extraction of those fitted values as well as the residuals is possible using a methods fitted
or residuals
. In case of allocation errors, set to FALSE
.
## [1] TRUE
In some SVC models, it might be an advantage to optimize not over all parameters, i.e. the mean and the covariance parameters. The motivation here is a dimension reduction of the parameter space. Setting profileLik
to TRUE
, the optimization will be done using the profile likelihood of the covariance parameters. The mean effects are then implicitly calculated using the generalized least squares estimate.
## [1] FALSE
Attention: Since the number of parameters is reduced, one should also be aware that the initial values as well as the lower and upper bounds are given correctly.
We implemented a usage of penalizing complexity priors. The argument pc.prior
takes a vector of length 4 as an input where the values are \(\rho_0, \alpha_\rho, \sigma_0, \alpha_\sigma\) to compute penalized complexity priors. One wants to penalize the lower tail on the range parameter as well as the upper tail of the standard deviation:
\[P(\rho < \rho_0) = \alpha_\rho, \quad P(\sigma > \sigma_0) = \alpha_\sigma.\]
We can now start the MLE. The following function call takes a few second.
The received object fit
is of class SVC_mle. For this class, there are numerous methods such as:
## SVC1.range SVC1.var SVC2.range SVC2.var SVC3.range SVC3.var
## 0.12198444 0.04901733 0.07933388 0.15372652 0.07486074 0.18419362
## nugget.var
## 0.03174272
## attr(,"GRF")
## [1] "exp"
## Var1 Var2 Var3
## 0.04752166 -0.01587562 0.18215334
##
## Call:
## SVC_mle with 3 fixed effect(s) and 3 SVC(s)
## using 400 observations at 400 different locations.
##
## Residuals:
## Min. 1st Qu. Median 3rd Qu. Max.
## -0.195059 -0.047189 -0.002415 0.047866 0.255630
##
## Residual standard error (nugget effect): 0.1782
## Multiple R-squared: 0.9862
##
##
## Coefficients of fixed effects:
## Var1 Var2 Var3
## 0.04752 -0.01588 0.18215
##
##
## Covariance parameters of the SVC(s):
## range variance
## SVC1 0.12198 0.04902
## SVC2 0.07933 0.15373
## SVC3 0.07486 0.18419
##
## The covariance parameters were estimated for the GRFs using
## exp. covariance functions. No covariance tapering applied.
##
##
## MLE:
## The MLE terminated after 86 function evaluations with convergence code 0
## (0 meaning that the optimization was succesful).
## The final Likelihood value for 10 parameters is -290.3.
Now, we can use our fit
object to make predictions:
# calling predictions without specifying new locations (newlocs) or
# new covariates (newX) gives estimates of SVC only at the training locations.
pred.SVC <- predict(fit)
Since we know the true SVC, we can compute the error in prediction and compare it to the true values.
colnames(pred.SVC)[1:p] <- paste0("pred.",colnames(pred.SVC)[1:p])
coordinates(pred.SVC) <- ~loc_x+loc_y
all.SVC <- cbind(pred.SVC, sp.SVC[, 1:3])
# compute errors
all.SVC$err.SVC_1 <- all.SVC$pred.SVC_1 - all.SVC$SVC_1
all.SVC$err.SVC_2 <- all.SVC$pred.SVC_2 - all.SVC$SVC_2
all.SVC$err.SVC_3 <- all.SVC$pred.SVC_3 - all.SVC$SVC_3
colnames(all.SVC@data) <- paste0(rep(c("pred.", "true.", "err."), each = p), "SVC_", rep(1:p, 3))
spplot(all.SVC[, paste0(rep(c("true.", "err.", "pred."), each = p),
"SVC_", 1:p)], colorkey = TRUE)
In this small example we already can see that the predicted SVC takes the general spatial structure of the true SVC. The error does not appear to have spatial structure for the SVC 2 and 3, respectively. However, the error for the intercept seems to have some spatial structure. If we increase the number of observations, the picture changes:
We do not run the code since it takes a couple hours to do the MLE, but here is the code to reproduce the figure:
# new m
m <- 50
# new SVC model
sp.SVC <- fullSVC_reggrid(m = m, p = p,
cov_pars = pars,
nugget = nugget.var)
spplot(sp.SVC, colorkey = TRUE)
# total number of observations
n <- m^2
X <- matrix(c(rep(1, n), rnorm((p-1)*n)), ncol = p)
y <- apply(X * as.matrix(sp.SVC@data[, 1:p]), 1, sum) + sp.SVC@data[, p+1]
fit <- SVC_mle(y = y, X = X, locs = coordinates(sp.SVC))
sp2500 <- predict(fit)
colnames(sp2500)[1:p] <- paste0("pred.",colnames(sp2500)[1:p])
coordinates(sp2500) <- ~loc_x+loc_y
all.SVC <- cbind(sp2500, sp.SVC[, 1:3])
# compute errors
all.SVC$err.SVC_1 <- all.SVC$pred.SVC_1 - all.SVC$SVC_1
all.SVC$err.SVC_2 <- all.SVC$pred.SVC_2 - all.SVC$SVC_2
all.SVC$err.SVC_3 <- all.SVC$pred.SVC_3 - all.SVC$SVC_3
colnames(all.SVC@data) <- paste0(rep(c("pred.", "true.", "err."), each = p), "SVC_", rep(1:p, 3))
png(filename = "figures/SVCs_result_n2500_p3.png", width = 960, height = 960)
spplot(all.SVC[, paste0(rep(c("true.", "err.", "pred."), each = p),
"SVC_", 1:p)], colorkey = TRUE,
as.table = TRUE, layout = c(3, 3))
dev.off()
In some cases, the assumption of a full SVC model might be wrong. Therefore, one might want to restrict the covariables to vary in some, but not all cases. In this case \(\textbf X \neq \textbf W\), hence we need to specify the W
argument of SVC_mle
independently of X
. We now have to differentiate between X
and W
, which is why we introduce pX
and pW
as the number of covariates for X
and W
, respectively.
# new m
m <- 20
# number of fixed effects
pX <- 4
# number of mean-zero SVC
pW <- 3
# mean values
mu <- 1:pX
# new SVC model
sp.SVC <- fullSVC_reggrid(m = m, p = pW,
cov_pars = pars,
nugget = nugget.var,
seed = 4)
# total number of observations
n <- m^2
X <- matrix(c(rep(1, n), rnorm((pX-1)*n,
mean = rep(1:(pX-1), each = n),
sd = rep(0.2*(1:(pX-1)), each = n))), ncol = pX)
W <- X[, 1:pW]
# calculate y
y <-
# X * mu
X %*% mu +
# W * beta_tilde
apply(W * as.matrix(sp.SVC@data[, 1:pW]), 1, sum) +
# nugget
sp.SVC@data[, pW+1]
# new initial values
control <- SVC_mle_control(init = c(init[1:(2*pW + 1)], mu))
# MLE
fit <- SVC_mle(y = y, X = X, W = W, locs = coordinates(sp.SVC), control = control)
As above, one can then use the SVC_mle
-class object to get predictions using a method of the predict
function. Depending on the available information, there are 2 types of preditions one can receive. First, just specifying the new locations using newlocs
will result in predictions for each individual SVC at corresponding locations. Additionally, when providing the corresponding data matrices newX
and newW
one gets the estiamted response variable y
. Further, setting the argument compute.y.var
to TRUE
will give the predictive variances to the predictions of y
.
set.seed(5)
newlocs <- matrix(c(0, 0), ncol = 2)
# only SVC prediction without y prediction
(pred.SVC <- predict(fit, newlocs = newlocs))
## SVC_1 SVC_2 SVC_3 loc_x loc_y
## 1 1.206589e-06 -0.3000649 -0.2761391 0 0
# SVC prediction and y prediction (with predictive variance)
newX <- matrix(c(1, rnorm(pX-1)), ncol = pX)
newW <- matrix(newX[, 1:pW], ncol = pW)
(pred.SVC <- predict(fit,
newlocs = newlocs,
newX = newX, newW = newW,
compute.y.var = TRUE))
## SVC_1 SVC_2 SVC_3 y.pred y.var loc_x loc_y
## 1 1.206589e-06 -0.3000649 -0.2761391 -1.82877 0.4142608 0 0
Furrer, Reinhard, Marc G Genton, and Douglas Nychka. 2006. “Covariance Tapering for Interpolation of Large Spatial Datasets.” Journal of Computational and Graphical Statistics 15 (3): 502–23. https://doi.org/10.1198/106186006X132178.
Furrer, Reinhard, and Stephan R. Sain. 2010. “spam
: A Sparse Matrix R Package with Emphasis on MCMC Methods for Gaussian Markov Random Fields.” Journal of Statistical Software 36 (10): 1–25. http://www.jstatsoft.org/v36/i10/.