Troubleshooting with glmmTMB


This vignette covers common problems that occur while using glmmTMB. The contents will expand with experience.

If your problem is not covered below, there’s a chance it has been solved in the development version; try updating to the latest version of glmmTMB on GitHub.


Model convergence problem; non-positive-definite Hessian matrix; NA values for likelihood/AIC/etc.

This warning (Model convergence problem; non-positive-definite Hessian matrix) states that at glmmTMB’s maximum-likelihood estimate, the curvature of the negative log-likelihood surface is inconsistent with glmmTMB really having found the best fit (minimum): instead, the surface is downward-curving, or flat, in some direction(s).

It will usually be accompanied by NA values for the standard errors, log-likelihood, AIC, and BIC, and deviance. When you run summary() on the resulting model, you’ll get the warning In sqrt(diag(vcov)) : NaNs produced.

These problems are most likely:

How do we diagnose the problem?

Example 1.

Consider this example:

zinbm0 = glmmTMB(count~spp + (1|site), zi=~spp, Salamanders, family=nbinom2)

First, see if any of the estimated coefficients are extreme. If you’re using a non-identity link function (e.g. log, logit), then parameter values with \(|\beta|>10\) are suspect (for a logit link, this implies probabilities very close to 0 or 1; for a log link, this implies mean counts that are close to 0 or extremely large).

Inspecting the fixed-effect estimates for this model:


The zero-inflation intercept parameter is tiny (\(\approx -17\)): since the parameters are estimated on the logit scale, we back-transform with plogis(-17) to see the at the zero-inflation probability for the baseline level is about \(4 \times 10^{-8}\))). Many of the other ZI parameters are very large, compensating for the intercept: the estimated zero-inflation probabilities for all species are

ff <- fixef(zinbm0)$zi

Since the baseline probability is already effectively zero, making the intercept parameter larger or smaller will have very little effect - the likelihood is flat, which leads to the non-positive-definite warning.

Now that we suspect the problem is in the zero-inflation component, we can try to come up with ways of simplifying the model: for example, we could use a model that compared the first species (“GP”) to the rest:

Salamanders <- transform(Salamanders, GP=as.numeric(spp=="GP"))
zinbm0_A = update(zinbm0, ziformula=~GP)

This fits without a warning, although the GP zero-inflation parameter is still extreme:


Another possibility would be to fit the variation among species in the zero-inflation parameter as a random effect, rather than a fixed effect: this is slightly more parsimonious. This again fits without an error, although both the average level of zero-inflation and the among-species variation are estimated as very small:

zinbm0_B = update(zinbm0, ziformula=~(1|spp))

The original analysis considered variation in zero-inflation by site status (mined or not mined) rather than by species - this simpler model only tries to estimate two parameters (mined + difference between mined and no-mining) rather than 7 (one per species) for the zero-inflation model.

zinbm1 = glmmTMB(count~spp + (1|site), zi=~mined, Salamanders, family=nbinom2)

This again fits without a warning, but we see that the zero-inflation is effectively zero in the unmined (“minedno”) condition (plogis(0.38-17.5) is approximately \(4 \times 10^{-8}\)). We can estimate the confidence interval, but it takes some extra work: the default Wald standard errors and confidence intervals are useless in this case.

## at present we need to specify the parameter by number; for
##  extreme cases need to specify the parameter range
## (not sure why the upper bound needs to be so high ... ?)
cc = confint(zinbm1,method="uniroot",parm=9, parm.range=c(-20,20))

The lower CI is not defined; the upper CI is -2.08, i.e. we can state that the zero-inflation probability is less than plogis(-2.08) = 0.11.

More broadly, general inspection of the data (e.g., plotting the response against potential covariates) should help to diagnose overly complex models.

Example 2.

In some cases, scaling predictor variables may help. For example, in this example from @phisanti, the results of glm and glmmTMB applied to a scaled version of the data set agree, while glmmTMB applied to the raw data set gives a non-positive-definite Hessian warning.

## data taken from, originally
load(system.file("vignette_data","plasma.rda", package="glmmTMB"))
m4.1 <- glm(calories ~ fat*fiber, family = Gamma(link = "log"), data = plasma)
m4.2 <- glmmTMB(calories ~ fat*fiber, family = Gamma(link = "log"), data = plasma)
ps  <- transform(plasma,fat=scale(fat,center=FALSE),fiber=scale(fiber,center=FALSE))
m4.3 <- update(m4.2, data=ps)
## scaling factor for back-transforming standard deviations
ss <- c(1,
        fatsc <- 1/attr(ps$fat,"scaled:scale"),
        fibsc <- 1/attr(ps$fiber,"scaled:scale"),
## combine SEs, suppressing the warning from the unscaled model
s_vals <- cbind(glm=sqrt(diag(vcov(m4.1))),

Example 3.

Here is another example (from Samantha Sherman):


The first model gives the specified warning when it runs, as well as the other symptoms such as NA values for the likelihood:


We can immediately see that the dispersion is very small and that the zero-inflation parameter is strongly negative. However, we’ll develop some fancier machinery that checks the variance-covariance matrix or Hessian of the model, finds eigenvalues that are negative or close to zero, and identifies which model components contribute to those eigenvalues:

diagnose_vcov <- function(model, tol=1e-5, digits=2, analyze_hessian=FALSE) {
    vv <- vcov(model, full=TRUE)
    nn <- rownames(vv)
    if (!all(is.finite(vv))) {
        if (missing(analyze_hessian)) warning("analyzing Hessian, not vcov")
        if (!analyze_hessian) stop("can't analyze vcov")
        analyze_hessian <- TRUE
    if (analyze_hessian) {
        par.fixed <- model$obj$env$
        r <- model$obj$env$random
        if (!is.null(r)) par.fixed <- par.fixed[-r]
        vv <- optimHess(par.fixed, fn=model$obj$fn, gr=model$obj$gr)
        ## note vv is now HESSIAN, not vcov
    ee <- eigen(vv)
    if (all(ee$values>tol)) {message("var-cov matrix OK"); return(invisible(NULL))}
    ## find negative or small-positive eigenvalues (flat/wrong curvature)
    bad_evals <- which(ee$values<tol)
    ## order worst to best
    bad_evals <- bad_evals[order(-ee$values[bad_evals])]
    ret <- lapply(bad_evals,
                  function(i) {
                      ## extract loadings
                      v <- setNames(ee$vectors[,i], nn)
                      ## order in decreasing magnitude & round

Running the diagnostics on the model:

(d1 <- diagnose_vcov(mod1))

This model has a very bad eigenvalue that is mostly driven by the zero-inflation parameter, and a little bit by the dispersion parameter. Let’s try dropping the zero-inflation term:

mod2 <- update(mod1, ziformula=~0)

We still get the warning, and the NA-valued likelihoods (and the very small dispersion parameter). Diagnose:


We can see that the dispersion parameter is still problematic. Simplify the model by switching from NB1 to Poisson:

mod3 <- update(mod2, family=poisson)

There are no warnings, the model looks OK now, and the diagnostic function agrees:


You can also check directly whether the model is OK by examining the pdHess (“positive-definite Hessian”) component of the sdr (“standard deviation report”) component of the model:


(FIXME: add an accessor method for this?)

In general models with non-positive definite Hessian matrices should be excluded from further consideration.

Model convergence problem: eigenvalue problems

m1 = glmmTMB(count~spp + mined + (1|site), zi=~spp + mined, Salamanders, family=genpois)

In this example, the fixed-effect covariance matrix is NaN. It may have to do with the generalized Poisson (genpois) distribution, which is known to have convergence problems; luckily, the negative binomial (nbinom1 and nbinom2) and/or Conway-Maxwell Poisson (compois) are good alternatives.

Models with convergence problems should be excluded from further consideration, in general.

In some cases, extreme eigenvalues may be caused by having predictor variables that are on very different scales: try rescaling, and centering, continuous predictors in the model.

NA/NaN function evaluation

Warning in nlminb(start = par, objective = fn, gradient = gr) : NA/NaN function evaluation

This warning occurs when the optimizer visits a region of parameter space that is invalid. It is not a problem as long as the optimizer has left that region of parameter space upon convergence, which is indicated by an absence of the model convergence warnings described above.

The following warnings indicate possibly-transient numerical problems with the fit, and can be treated in the same way (i.e. ignored if there are no errors or convergence warnings about the final fitted model).

Cholmod warning ‘matrix not positive definite’

In older versions of R (< 3.6.0):

Warning in f(par, order = order, …) : value out of range in ‘lgamma’

false convergence

This warning:

false convergence: the gradient ∇f(x) may be computed incorrectly, the other stopping tolerances may be too tight, or either f or ∇f may be discontinuous near the current iterate x

comes from the nlminb optimizer used by default in glmmTMB. It’s usually hard to diagnose the source of this warning (this Stack Overflow answer explains a bit more about what it means). Reasonable methods for making sure your model is OK are:

and see if the results are sufficiently similar to the original fit.


NA/NaN gradient evaluation

dat1 = expand.grid(y=-1:1, rep=1:10)
m1 = glmmTMB(y~1, dat1, family=nbinom2)

The error occurs here because the negative binomial distribution is inappropriate for data with negative values.

If you see this error, check that the response variable meets the assumptions of the specified distribution.

gradient length

Error in nlminb(start = par, objective = fn, gradient = gr) : gradient function must return a numeric vector of length x

Error in optimHess(par.fixed, obj\(fn, obj\)gr): gradient in optim evaluated to length x

Try rescaling predictor variables. Try a simpler model and build up. (If you have a simple reproducible example of these errors, please post them to the issues list.)