Jags User Manual
Jags User Manual
0 user manual
Martyn Plummer
1 October 2012
Contents
1 Introduction 3
2 Running a model in JAGS 4
2.1 Denition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
2.1.1 Model denition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
2.1.2 Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
2.1.3 Node Array dimensions . . . . . . . . . . . . . . . . . . . . . . . . . . 6
2.2 Compilation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
2.3 Initialization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
2.3.1 Parameter values . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
2.3.2 RNGs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
2.3.3 Samplers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
2.4 Adaptation and burn-in . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
2.5 Monitoring . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
3 Running JAGS 12
3.1 Scripting commands . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
3.1.1 MODEL IN . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
3.1.2 DATA IN . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
3.1.3 COMPILE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
3.1.4 PARAMETERS IN . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
3.1.5 INITIALIZE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
3.1.6 UPDATE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
3.1.7 ADAPT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
3.1.8 MONITOR . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
3.1.9 CODA . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
3.1.10 EXIT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
3.1.11 DATA TO . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
3.1.12 PARAMETERS TO . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
3.1.13 SAMPLERS TO . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
3.1.14 LOAD . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
3.1.15 UNLOAD . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
3.1.16 LIST MODULES . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
3.1.17 LIST FACTORIES . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
3.1.18 SET FACTORY . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
3.1.19 MODEL CLEAR . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
1
3.1.20 Print Working Directory (PWD) . . . . . . . . . . . . . . . . . . . . . 16
3.1.21 Change Directory (CD) . . . . . . . . . . . . . . . . . . . . . . . . . . 16
3.1.22 Directory list (DIR) . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
3.1.23 RUN . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
3.2 Errors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
4 Modules 18
4.1 The base module . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
4.1.1 Base Samplers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
4.1.2 Base RNGs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
4.1.3 Base Monitors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
4.2 The bugs module . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
4.3 The mix module . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
4.4 The dic module . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
4.4.1 The deviance monitor . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
4.4.2 The pD monitor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
4.4.3 The popt monitor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
4.5 The msm module . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
4.6 The glm module . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
5 Functions 22
5.1 Base functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
5.2 Functions in the bugs module . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
5.2.1 Scalar functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
5.2.2 Scalar-valued functions with vector arguments . . . . . . . . . . . . . 24
5.2.3 Vector- and array-valued functions . . . . . . . . . . . . . . . . . . . . 24
5.3 Function aliases . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24
6 Distributions 28
6.1 Distribution aliases . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
7 Dierences between JAGS and OpenBUGS 31
7.0.1 Data format . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
7.0.2 Distributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
7.0.3 Observable Functions . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
7.0.4 Data transformations . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
7.0.5 Directed cycles . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34
7.0.6 Censoring, truncation and prior ordering . . . . . . . . . . . . . . . . . 35
8 Feedback 37
9 Acknowledgments 38
2
Chapter 1
Introduction
JAGS is Just Another Gibbs Sampler. It is a program for the analysis of Bayesian models
using Markov Chain Monte Carlo (MCMC) which is not wholly unlike OpenBUGS (http:
//www.openbugs.info). JAGS was written with three aims in mind: to have an engine for
the BUGS language that runs on Unix; to be extensible, allowing users to write their own
functions, distributions, and samplers; and to be a platform for experimentation with ideas
in Bayesian modelling.
JAGS is designed to work closely with the R language and environment for statistical
computation and graphics (http://www.r-project.org). You will nd it useful to install
the coda package for R to analyze the output. You can also use the rjags package to work
directly with JAGS from within R (but note that the rjags package is not described in this
manual).
JAGS is licensed under the GNU General Public License version 2. You may freely modify
and redistribute it under certain conditions (see the le COPYING for details).
3
Chapter 2
Running a model in JAGS
JAGS is designed for inference on Bayesian models using Markov Chain Monte Carlo (MCMC)
simulation. Running a model refers to generating samples from the posterior distribution of
the model parameters. This takes place in ve steps:
1. Denition of the model
2. Compilation
3. Initialization
4. Adaptation and burn-in
5. Monitoring
The next stages of analysis are done outside of JAGS: convergence diagnostics, model criticism,
and summarizing the samples must be done using other packages more suited to this task.
There are several R packages designed for analyzing MCMC output, and JAGS can be used
from within R using the rjags package.
2.1 Denition
There are two parts to the denition of a model in JAGS: a description of the model and the
denition of the data.
2.1.1 Model denition
The model is dened in a text le using a dialect of the BUGS language. The model denition
consists of a series of relations inside a block delimited by curly brackets { and } and preceded
by the keyword model. Here is the standard linear regression example:
model {
for (i in 1:N) {
Y[i] ~ dnorm(mu[i], tau)
mu[i] <- alpha + beta * (x[i] - x.bar)
}
x.bar <- mean(x)
4
alpha ~ dnorm(0.0, 1.0E-4)
beta ~ dnorm(0.0, 1.0E-4)
sigma <- 1.0/sqrt(tau)
tau ~ dgamma(1.0E-3, 1.0E-3)
}
Each relation denes a node in the model in terms of other nodes that appear on the right
hand side. These are referred to as the parent nodes. Taken together, the nodes in the model
(together with the parent/child relationships represented as directed edges) form a directed
acyclic graph. The very top-level nodes in the graph, with no parents, are constant nodes,
which are dened either in the model denition (e.g. 1.0E-3), or in the data le (e.g. x[1]).
Relations can be of two types. A stochastic relation (~) denes a stochastic node, repre-
senting a random variable in the model. A deterministic relation (<-) denes a deterministic
node, the value of which is determined exactly by the values of its parents.
Nodes dened by a relation are embedded in named arrays. Array names may contain
letters, numbers, decimal points and underscores, but they must start with a letter. The
node array mu is a vector of length N containing N nodes (mu[1], . . ., mu[N]). The node
array alpha is a scalar. JAGS follows the S language convention that scalars are considered
as vectors of length 1. Hence the array alpha contains a single node alpha[1].
Deterministic nodes do not need to be embedded in node arrays. The node Y[i] could
equivalently be dened as
Y[i] ~ dnorm(alpha + beta * (x[i] - x.bar), tau)
In this version of the model denition, the node previously dened as mu[i] still exists, but is
not accessible to the user as it does not have a name. This ability to hide deterministic nodes
by embedding them in other expressions underscores an important fact: only the stochas-
tic nodes in a model are really important. Deterministic nodes are merely a syntactically
convenient way of describing the relations between stochastic nodes, or transformations of
them.
2.1.2 Data
The data are dened in a separate le from the model denition, in the format created by the
dump() function in R(See appendix [? ]). Data values may be supplied for stochastic nodes
and constants (including constant values used inside for loops). It is an error to supply a data
value for a deterministic node. (See, however, section 7.0.3 on observable functions).
Here are the data for the LINE example:
x <- c(1, 2, 3, 4, 5)
#R-style comments, like this one, can be embedded in the data file
Y <- c(1, 3, 3, 3, 5)
N <- 5
The unobserved stochastic nodes are referred to as the parameters of the model. The
data le therefore denes the parameters of the model by omission. In the LINE example, the
parameters are alpha, beta and tau.
If a node array contains both observed and unobserved nodes, then the data should contain
missing values (NA) for the unobserved elements. In the LINE example, if Y[2] and Y[5] were
unobserved, then the data would be
5
Y <- c(1, NA, 3, 3, NA)
Multivariate nodes cannot be partially observed, so if a node takes up two or more elements
of a node array, then the corresponding data values must be all present or all missing.
2.1.3 Node Array dimensions
Array declarations
JAGS allows the option of declaring the dimensions of node arrays in the model le. The
declarations consist of the keyword var (for variable) followed by a comma-separated list of
array names, with their dimensions in square brackets. The dimensions may be given in terms
of any expression of the data that returns a single integer value.
In the linear regression example, the model block could be preceded by
var x[N], Y[N], mu[N], alpha, beta, tau, sigma, x.bar;
Undeclared nodes
If a node array is not declared then JAGS has three methods of determining its size.
1. Using the data. The dimension of an undeclared node array may be inferred if it is
supplied in the data le.
2. Using the left hand side of the relations. The maximal index values on the left
hand side of a relation are taken to be the dimensions of the node array. For example,
in this case:
for (i in 1:N) {
for (j in 1:M) {
Y[i,j] ~ dnorm(mu[i,j], tau)
}
}
Y would be inferred to be an N M matrix. This method cannot be used when there
are empty indices (e.g. Y[i,]) on the left hand side of the relation.
3. Using the dimensions of the parents If a whole node array appears on the left hand
side of a relation, then its dimensions can be inferred from the dimensions of the nodes
on the right hand side. For example, if A is known to be an N N matrix and
B <- inverse(A)
Then B is also an N N matrix.
6
Querying array dimensions
The JAGS compiler has two built-in functions for querying array sizes. The length() function
returns the number of elements in a node array, and the dim() function returns a vector
containing the dimensions of an array. These two functions may be used to simplify the data
preparation. For example, if Y represents a vector of observed values, then using the length()
function in a for loop:
for (i in 1:length(Y)) {
Y[i] ~ dnorm(mu[i], tau)
}
avoids the need to put a separate data value N in the le representing the length of Y.
For multi-dimensional arrays, the dim function serves a similar purpose. The dim function
returns a vector, which must be stored in an array before its elements can be accessed. For
this reason, calls to the dim function must always be in a data block (see section 7.0.4).
data {
D <- dim(Z)
}
model {
for (i in 1:D[1]) {
for (j in 1:D[2]) {
Z[i,j] <- dnorm(alpha[i] + beta[j], tau)
}
}
...
}
Clearly, the length() and dim() functions can only work if the size of the node array can be
inferred, using one of the three methods outlined above.
Note: the length() and dim() functions are dierent from all other functions in JAGS:
they do not act on nodes, but only on node arrays. As a consequence, an expression such as
dim(a %*% b) is syntactically incorrect.
2.2 Compilation
When a model is compiled, a graph representing the model is created in computer memory.
Compilation can fail for a number of reasons:
1. The graph contains a directed cycle. These are forbidden in JAGS.
2. A top-level parameter is undened. Any node that is used on the right hand side of
a relation, but is not dened on the left hand side of any relation, is assumed to be a
constant node. Its value must be supplied in the data le.
3. The model uses a function or distribution that has not been dened in any of the loaded
modules.
The number of parallel chains to be run by JAGS is also dened at compilation time. Each
parallel chain should produce an independent sequence of samples from the posterior distri-
bution. By default, JAGS only runs a single chain.
7
2.3 Initialization
Before a model can be run, it must be initialized. There are three steps in the initialization
of a model:
1. The initial values of the model parameters are set.
2. A Random Number Generator (RNG) is chosen for each parallel chain, and its seed is
set.
3. The Samplers are chosen for each parameter in the model.
2.3.1 Parameter values
The user may supply initial value les one for each chain containing initial values for the
model parameters. The les may not contain values for logical or constant nodes. The format
is the same as the data le (see (See appendix [? ]). Section describes how to read the initial
value les into the model.
As with the data le, you may supply missing values in the initial values le. This need
typically arises with contrast parameters. Suppose X is a categorical covariate taking values
from 1 to 4. There is one parameter for each level of x (
1
. . .
4
) , but the rst parameter
1
is set to zero for identiability. The remaining parameters
2
. . .
4
represent contrasts with
respect to the rst level of X.
for (i in 1:N) {
Y[i] ~ alpha + beta[x[i]]
}
# Prior distribution
alpha ~ dnorm(0, 1.0E-3)
beta[1] <- 0
for(i in 2:4) {
beta[i] ~ dnorm(0, 1.0E-3)
}
A suitable initial value for beta would be
beta <- c(NA, 1.03, -2.03, 0.52)
This allows parameter values to be supplied for the stochastic elements of beta but not the
constant rst element.
If initial values are not supplied by the user, then each parameter chooses its own initial
value based on the values of its parents. The initial value is chosen to be a typical value
from the prior distribution. The exact meaning of typical value depends on the distribution
of the stochastic node, but is usually the mean, median, or mode.
If you rely on automatic initial value generation and are running multiple parallel chains,
then the initial values will be the same in all chains. You may not want this behaviour,
especially if you are using the Gelman and Rubin convergence diagnostic, which assumes that
the initial values are over-dispersed with respect to the posterior distribution. In this case,
you are advised to set the starting values manually using the parameters in statement.
8
2.3.2 RNGs
Each chain in JAGS has its own random number generator (RNG). RNGs are more correctly
referred to as pseudo-random number generators. They generate a sequence of numbers
that merely looks random but is, in fact, entirely determined by the initial state. You may
optionally set the name of the RNG and its initial state in the initial values le.
The name of the RNG is set as follows.
.RNG.name <- "name"
There are four RNGs supplied by the base module in JAGS with the following names:
"base::Wichmann-Hill"
"base::Marsaglia-Multicarry"
"base::Super-Duper"
"base::Mersenne-Twister"
There are two ways to set the starting state of the RNG. The simplest is to supply an
integer value to .RNG.seed, e.g.
".RNG.seed" <- 314159
The second is way to save the state of the RNG from one JAGS session (see the PARAM-
ETERS TO statement, section 3.1.12) and use this as the initial state of a new chain. The
state of any RNG in JAGS can be saved and loaded as an integer vector with the name
.RNG.state. For example,
".RNG.state" <- as.integer(c(20899,10892,29018))
is a valid state for the Marsaglia-Multicarry generator. You cannot supply an arbitrary integer
to .RNG.state. Both the length of the vector and the permitted values of its elements are
determined by the class of the RNG. The only safe way to use .RNG.state is to re-use a
previously saved state.
If no RNG names are supplied, then RNGs will be chosen automatically so that each
chain has its own independent random number stream. The exact behaviour depends on
which modules are loaded. The base module uses the four generators listed above for the
rst four chains, then recycles them with dierent seeds for the next four chains, and so on.
By default, JAGS bases the initial state on the time stamp. This means that, when a
model is re-run, it generates an independent set of samples. If you want your model run to
be reproducible, you must explicitly set the .RNG.seed for each chain.
2.3.3 Samplers
A Sampler is an object that acts on a set of parameters and updates them from one iteration
to the next. During initialization of the model, Samplers are chosen automatically for all
parameters.
The Model holds an internal list of Sampler Factory objects, which inspect the graph,
recognize sets of parameters that can be updated with specic methods, and generate Sampler
objects for them. The list of Sampler Factories is traversed in order, starting with sampling
methods that are ecient, but limited to certain specic model structures and ending with
9
the most generic, possibly inecient, methods. If no suitable Sampler can be generated for
one of the model parameters, an error message is generated.
The user has no direct control over the process of choosing Samplers. However, you may
indirectly control the process by loading a module that denes a new Sampler Factory. The
module will insert the new Sampler Factory at the beginning of the list, where it will be
queried before all of the other Sampler Factories. You can also optionally turn on and o
sampler factories using the SET FACTORY command. See 3.1.18.
A report on the samplers chosen by the model, and the stochastic nodes they act on, can
be generated using the SAMPLERS TO command. See section 3.1.13.
2.4 Adaptation and burn-in
In theory, output from an MCMC sampler converges to the target distribution (i.e. the
posterior distribution of the model parameters) in the limit as the number of iterations tends
to innity. In practice, all MCMC runs are nite. By convention, the MCMC output is
divided into two parts: an initial burn-in period, which is discarded, and the remainder
of the run, in which the output is considered to have converged (suciently close) to the
target distribution. Samples from the second part are used to create approximate summary
statistics for the target distribution.
By default, JAGS keeps only the current value of each node in the model, unless a monitor
has been dened for that node. The burn-in period of a JAGS run is therefore the interval
between model initialization and the creation of the rst monitor.
When a model is initialized, it may be in adaptive mode, meaning that the Samplers
used by the model may modify their behaviour for increased eciency. Since this adaptation
may depend on the entire sample history, the sequence generated by an adapting sampler
is no longer a Markov chain, and is not guaranteed to converge to the target distribution.
Therefore, adaptive mode must be turned o at some point during burn-in, and a sucient
number of iterations must take place after the adaptive phase to ensure successful burnin.
By default, adaptive mode is turned o half way through rst update of a JAGS model.
All samplers have a built in test to determine whether they have converged to their optimal
sampling behaviour. If any sampler fails this validation test, a warning will be printed. To
ensure optimal sampling behaviour, the model should be run again from scratch using a longer
adaptation period.
The adapt command (see section 3.1.7) can be used for more control over the adaptive
phase. The adapt command updates the model but keeps it in adaptive mode. At the end
of each update, the convergence test is called. The message Adaptation successful will
be printed if the convergence test is successful, otherwise the message will read Adaptation
incomplete. Successive calls to adapt are possible while keeping the model in adaptive mode.
The next update command will immediately turn o adaptive mode.
2.5 Monitoring
A monitor in JAGS is an object that records sampled values. The simplest monitor is a trace
monitor, which stores the sampled value of a node at each iteration.
JAGS cannot monitor a node unless it has been dened in the model le. For vector- or
array-valued nodes, this means that every element must be dened. Here is an example of a
10
simple for loop that only denes elements 2 to N of theta
for (i in 2:N) {
theta[i] ~ dnorm(0,1);
}
Unless theta[1] is dened somewhere else in the model le, the multivariate node theta
is undened and therefore it will not be possible to monitor theta as a whole. In such cases
you can request each element separately , e.g. theta[2], theta[3], etc., or request a subset
that is fully dened, e.g. theta[2:6].
Monitors can be classied according to whether they pool values over iterations and
whether they pool values over parallel chains (The standard trace monitor does neither).
When monitor values are written out to le using the CODA command, the output les cre-
ated depend on the pooling of the monitor, as shown in table 2.1. By default, all of these les
have the prex CODA, but this may be changed to any other name using the stem option
to the CODA command (See 3.1.9).
Pool Pool Output les
iterations chains
no no CODAindex.txt, CODAchain1.txt, ... CODAchainN.txt
no yes CODAindex0.txt, CODAchain0.txt
yes no CODAtable1.txt, ... CODAtableN.txt
yes yes CODAtable0.txt
Table 2.1: Output les created by the CODA command depending on whether a monitor
pools its values over chains or over iterations
The standard CODA format for monitors that do not pool values over iterations is to
create an index le and one or more output les. The index le is has three columns with,
one each line,
1. A string giving the name of the (scalar) value being recorded
2. The rst line in the output le(s)
3. The last line in the output le(s)
The output le(s) contain two columns:
1. The iteration number
2. The value at that iteration
Some monitors pool values over iterations. For example a mean monitor may record only
the sample mean of a node, without keeping the individual values from each iteration. Such
monitors are written out to a table le with two columns:
1. A string giving the name of the (scalar) value being recorded
2. The value (pooled over all iterations)
11
Chapter 3
Running JAGS
JAGS has a command line interface. To invoke jags interactively, simply type jags at the
shell prompt on Unix, a Terminal window on MacOS, or the Windows command prompt on
Windows. To invoke JAGS with a script le, type
jags <script file>
A typical script le has the following commands:
model in "line.bug" # Read model file
data in "line-data.R" # Read data in from file
compile, nchains(2) # Compile a model with two parallel chains
parameters in "line-inits1.R", chain(1) # Read initial values from file for chain 1
parameters in "line-inits2.R", chain(2) # Read initial values from file for chain 2
initialize # Initialize the model
update 1000 # Adaptation (if necessary) and burnin for 1000 iterations
monitor alpha # Set trace monitor for node alpha ...
monitor beta # ... and beta
monitor sigma # ... and sigma
update 10000 # Update model for 10000 iterations
coda * # All monitored values are written out to file
More examples can be found in the le classic-bugs.tar.gz which is available from Source-
forge (http://sourceforge.net/projects/mcmc-jags/files.
Output from JAGS is printed to the standard output, even when a script le is being used.
3.1 Scripting commands
JAGS has a simple set of scripting commands with a syntax loosely based on Stata. Commands
are shown below preceded by a dot (.). This is the JAGS prompt. Do not type the dot in
when you are entering the commands.
C-style block comments taking the form /* ... */ can be embedded anywhere in the script
le. Additionally, you may use R-style single-line comments starting with #.
If a scripting command takes a le name, then the name may be optionally enclosed in
quotes. Quotes are required when the le name contains space, or any character which is not
alphanumeric, or one of the following: _, -, ., /, \.
12
In the descriptions below, angular brackets <>, and the text inside them, represents a
parameter that should be replaced with the correct value by you. Anything inside square
brackets [] is optional. Do not type the square brackets if you wish to use an option.
3.1.1 MODEL IN
. model in <file>
Checks the syntactic correctness of the model description in file and reads it into memory.
The next compilation statement will compile this model.
See also: MODEL CLEAR (3.1.19)
3.1.2 DATA IN
. data in <file>
JAGS keeps an internal data table containing the values of observed nodes inside each node
array. The DATA IN statement reads data from a le into this data table.
Several data statements may be used to read in data from more than one le. If two data
les contain data for the same node array, the second set of values will overwrite the rst,
and a warning will be printed.
See also: DATA TO (3.1.11).
3.1.3 COMPILE
. compile [, nchains(<n>)]
Compiles the model using the information provided in the preceding model and data state-
ments. By default, a single Markov chain is created for the model, but if the nchains option
is given, then n chains are created
Following the compilation of the model, further DATA IN statements are legal, but have
no eect. A new model statement, on the other hand, will replace the current model.
3.1.4 PARAMETERS IN
. parameters in <file> [, chain(<n>)]
Reads the values in file and writes them to the corresponding parameters in chain n. The
le has the same format as the one in the DATA IN statement. The chain option may be
omitted, in which case the parameter values in all chains are set to the same value.
The PARAMETERS IN statement must be used after the COMPILE statement and
before the INITIALIZE statement. You may only supply the values of unobserved stochastic
nodes in the parameters le, not logical or constant nodes.
See also: PARAMETERS TO (3.1.12)
3.1.5 INITIALIZE
. initialize
Initializes the model using the previously supplied data and parameter values supplied for
each chain.
13
3.1.6 UPDATE
. update <n> [,by(<m>)]
Updates the model by n iterations.
If JAGS is being run interactively, a progress bar is printed on the standard output
consisting of 50 asterisks. If the by option is supplied, a new asterisk is printed every m
iterations. If this entails more than 50 asterisks, the progress bar will be wrapped over
several lines. If m is zero, the printing of the progress bar is suppressed.
If JAGS is being run in batch mode, then the progress bar is suppressed by default, but
you may activate it by supplying the by option with a non-zero value of m.
If the model has an adaptive sampling phase, the rst UPDATE statement turns o
adaptive mode for all samplers in the model after n/2 iterations. A warning is printed if
adaptation is incomplete. Incomplete adaptation means that the mixing of the Markov chain
is not optimal. It is still possible to continue with a model that has not completely adapted,
but it may be preferable to run the model again with a longer adaptation phase, starting
from the MODEL IN statement. Alternatively, you may use an ADAPT statement (see
below) immediately after initialization.
3.1.7 ADAPT
. adapt <n> [,by(<m>)]
Updates the model by n iterations keeping the model in adaptive mode and prints a message to
indicate whether adaptation is successful. Successive calls to ADAPT may be made until the
adaptation is successful. The next call to UPDATE then turns o adaptive mode immediately.
Use this instead of the rst UPDATE statement if you want explicit control over the
length of the adaptive sampling phase.
Like the UPDATE statement, the ADAPT statement prints a progress bar, but with plus
signs instead of asterisks.
3.1.8 MONITOR
In JAGS, a monitor is an object that calculates summary statistics from a model. The most
commonly used monitor simply records the value of a single node at each iteration. This is
called a trace monitor.
. monitor <varname> [, thin(n)] [type(<montype>)]
The thin option sets the thinning interval of the monitor so that it will only record every nth
value. The thin option selects the type of monitor to create. The default type is trace.
More complex monitors can be dened that do additional calculations. For example,
the dic module denes a deviance monitor that records the deviance of the model at
each iteration, and a pD monitor that calculates an estimate of the eective number of
parameters on the model [? ].
. monitor clear <varname> [type(<montype>)]
Clears the monitor of the given type associated with variable <varname>.
14
3.1.9 CODA
. coda <varname> [, stem(<filename>)]
If the named node has a trace monitor, this dumps the monitored values of to les CODAindex.txt,
CODAindex1.out, CODAindex2.txt, . . . in a form that can be read by the coda package of R.
The stem option may be used to modify the prex from CODA to another string. The
wild-card character * may be used to dump all monitored nodes
3.1.10 EXIT
. exit
Exits JAGS. JAGS will also exit when it reads an end-of-le character.
3.1.11 DATA TO
. data to <filename>
Writes the data (i.e. the values of the observed nodes) to a le in the R dump format. The
same le can be used in a DATA IN statement for a subsequent model.
See also: DATA IN (3.1.2)
3.1.12 PARAMETERS TO
. parameters to <file> [, chain(<n>)]
Writes the current parameter values (i.e. the values of the unobserved stochastic nodes) in
chain <n> to a le in R dump format. The name and current state of the RNG for chain
<n> is also dumped to the le. The same le can be used as input in a PARAMETERS IN
statement in a subsequent run.
See also: PARAMETERS IN (3.1.4)
3.1.13 SAMPLERS TO
. samplers to <file>
Writes out a summary of the samplers to the given le. The output appears in three tab-
separated columns, with one row for each sampled node
The index number of the sampler (starting with 1). The index number gives the order
in which Samplers are updated at each iteration.
The name of the sampler, matching the index number
The name of the sampled node.
If a Sampler updates multiple nodes then it is represented by multiple rows with the same
index number.
Note that the list includes only those nodes that are updated by a Sampler. Stochastic
nodes that are updated by forward sampling from the prior are not listed.
15
3.1.14 LOAD
. load <module>
Loads a module into JAGS (see chapter 4). Loading a module does not aect any previously
initialized models, but will aect the future behaviour of the compiler and model initialization.
3.1.15 UNLOAD
. unload <module>
Unloads a module. Currently initialized models are unaected, but the functions, distribution,
and factory objects in the model will not be accessible to future models.
3.1.16 LIST MODULES
. list modules
Prints a list of the currently loaded modules.
3.1.17 LIST FACTORIES
. list factories, type(<factype>)
List the currently loaded factory objects and whether or not they are active. The type option
must be given, and the three possible values of <factype> are sampler, monitor, and rng.
3.1.18 SET FACTORY
. set factory "<facname>" <status>, type(<factype>)
Sets the status of a factor object. The possible values of <status> are on and off. Possible
factory names are given from the LIST MODULES command.
3.1.19 MODEL CLEAR
. model clear
Clears the current model. The data table (see section 3.1.2) remains intact
3.1.20 Print Working Directory (PWD)
. pwd
Prints the name of the current working directory. This is where JAGS will look for les when
the le name is given without a full path, e.g. "mymodel.bug".
3.1.21 Change Directory (CD)
. cd <dirname>
Changes the working directory to <dirname>
16
3.1.22 Directory list (DIR)
. dir
Lists the les in the current working directory.
3.1.23 RUN
. run <cmdfile>
Opens the le <cmdfile> and reads further scripting commands until the end of the le. Note
that if the le contains an EXIT statement, then the JAGS session will terminate.
3.2 Errors
There are two kinds of errors in JAGS: runtime errors, which are due to mistakes in the model
specication, and logic errors which are internal errors in the JAGS program.
Logic errors are generally created in the lower-level parts of the JAGS library, where it is
not possible to give an informative error message. The upper layers of the JAGS program are
supposed to catch such errors before they occur, and return a useful error message that will
help you diagnose the problem. Inevitably, some errors slip through. Hence, if you get a logic
error, there is probably an error in your input to JAGS, although it may not be obvious what
it is. Please send a bug report (see Feedback below) whenever you get a logic error.
Error messages may also be generated when parsing les (model les, data les, command
les). The error messages generated in this case are created automatically by the program
bison. They generally take the form syntax error, unexpected FOO, expecting BAR and
are not always abundantly clear.
If a model compiles and initializes correctly, but an error occurs during updating, then
the current state of the model will be dumped to a le named jags.dumpN.R where N is the
chain number. You should then load the dumped data into R to inspect the state of each
chain when the error occurred.
17
Chapter 4
Modules
The JAGS library is distributed along with certain dynamically loadable modules that extend
its functionality. A module can dene new objects of the following classes:
1. functions and distributions, the basic building blocks of the BUGS language.
2. samplers, the objects which update the parameters of the model at each iteration, and
sampler factories, the objects that create new samplers for specic model structures.
If the module denes a new distribution, then it will typically also dene a new sampler
for that distribution.
3. monitors, the objects that record sampled values for later analysis, and monitor
factories that create them.
4. random number generators, the objects that drive the MCMC algorithm and RNG
factories that create them.
The base module and the bugs module are loaded automatically at start time. Others
may be loaded by the user.
4.1 The base module
The base module supply the base functionality for the JAGS library to function correctly. It
is loaded rst by default.
4.1.1 Base Samplers
The base module denes samplers that use highly generic update methods. These sampling
methods only require basic information about the stochastic nodes they sample. Conversely,
they may not be fully ecient.
Three samplers are currently dened:
1. The Finite sampler can sample a discrete-valued node with xed support of less than
20 possible values. The node must not be bounded using the T(,) construct
2. The Real Slice Sampler can sample any scalar real-valued stochastic node.
3. The Discrete Slice Sampler can sample any scalar discrete-valued stochastic node.
18
4.1.2 Base RNGs
The base module denes four RNGs, taken directly from R, with the following names:
1. "base::Wichmann-Hill"
2. "base::Marsaglia-Multicarry"
3. "base::Super-Duper"
4. "base::Mersenne-Twister"
A single RNG factory object is also dened by the base module which will supply these
RNGs for chains 1 to 4 respectively, if RNG.name is not specied in the initial values le.
All chains generated by the base RNG factory are initialized using the current time stamp.
If you have more than four parallel chains, then the base module will recycle the same for
RNGs, but using dierent seeds. If you want many parallel chains then you may wish to load
the lecuyer module.
4.1.3 Base Monitors
The base module denes the TraceMonitor class (type trace). This is the monitor class
that simply records the current value of the node at each iteration.
4.2 The bugs module
The bugs module denes some of the functions and distributions from OpenBUGS. These are
described in more detail in sections 5 and 6. The bugs module also denes conjugate samplers
for ecient Gibbs sampling.
4.3 The mix module
The mix module denes a novel distribution dnormmix(mu,tau,pi) representing a nite mix-
ture of normal distributions. In the parameterization of the dnormmix distribution, , , and
are vectors of the same length, and the density of y ~ dnormmix(mu, tau, pi) is
f(y|, , ) =
1
2
i
(
1
2
i
(y
i
))
where () is the probability density function of a standard normal distribution.
The mix module also denes a sampler that is designed to act on nite normal mixtures.
It uses tempered transitions to jump between distant modes of the multi-modal posterior
distribution generated by such models [? ? ]. The tempered transition method is computa-
tionally very expensive. If you want to use the dnormmix distribution but do not care about
label switching, then you can disable the tempered transition sampler with
set factory "mix::TemperedMix" off, type(sampler)
19
4.4 The dic module
The dic module denes new monitor classes for Bayesian model criticism using deviance-
based measures.
4.4.1 The deviance monitor
The deviance monitor records the deviance of the model (i.e. the sum of the deviances of all
the observed stochastic nodes) at each iteration. The command
monitor deviance
will create a deviance monitor unless you have dened a node called deviance in your model.
In this case, you will get a trace monitor for your deviance node.
4.4.2 The pD monitor
The pD monitor is used to estimate the eective number of parameters (p
D
) of the model [?
]. It requires at least two parallel chains in the model, but calculates a single estimate of p
D
across all chains [? ]. A pD monitor can be created using the command:
monitor pD
Like the deviance monitor, however, if you have dened a node called pD in your model
then this will take precedence, and you will get a trace monitor for your pD node.
Since the p
D
monitor pools its value across all chains, its values will be written out to the
index le CODAindex0.txt and output le CODAoutput0.txt when you use the CODA
command.
The eective number of parameters is the sum of separate contributions from all observed
stochastic nodes: p
D
=
i
p
D
i
. There is also a monitor that stores the sample mean of
p
D
i
. These statistics may be used as inuence diagnostics [? ]. The mean monitor for p
D
i
is
created with:
monitor pD, type(mean)
Its values can be written out to a le PDtable0.txt with
coda pD, type(mean) stem(PD)
4.4.3 The popt monitor
The popt monitor works exactly like the mean monitor for p
D
, but records contributions to
the optimism of the expected deviance (p
opt
i
). The total optimism p
opt
=
i
p
opt
i
can be
added to the mean deviance to give the penalized expected deviance [? ].
The mean monitor for p
opt
i
is created with
monitor popt, type(mean)
Its values can be written out to a le POPTtable0.txt with
coda popt, type(mean) step(POPT)
20
Under asymptotically favourable conditions in which p
D
i
1i,
p
opt
2p
D
For generalized linear models, a better approximation is
p
opt
n
i=1
p
D
i
1 p
D
i
The popt monitor uses importance weights to estimate p
opt
. The resulting estimates may
be numerically unstable when p
D
i
is not small. This typically occurs in random-eects models,
so it is recommended to use caution with the popt until I can nd a better way of estimating
p
opt
i
.
4.5 The msm module
The msm module denes the matrix exponential function mexp and the multi-state distribution
dmstate which describes the transitions between observed states in continuous-time multi-
state Markov transition models.
4.6 The glm module
The glm module implements samplers for ecient updating of generalized linear mixed mod-
els. The fundamental idea is to do block updating of the parameters in the linear predictor.
The glm module is built on top of the Csparse and CHOLMOD sparse matrix libraries [? ?
] which allows updating of both xed and random eects in the same block. Currently, the
methods only work on parameters that have a normal prior distribution.
Some of the samplers are based in the idea of introducing latent normal variables that
reduce the GLM to a linear model. This idea was introduced by Albert and Chib [? ]
for probit regression with a binary outcome, and was later rened and extended to logistic
regression with binary outcomes by Holmes and Held [? ]. Another approach, auxiliary
mixture sampling, was developed by Fr uhwirth-Schnatter et al [? ] and is used for more
general Poisson regression and logistic regression models with binomial outcomes. Gamerman
[? ] proposed a stochastic version of the iteratively weighted least squares algorithm for
GLMs, which is also implemented in the glm module. However the IWLS sampler tends to
break down when there are many random eects in the model. It uses Metropolis-Hastings
updates, and the acceptance probability may be very small under these circumstances.
Block updating in GLMMs frees the user from the need to center predictor variables, like
this:
y[i] ~ dnorm(mu[i], tau)
mu[i] <- alpha + beta * (x[i] - mean(x))
The second line can simply be written
mu[i] <- alpha + beta * x[i]
without aecting the mixing of the Markov chain.
21
Chapter 5
Functions
Functions allow deterministic nodes to be dened using the <- (gets) operator. Most of the
functions in JAGS are scalar functions taking scalar arguments. However, JAGS also allows
arbitrary vector- and array-valued functions, such as the matrix multiplication operator %*%
and the transpose function t() dened in the bugs module, and the matrix exponential
function mexp() dened in the msm module. JAGS also uses an enriched dialect of the BUGS
language with a number of operators that are used in the S language.
Scalar functions taking scalar arguments are automatically vectorized. They can also be
called when the arguments are arrays with conforming dimensions, or scalars. So, for example,
the scalar c can be added to the matrix A using
B <- A + c
instead of the more verbose form
D <- dim(A)
for (i in 1:D[1])
for (j in 1:D[2]) {
B[i,j] <- A[i,j] + c
}
}
Inverse link functions are an exception to this rule (i.e. exp, icloglog, ilogit, phi) and cannot
be vectorized.
5.1 Base functions
The functions dened by the base module all appear as inx or prex operators. The syntax
of these operators is built into the JAGS parser. They are therefore considered part of the
modelling language. Table 5.1 lists them in reverse order of precedence.
Logical operators convert numerical arguments to logical values: zero arguments are con-
verted to FALSE and non-zero arguments to TRUE. Logical and comparison operators return
the value 1 if the result is TRUE and 0 if the result is FALSE. Comparison operators are
non-associative: the expression x < y < z, for example, is syntactically incorrect.
The %special% function is an exception in table 5.1. It is not a function dened by the
base module, but is a place-holder for any function with a name starting and ending with
22
Type Usage Description
Logical x || y Or
operators x && y And
!x Not
Comparison x > y Greater than
operators x >= y Greater than or equal to
x < y Less than
x <= y Less than or equal to
x == y Equal
x != y Not equal
Arithmetic x + y Addition
operators x - y Subtraction
x * y Multiplication
x / y Division
x %special% y User-dened operators
-x Unary minus
Power function x^y
Table 5.1: Base functions listed in reverse order of precedence
the character % Such functions are automatically recognized as inx operators by the JAGS
model parser, with precedence dened by table 5.1.
5.2 Functions in the bugs module
5.2.1 Scalar functions
Table 5.2 lists the scalar-valued functions in the bugs module that also have scalar arguments.
These functions are automatically vectorized when they are given vector, matrix, or array
arguments with conforming dimensions.
Table 5.4 lists the link functions in the bugs module. These are smooth scalar-valued func-
tions that may be specied using an S-style replacement function notation. So, for example,
the log link
log(y) <- x
is equivalent to the more direct use of its inverse, the exponential function:
y <- exp(x)
This usage comes from the use of link functions in generalized linear models.
Table 5.3 shows functions to calculate the probability density, probability function, and
quantiles of some of the distributions provided by the bugs module. These functions are
parameterized in the same way as the corresponding distribution. For example, if x has a
normal distribution with mean and precision
x ~ dnorm(mu, tau)
23
Then the usage of the corresponding density, probability, and quantile functions is:
density.x <- dnorm(x, mu, tau) # Density of normal distribution at x
prob.x <- pnorm(x, mu, tau) # P(X <= x)
quantile90.x <- qnorm(0.9, mu, tau) # 90th percentile
For details of the parameterization of the other distributions, see tables 6.1 and 6.2.
5.2.2 Scalar-valued functions with vector arguments
Table 5.5 lists the scalar-valued functions in the bugs module that take general arguments.
Unless otherwise stated in table 5.5, the arguments to these functions may be scalar, vector,
or higher-dimensional arrays.
The max() and min() functions work like the corresponding R functions. They take a
variable number of arguments and return the maximum/minimum element over all supplied
arguments. This usage is compatible with OpenBUGS, although more general.
5.2.3 Vector- and array-valued functions
Table 5.6 lists vector- or matrix-valued functions in the bugs module.
The sort and rank functions behaves like their R namesakes: sort accepts a vector and
returns the same values sorted in ascending order; rank returns a vector of ranks. This is
distinct from OpenBUGS, which has two scalar-valued functions rank and ranked.
5.3 Function aliases
A function may optionally have an alias, which can be used in the model denition in place
of the canonical name. Aliases are used to to avoid confusion with other software in which
functions may have dierent names. Table 5.7 shows the functions in the bugs module with
an alias.
24
Usage Description Value Restrictions on arguments
abs(x) Absolute value Real
arccos(x) Arc-cosine Real 1 < x < 1
arccosh(x) Hyperbolic arc-cosine Real 1 < x
arcsin(x) Arc-sine Real 1 < x < 1
arcsinh(x) Hyperbolic arc-sine Real
arctan(x) Arc-tangent Real
arctanh(x) Hyperbolic arc-tangent Real 1 < x < 1
cos(x) Cosine Real
cosh(x) Hyperbolic Cosine Real
cloglog(x) Complementary log log Real 0 < x < 1
equals(x,y) Test for equality Logical
exp(x) Exponential Real
icloglog(x) Inverse complementary Real
log log function
ifelse(x,a,b) If x then a else b Real
ilogit(x) Inverse logit Real
log(x) Log function Real x > 0
logfact(x) Log factorial Real x > 1
loggam(x) Log gamma Real x > 0
logit(x) Logit Real 0 < x < 1
phi(x) Standard normal cdf Real
pow(x,z) Power function Real If x < 0 then z is integer
probit(x) Probit Real 0 < x < 1
round(x) Round to integer Integer
away from zero
sin(x) Sine Real
sinh(x) Hyperbolic Sine Real
sqrt(x) Square-root Real x >= 0
step(x) Test for x 0 Logical
tan(x) Tangent Real
tanh(x) Hyperbolic Tangent Real
trunc(x) Round to integer Integer
towards zero
Table 5.2: Scalar functions in the bugs module
25
Distribution Density Distribution Quantile
Bernoulli dbern pbern qbern
Beta dbeta pbeta qbeta
Binomial dbin pbin qbin
Chi-square dchisqr pchisqr qchisqr
Double exponential ddexp pdexp qdexp
Exponential dexp pexp qexp
F df pf qf
Gamma dgamma pgamma qgamma
Generalized gamma dgen.gamma pgen.gamma qgen.gamma
Noncentral hypergeometric dhyper phyper qhyper
Logistic dlogis plogis qlogis
Log-normal dlnorm plnorm qlnorm
Negative binomial dnegbin pnegbin qnegbin
Noncentral Chi-square dnchisqr pnchisqr qnchisqr
Normal dnorm pnorm qnorm
Pareto dpar ppar qpar
Poisson dpois ppois qpois
Student t dt pt qt
Weibull dweib pweib qweib
Table 5.3: Functions to calculate the probability density, probability function, and quantiles
of some of the distributions provided by the bugs module.
Link function Description Range Inverse
cloglog(y) <- x Complementary log log 0 < y < 1 y <- icloglog(x)
log(y) <- x Log 0 < y y <- exp(x)
logit(y) <- x Logit 0 < y < 1 y <- ilogit(x)
probit(y) <- x Probit 0 < y < 1 y <- phi(x)
Table 5.4: Link functions in the bugs module
Function Description Restrictions
inprod(x1,x2) Inner product Dimensions of x1, x2 conform
interp.lin(e,v1,v2) Linear Interpolation e scalar,
v1, v2 conforming vectors
logdet(m) Log determinant m is a symmetric positive denite matrix
max(x1,x2,...) Maximum element among all arguments
mean(x) Mean of elements of x
min(x1,x2,...) Minimum element among all arguments
prod(x) Product of elements of x
sum(x) Sum of elements of x
sd(x) Standard deviation of elements of x
Table 5.5: Scalar-valued functions with general arguments in the bugs module
26
Usage Description Restrictions
inverse(a) Matrix inverse a is a symmetric positive denite matrix
rank(v) Ranks of elements of v v is a vector
sort(v) Elements of v in order v is a vector
t(a) Transpose a is a matrix
a %*% b Matrix multiplication a, b conforming vector or matrices
Table 5.6: Vector- or matrix-valued functions in the bugs module
Function Canonical Alias Compatible
name with
Arc-cosine arccos acos R
Hyperbolic arc-cosine arccosh acosh R
Arc-sine arcsin asin R
Hyperbolic arc-sine arcsinh asinh R
Arc-tangent arctan atan R
Table 5.7: Functions with aliases in bugs module
27
Chapter 6
Distributions
Distributions are used to dene stochastic nodes using the ~ operator. The distributions
dened in the bugs module are listed in table 6.1 (real-valued distributions), 6.2 (discrete-
valued distributions), and 6.3 (multivariate distributions).
Some distributions have restrictions on the valid parameter values, and these are indicated
in the tables. If a Distribution is given invalid parameter values when evaluating the log-
likelihood, it returns . When a model is initialized, all stochastic nodes are checked to
ensure that the initial parameter values are valid for their distribution.
6.1 Distribution aliases
A distribution may optionally have an alias, which can be used in the model denition in
place of the canonical name. Aliases are used to to avoid confusion with other statistical
software in which distributions may have dierent names. Table 6.4 shows the distributions
in the bugs module with an alias.
28
Name Usage Density Lower Upper
Beta dbeta(a,b)
x
a1
(1 x)
b1
(a, b)
0 1
a > 0, b > 0
Chi-square dchisqr(k)
x
k
2
1
exp(x/2)
2
k
2
(
k
2
)
0
k > 0
Double ddexp(mu,tau)
exp(|x |)/2
exponential > 0
Exponential dexp(lambda)
exp(x)
0
> 0
F df(n,m)
(
n+m
2
)
(
n
2
)(
m
2
)
_
n
m
_n
2
x
n
2
1
_
1 +
nx
m
_
(n+m)
2
0
n > 0, m > 0
Gamma dgamma(r, lambda)
r
x
r1
exp(x)
(r)
0
> 0, r > 0
Generalized dgen.gamma(r,lambda,b)
b
br
x
br1
exp{(x)
b
}
(r)
0
gamma > 0, b > 0, r > 0
Logistic dlogis(mu, tau) exp{(x )}
[1 + exp{(x )}]
2
> 0
Log-normal dlnorm(mu,tau) _
2
_1
2
x
1
exp
_
(log(x) )
2
/2
_ 0
> 0
Noncentral dnchisqr(k, delta)
r=0
exp(
2
)(
2
)
r
r!
x
(k/2+r1)
exp(
x
2
)
2
(k/2+r)
(
k
2
+r)
0
Chi-squre k > 0, 0
Normal dnorm(mu,tau) _
2
_1
2
exp{(x )
2
/2}
> 0
Pareto dpar(alpha, c)
c
x
(+1)
c
> 0, c > 0
Student t dt(mu,tau,k)
(
k+1
2
)
(
k
2
)
_
k
_1
2
_
1 +
(x)
2
k
_
(k+1)
2
> 0, k > 0
Uniform dunif(a,b)
1
b a
a b
a < b
Weibull dweib(v, lambda)
vx
v1
exp(x
v
)
0
v > 0, > 0
Table 6.1: Univariate real-valued distributions in the bugs module
29
Name Usage Density Lower Upper
Beta dbetabin(a, b, n) _
a+x1
x
__
b+nx1
nx
__
a+b+n1
n
_
1 0 n
binomial a > 0, b > 0, n N
Bernoulli dbern(p)
p
x
(1 p)
1x
0 1
0 < p < 1
Binomial dbin(p,n)
_
n
x
_
p
x
(1 p)
nx
0 n
0 < p < 1, n N
Categorical dcat(pi)
x
i
i
1 N
(R
+
)
N
Noncentral dhyper(n1,n2,m1,psi) (
n
1
x
)(
n
2
m
1
x
)
x
i
(
n
1
i
)(
n
2
m
1
i
)
i
max(0,n
+
m
1
) min(n
1
,m
1
)
hypergeometric 0 n
i
, 0 < m
1
n
+
Negative dnegbin(p, r)
_
x+r1
x
_
p
r
(1 p)
x
0
binomial 0 < p < 1, r N
+
Poisson dpois(lambda)
exp()
x
x!
0
> 0
Table 6.2: Discrete univariate distributions in the bugs module
Name Usage Density
Dirichlet p ~ ddirch(alpha)
(
i
i
)
j
p
j
1
j
(
j
)
j
0
Multivariate x ~ dmnorm(mu,Omega)
_
||
2
_1
2
exp{(x )
T
(x )/2}
normal positive denite
Wishart Omega ~ dwish(R,k) ||
(kp1)/2
|R|
k/2
exp{Tr(R/2)}
2
pk/2
p
(k/2) R p p pos. def., k p
Multivariate x ~ dmt(mu,Omega,k) {(k + p)/2}
(k/2)(n)
p/2
||
1/2
_
1 +
1
k
(x )
T
(x )
_
(k+p)
2
Student t pos. def.
Multinomial x ~ dmulti(pi, n)
n!
x
j
j
x
j
!
j
x
j
= n
Table 6.3: Multivariate distributions in the bugs module
Distribution Canonical Alias Compatibile
name with
Binomial dbin dbinom R
Chi-square dchisqr dchisq R
Negative binomial dnegbin dnbinom R
Weibull dweib dweibull R
Dirichlet ddirch ddirich OpenBUGS
Table 6.4: Distributions with aliases in bugs module
30
Chapter 7
Dierences between JAGS and
OpenBUGS
Although JAGS aims for the same functionality as OpenBUGS, there are a number of important
dierences.
7.0.1 Data format
There is no need to transpose matrices and arrays when transferring data between R and JAGS,
since JAGS stores the values of an array in column major order, like R and FORTRAN (i.e.
lling the left-hand index rst).
If you have an S-style data le for OpenBUGS and you wish to convert it for JAGS, then
use the command bugs2jags, which is supplied with the coda package.
7.0.2 Distributions
Structural zeros are allowed in the Dirichlet distribution. If
p ~ ddirch(alpha)
and some of the elements of alpha are zero, then the corresponding elements of p will be xed
to zero.
The Multinomial (dmulti) and Categorical (dcat) distributions, which take a vector of
probabilities as a parameter, may use unnormalized probabilities. The probability vector is
normalized internally so that
p
i
p
i
j
p
j
The non-central hypergeometric distribution (dhyper) uses the same parameterization as
R, which is dierent from the parameterization used in OpenBUGS 3.2.2. OpenBUGS is
parameterized as
X ~ dhyper(n, m, N, psi) #OpenBUGS
where n, m, N are the following table margins:
x - n
- - -
m - N
31
This parameterization is symmetric in n, m. In JAGS, dhyper is parameterized as
X ~ dhyper(n1, n2, m1, psi) #JAGS
where n1, n2, m1 are
x - m1
- - -
n1 n2 -
7.0.3 Observable Functions
Logical nodes in the BUGS language are a convenient way of describing the relationships
between observables (constant and stochastic nodes), but are not themselves observable. You
cannot supply data values for a logical node.
This restriction can occasionally be inconvenient, as there are important cases where the
data are a deterministic function of unobserved variables. Three important examples are
1. Censored data, which commonly occurs in survival analysis. In the most general case,
we know that unobserved failure time T lies in the interval (L, U].
2. Rounded data, when there is a discrete underlying distribution but the measurements
are rounded to a xed number of decimal places.
3. Aggregate data when we observe the sum of two or more unobserved variables.
JAGS contains novel distributions to handle these situations.
Interval censored data: dinterval
t <- c(1.2, 4.7, NA, NA, 3.2)
is.censored <- c(0, 0, 1, 1, 0)
cutpoint <- 5
The dinterval distribution represents interval-censored data. It has two parameters: t
the original continuous variable, and c[], a vector of cut points of length M, say. If Y
dinterval(t, c) then
Y = 0 if t c[1]
Y = m if c[m] < t c[m + 1] for 1 m < M
Y = M if c[M] < t.
Rounded data: dround
The dround distribution represents rounded data. It has two parameters: t the original
continuous variable and d, the number of decimal places to which the measurements are
rounded. Thus if t = 1.2345 and d = 2 then the rounded value is 1.23. Note that d can be
negative: if d = 2 then the data are rounded to the nearest 100.
32
Summed data: dsum
The dsum distribution represents the sum of two or more variables. It takes a variable number
of parameters. If Y dsum(x1,x2,x3) then Y = x1 + x2 + x3.
These distributions exist to give a likelihood to data that is, in fact, a deterministic
function of the parameters. The relation
Y ~ dsum(x1, x2)
is logically equivalent to
Y <- x1 + x2
But the latter form does not create a contribution to the likelihood, and does not allow you
to dene Y as data. The likelihood function is trivial: it is 1 if the parameters are consistent
with the data and 0 otherwise. The dsum distribution also requires a special sampler, which
can currently only handle the case where the parameters of dsum are unobserved stochastic
nodes, and where the parameters are either all discrete-valued or all continuous-valued. A
node cannot be subject to more than one dsum constraint.
7.0.4 Data transformations
JAGS allows data transformations, but the syntax is dierent from BUGS. BUGS allows you
to put a stochastic node twice on the left hand side of a relation, as in this example taken
from the manual
for (i in 1:N) {
z[i] <- sqrt(y[i])
z[i] ~ dnorm(mu, tau)
}
This is forbidden in JAGS. You must put data transformations in a separate block of relations
preceded by the keyword data:
data {
for (i in 1:N) {
z[i] <- sqrt(y[i])
}
}
model {
for (i in 1:N) {
z[i] ~ dnorm(mu, tau)
}
...
}
This syntax preserves the declarative nature of the BUGS language. In eect, the data block
denes a distinct model, which describes how the data is generated. Each node in this model
is forward-sampled once, and then the node values are read back into the data table. The
data block is not limited to logical relations, but may also include stochastic relations. You
33
may therefore use it in simulations, generating data from a stochastic model that is dierent
from the one used to analyse the data in the model statement.
This example shows a simple location-scale problem in which the true values of the
parameters mu and tau are generated from a given prior in the data block, and the generated
data is analyzed in the model block.
data {
for (i in 1:N) {
y[i] ~ dnorm(mu.true, tau.true)
}
mu.true ~ dnorm(0,1);
tau.true ~ dgamma(1,3);
}
model {
for (i in 1:N) {
y[i] ~ dnorm(mu, tau)
}
mu ~ dnorm(0, 1.0E-3)
tau ~ dgamma(1.0E-3, 1.0E-3)
}
Beware, however, that every node in the data statement will be considered as data in the sub-
sequent model statement. This example, although supercially similar, has a quite dierent
interpretation.
data {
for (i in 1:N) {
y[i] ~ dnorm(mu, tau)
}
mu ~ dnorm(0,1);
tau ~ dgamma(1,3);
}
model {
for (i in 1:N) {
y[i] ~ dnorm(mu, tau)
}
mu ~ dnorm(0, 1.0E-3)
tau ~ dgamma(1.0E-3, 1.0E-3)
}
Since the names mu and tau are used in both data and model blocks, these nodes will be
considered as observed in the model and their values will be xed at those values generated
in the data block.
7.0.5 Directed cycles
Directed cycles are forbidden in JAGS. There are two important instances where directed
cycles are used in BUGS.
34
Dening autoregressive priors
Dening ordered priors
For the rst case, the GeoBUGS extension to OpenBUGS provides some convenient ways of
dening autoregressive priors. These should be available in a future version of JAGS.
7.0.6 Censoring, truncation and prior ordering
These are three, closely related issues that are all handled using the I(,) construct in BUGS.
Censoring occurs when a variable X is not observed directly, but is observed only to lie
in the range (L, U]. Censoring is an a posteriori restriction of the data, and is represented in
OpenBUGS by the I(,) construct, e.g.
X ~ dnorm(theta, tau) I(L,U)
where L and U are constant nodes.
Truncation occurs when a variable is known a priori to lie in a certain range. Although
BUGS has no construct for representing truncated variables, it turns out that there is no
dierence between censoring and truncation for top-level parameters (i.e. variables with no
unobserved parents). Hence, for example, this
theta ~ dnorm(0, 1.0E-3) I(0, )
is a perfectly valid way to describe a parameter with a half-normal prior distribution.
Prior ordering occurs when a vector of nodes is known a priori to be strictly increasing
or decreasing. It can be represented in OpenBUGS with symmetric I(, ) constructs, e.g.
X[1] ~ dnorm(0, 1.0E-3) I(,X[2])
X[2] ~ dnorm(0, 1.0E-3) I(X[1],)
ensures that X[1] X[2].
JAGS makes an attempt to separate these three concepts.
Censoring is handled in JAGS using the new distribution dinterval (section 7.0.3). This
can be illustrated with a survival analysis example. A right-censored survival time t
i
with a
Weibull distribution is described in OpenBUGS as follows:
t[i] ~ dweib(r, mu[i]) I(c[i], )
where t
i
is unobserved if t
i
> c
i
. In JAGS this becomes
is.censored[i] ~ dinterval(t[i], c[i])
t[i] ~ dweib(r, mu[i])
where is.censored[i] is an indicator variable that takes the value 1 if t
i
is censored and 0
otherwise. See the MICE and KIDNEY examples in the classic bugs set of examples.
Truncation is represented in JAGS using the T(,) construct, which has the same syntax
as the I(,) construct in OpenBUGS, but has a dierent interpretation. If
X ~ dfoo(theta) T(L,U)
35
then a priori X is known to lie between L and U. This generates a likelihood
p(x | )
P(L X U | )
if L X U and zero otherwise, where p(x | ) is the density of X given according
to the distribution foo. Note that calculation of the denominator may be computationally
expensive.
For compatibility with OpenBUGS, JAGS permits the use of I(,) for truncation when
the the parameters of the truncated distribution are xed. For example, this is permitted:
mu ~ dnorm(0, 1.0E-3) I(0, )
because the truncated distribution has xed parameters (mean 0, precision 1.0E-3). In this
case, there is no dierence between censoring and truncation. Conversely, this is not permit-
ted:
for (i in 1:N) {
x[i] ~ dnorm(mu, tau) I(0, )
}
mu ~ dnorm(0, 1.0E-3)
tau ~ dgamma(1, 1)
}
because the mean and precision of x
1
. . . x
N
are parameters to be estimated. JAGS does not
know if the aim is to model truncation or censoring and so the compiler will reject the model.
Use either T(,) or the dinterval distribution to resolve the ambiguity.
Prior ordering of top-level parameters in the model can be achieved using the sort func-
tion, which sorts a vector in ascending order.
Symmetric truncation relations like this
alpha[1] ~ dnorm(0, 1.0E-3) I(,alpha[2])
alpha[2] ~ dnorm(0, 1.0E-3) I(alpha[1],alpha[3])
alpha[3] ~ dnorm(0, 1.0E-3) I(alpha[2],)
Should be replaced by this
for (i in 1:3) {
alpha0[i] ~ dnorm(0, 1.0E-3)
}
alpha[1:3] <- sort(alpha0)
36
Chapter 8
Feedback
Please send feedback to martyn_plummer@users.sourceforge.net. I am particularly inter-
ested in the following problems:
Crashes, including both segmentation faults and uncaught exceptions.
Incomprehensible error messages
Models that should compile, but dont
Output that cannot be validated against OpenBUGS
Documentation erors
If you want to send a bug report, it must be reproducible. Send the model le, the data
le, the initial value le and a script le that will reproduce the problem. Describe what you
think should happen, and what did happen.
37
Chapter 9
Acknowledgments
Many thanks to the BUGS development team, without whom JAGS would not exist. Thanks
also to Simon Frost for pioneering JAGS on Windows and Bill Northcott for getting JAGS
on Mac OS X to work. Kostas Oikonomou found many bugs while getting JAGS to work
on Solaris using Sun development tools and libraries. Bettina Gruen, Chris Jackson, Greg
Ridgeway and Geo Evans also provided useful feedback. Special thanks to Jean-Baptiste
Denis who has been very diligent in providing feedback on JAGS.
38
Appendix A
Data format
The le format used by JAGS for representing data and initial values is the dump() format
used by R. This format is valid R code, so a JAGS data le can be read into R using the
source() function. Since R is a functional language, the code consists of a sequence of
assignments and function calls.
Assignments are represented by a left arrow (<-) with the variable name on the left hand
side and a numeric value or function call on the right hand side. The variable name
may optionally be enclosed in single quotes, double quotes, or back-ticks.
Function calls are represented by the function name, followed by a list of comma-separated
arguments inside round brackets. Optional arguments need to be tagged (i.e. given in
the form tag=value).
The JAGS parser ignores all white space. Long expressions can therefore be split over several
lines.
Scalar values are represented by a simple assignment statement
theta <- 0.1243
All numeric values are read in to JAGS as doubles, even if they are represented as integers
(i.e. 12 and 12.0 are equivalent). Engineering notation may be used to represent large or
small values (e.g 1.5e-3 is equivalent to 0.0015).
Vectors are denoted by the R collection function c, which takes a number of arguments
equal to the length of the vector. The following code denotes a vector of length 4:
x <- c(2, 3.5, 1.3e-2, 88)
There is no distinction between row and column-vectors.
Matrices and higher-dimensional arrays in R are created by adding a dimension attribute
(.Dim) to a vector. In the R dump format this is done using the structure function. The
rst argument to the structure function is a vector of numeric values of the matrix, given
in column major order (i.e. lling the left index rst). The second argument, which must
be given the tag .Dim, is the number of dimensions of the array, represented by a vector of
integer values. For example, if the matrix A takes values
_
_
1 4
2 5
3 6
_
_
39
it is represented as
A <- structure(c(1, 2, 3, 4, 5, 6), .Dim=c(3,2))
The simplest way to prepare your data is to read them into R and then dump them. Only
numeric vectors, matrices and arrays are allowed. More complex data structures such as
factors, lists and data frames cannot be parsed by JAGS nor can non-numeric vectors. Any
R attributes of the data (such as names and dimnames) are ignored when they are read into
JAGS.
40
Bibliography
[1] James Albert and Siddharta Chib. Bayesian analysis of binary and polychotomous re-
sponse data. Journal of the American Statistical Association, 88:669679, 1993.
[2] Gilles Celeux, Merrilee Hurn, and Christian P. Robert. Computational and inferential
diculties with mixture posterior distributions. Journal of the American Statistical
Association, 95:957970, 1999.
[3] T A Davis and W W Hager. Modifying a sparse cholesky factorization. SIAM Journal
on Matrix Analysis and Applications, 20:606627, 1999.
[4] Timothy A Davis. Direct Methods for Sparse Linear Systems. SIAM, 2006.
[5] Sylvia Fr uhwirth-Schnatter, Rudolf Fr uhwirth, Leonhard Held, and Havard Rue. Im-
proved auxiliary mixture sampling for hierarchical models of non-gaussian data. Statistics
and Computing, 19(4):479492, 2009.
[6] Dani Gamerman. Ecient sampling from the posterior distribution in generalized linear
mixed models. Statistics and Computing, 7:5768, 1997.
[7] Chris Holmes and Leonard Held. Bayesian auxiliary variable models for binary and
multinomial regression. Bayesian Analysis, 1(1):145168, 2006.
[8] Radford Neal. Sampling from multimodal distributions using tempered transitions.
Statistics and Computing, 6:353366, 1994.
[9] M Plummer. Discussion of the paper by Spiegelhalter et al. Journal of the Royal Statis-
tical Society Series B, 64:620, 2002.
[10] M Plummer. Penalized loss functions for Bayesian model comparison. Biostatistics,
9(3):523539, 2008.
[11] DJ Spiegelhalter, NG Best, BP Carlin, and A van der Linde. Bayesian measures of model
complexity and t (with discussion). Journal of the Royal Statistical Societey Series B,
64:583639, 2002.
41