Slice sampling is a type of Markov chain Monte Carlo algorithm used to draw pseudo-random samples from a statistical distribution. The method is based on the observation that to sample a random variable one can sample uniformly from the region under the graph of its density function.

Motivation

Suppose you want to sample some random variable X with distribution f(x). Suppose that the following is the graph of f(x). The height of f(x) corresponds to the likelihood at that point.

350px

If you were to uniformly sample X, each value would have the same likelihood of being sampled, and your distribution would be of the form f(x)=y for some y value instead of some non-uniform function f(x). Instead of the original black line, your new distribution would look more like the blue line.

350px

In order to sample X in a manner which will retain the distribution f(x), some sampling technique must be used which takes into account the varied likelihoods for each range of f(x).

Compared to Other Methods

Slice sampling is a Markov chain method and as such serves the same purpose as Gibbs sampling and Metropolis. Unlike Metropolis, there is no need to manually tune the candidate function or candidate standard deviation.

Recall that Metropolis is sensitive to step size. If the step size is too small random walk causes slow decorrelation. If the step size is too large there is great inefficiency due to a high rejection rate.

In contrast to Metropolis, slice sampling automatically adjusts the step size to match the local shape of the density function. Implementation is arguably easier and more efficient than Gibbs sampling or simple Metropolis updates.

Note that, in contrast to many available methods for generating random numbers from non-uniform distributions, random variates generated directly by this approach will exhibit serial statistical dependence. In other words, not all points have the same independent likelihood of selection. This is a desired behavior, as the sampling likelihood should follow the properties of the sampled distribution (i.e. x value ranges in f(x) with a greater area under the curve should have a higher likelihood of selection over value ranges with a lesser area under the curve).

Slice Sampling requires that the distribution to be sampled be evaluable. One way to relax this requirement is to substitute an evaluable distribution which is proportional to the true unevaluable distribution.

Univariate Case

thumb|350pxFor a given sample x, a value for y is chosen from [0, f(x)], which defines a "slice" of the distribution (shown by the solid horizontal line). In this case, there are two slices separated by an area outside the range of the distribution. To sample a random variable X with density f(x) we introduce an auxiliary variable Y and iterate as follows:

  • Given a sample x we choose y uniformly at random from the interval [0, f(x)];
  • given y we choose x uniformly at random from the set $y^{-1}[y, +\infty)$.
  • The sample of x is obtained by ignoring the y values.

Our auxiliary variable Y represents a horizontal “slice” of the distribution. The rest of each iteration is dedicated to sampling an x value from the slice which is representative of the density of the region being considered.

In practice, sampling from a horizontal slice of a multimodal distribution is difficult. There is a tension between obtaining a large sampling region and thereby making possible large moves in the distribution space, and obtaining a simpler sampling region to increase efficiency. One option for simplifying this process is regional expansion and contraction.

  • First, a width parameter w is used to define the area containing the given x value. Each endpoint of this area is tested to see if it lies outside the given slice. If not, the region is extended in the appropriate direction(s) by w until the end both endpoints lie outside the slice.
  • A candidate sample is selected uniformly from within this region. If the candidate sample lies inside of the slice, then it is accepted as the new sample. If it lies outside of the slice, the candidate point becomes the new boundary for the region. A new candidate sample is taken uniformly. The process repeats until the candidate sample is within the slice. (See diagram for a visual example).

500px-slice.png Finding a sample given a set of slices (the slices are represented here as blue lines and correspond to the solid line slices in the previous graph of f(x) ). a) A width parameter w is set. b) A region of width w is identified around a given point $x_0$. c) The region is expanded by w until both endpoints are outside of the considered slice. d) $x_1$ is selected uniformly from the region. e) Since $x_1$ lies outside the considered slice, the region's left bound is adjusted to $x_1$. f) Another uniform sample $x$ is taken and accepted as the sample since it lies within the considered slice.

Multivariate Methods

Treating each variable independently

Single variable slice sampling can be used in the multivariate case by sampling each variable in turn repeatedly, as in Gibbs sampling. To do so requires that we can compute, for each component $x_i$ a function that is proportional to $p(x_i|x_0\ldotsx_n)$.

To prevent random walk behavior, overrelaxation methods can be used to update each variable in turn. Overrelaxation chooses a new value on the opposite side of the mode from the current value, as opposed to choosing a new independent value from the distribution as done in Gibbs.

Hyperrectangle slice sampling

This method adapts the univariate algorithm to the multivariate case by substituting a hyperrectangle for the one-dimensional w region used in the original. The hyperrectangle H is initialized to a random position over the slice. H is then shrunken as points from it are rejected.

Reflective slice sampling

Reflective slice sampling is a technique to suppress random walk behavior in which the successive candidate samples of distribution f(x) are kept within the bounds of the slice by “reflecting” the direction of sampling inward toward the slice once the boundary has been hit.

In this graphical representation of reflective sampling, the shape indicates the bounds of a sampling slice. The dots indicate start and stopping points of a sampling walk. When the samples hit the bounds of the slice, the direction of sampling is “reflected” back into the slice.

350px|alt=alt text

Example

Let us consider a single variable example. Suppose our true distribution $g(x)~N(0,5)$. So: $f(x) = \frac{1}{\sqrt{2\pi*5^2}} * e^{ -\frac{(x-0)^2}{2*5^2} }$

  • We first draw a uniform random value y from the range of f(x) in order to define our slice(es). Suppose y=0.01.
  • Next, we set our width parameter w which we will use to expand our region of consideration. Suppose w=2.
  • Next, we need an initial value for x. We draw x from the uniform distribution within the domain of f(x) at our current y. Suppose x=2.
  • Because x=2 and w=2, our current region of interest is bounded by (1,3).
  • Now, each endpoint of this area is tested to see if it lies outside the given slice. Our right bound lies outside our slice, but the left value does not. We expand the left bound by adding w to it until it extends past the limit of the slice. After this process, the new bounds of our region of interest are (-4,3).
  • Next, we take a uniform sample within (-4,3). Suppose this sample yields x=-3.9. Though this sample is within our region of interest, it does not lie within our slice, so we modify the left bound of our region of interest to this point. Now we take a uniform sample from (-3.9,3). This time our sample yields x=1, which is within our slice, and thus is our accepted sample. Had our new x not been within our slice, we would continue the shrinking/resampling process until a valid x within bounds is found.

Candidate Question

Explain how a slice sampling algorithm would work. Assume the univariate case using an overrelaxation update method. Be explicit in how the overrelaxation method would work.

See also

References

  • Neal, Radford M. (2003). “Slice Sampling”, Annals of Statistics, 31(3).
  • Bishop, Christopher (2006). Pattern Recognition and Machine Learning, Springer, 545-547.

Markov chain Monte Carlo Non-uniform random numbers

cs-677sp2010/slice-sampling.txt · Last modified: 2014/12/11 16:49 by ryancha
Back to top
CC Attribution-Share Alike 4.0 International
chimeric.de = chi`s home Valid CSS Driven by DokuWiki do yourself a favour and use a real browser - get firefox!! Recent changes RSS feed Valid XHTML 1.0