Latin hypercube sampling (LHS) is a statistical method for generating a near-random sample of parameter values from a multidimensional distribution. The sampling method is often used to construct computer experiments or for Monte Carlo integration.[1]
LHS was described by Michael McKay of Los Alamos National Laboratory in 1979.[1] An equivalent technique was independently proposed by Vilnis Eglājs in 1977.[2] It was further elaborated by Ronald L. Iman and coauthors in 1981.[3] Detailed computer codes and manuals were later published.[4]
In the context of statistical sampling, a square grid containing sample positions is a Latin square if (and only if) there is only one sample in each row and each column. A Latin hypercube is the generalisation of this concept to an arbitrary number of dimensions, whereby each sample is the only one in each axis-aligned hyperplane containing it.[1]
When sampling a function of variables, the range of each variable is divided into equally probable intervals. sample points are then placed to satisfy the Latin hypercube requirements; this forces the number of divisions, , to be equal for each variable. This sampling scheme does not require more samples for more dimensions (variables); this independence is one of the main advantages of this sampling scheme. Another advantage is that random samples can be taken one at a time, remembering which samples were taken so far.
In two dimensions the difference between random sampling, Latin hypercube sampling, and orthogonal sampling can be explained as follows:
Thus, orthogonal sampling ensures that the set of random numbers is a very good representative of the real variability, LHS ensures that the set of random numbers is representative of the real variability whereas traditional random sampling (sometimes called brute force) is just a set of random numbers without any guarantees.