The set of all possible states that a system can be in defines its state space. An essential component of cybernetic modelling is a quantitative measure for the size of that state space, or the number of distinct states. This measure is called variety. Variety represents the freedom the system has in choosing a particular state, and thus the uncertainty we have about which state the system occupies. Variety V is defined as the number of elements in the state space S, or, more commonly, as the logarithm to the basis two of that number:
V = log2 (|S|)
The unit of variety in the logarithmic form is the bit. A variety of one bit, V=1, means that the system has two possible states, that is, one difference or distinction. In the simplest case of n binary variables, V = log2(2n) = n is therefore equal to the minimal number of independent dimensions.
Background
Variety has always been a fundamental idea in Cybernetics and Systems Science, and is
so in Metasystem Transition Theory. Variety is defined as a multiplicity of distinctions. The
existence of variety is necessary for all change, choice, and information. A reduction in the quantity of variety is the process of
selection. If variety has thus been reduced, i.e. if actual variety is less than potential variety, then we say that there is constraint.
Frequently the quantity of variety and the change in the quantity of
variety (positive increase or negative decrease) is critical to understand
system evolution. Where variety is manifest in a process, then we sometimes
want to say that there is uncertainty about the outcome of the
process; when that uncertainty is relieved but the occurrence of one of the
possibilities, then we gain information. The are many possible ways
to measure the quantity of variety, uncertainty, or information. As defined above, the simplest is the count of the number of distinct states.
More useful can be the logarithm of that number as a quantity of
information, which is called the Hartley entropy. When sets and subsets of
distinctions are considered, possibilistic nonspecificities result . The most celebrated are the stochastic entropies of
classical information theory, which result from applying probabilistic
distributions to the various distinctions.