In the
design of experiments
The design of experiments (DOE), also known as experiment design or experimental design, is the design of any task that aims to describe and explain the variation of information under conditions that are hypothesized to reflect the variation. ...
, optimal experimental designs (or optimum designs) are a class of
experimental designs that are
optimal with respect to some
statistical
Statistics (from German language, German: ', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a s ...
criterion. The creation of this field of statistics has been credited to Danish statistician
Kirstine Smith.
In the
design of experiments
The design of experiments (DOE), also known as experiment design or experimental design, is the design of any task that aims to describe and explain the variation of information under conditions that are hypothesized to reflect the variation. ...
for
estimating statistical model
A statistical model is a mathematical model that embodies a set of statistical assumptions concerning the generation of Sample (statistics), sample data (and similar data from a larger Statistical population, population). A statistical model repre ...
s, optimal designs allow parameters to be
estimated without bias and with
minimum variance. A non-optimal design requires a greater number of
experimental runs to
estimate
Estimation (or estimating) is the process of finding an estimate or approximation, which is a value that is usable for some purpose even if input data may be incomplete, uncertain, or unstable. The value is nonetheless usable because it is de ...
the
parameters with the same
precision as an optimal design. In practical terms, optimal experiments can reduce the costs of experimentation.
The optimality of a design depends on the
statistical model
A statistical model is a mathematical model that embodies a set of statistical assumptions concerning the generation of Sample (statistics), sample data (and similar data from a larger Statistical population, population). A statistical model repre ...
and is assessed with respect to a statistical criterion, which is related to the variance-matrix of the estimator. Specifying an appropriate model and specifying a suitable criterion function both require understanding of
statistical theory
The theory of statistics provides a basis for the whole range of techniques, in both study design and data analysis, that are used within applications of statistics.
The theory covers approaches to statistical-decision problems and to statistica ...
and practical knowledge with
designing experiments.
Advantages
Optimal designs offer three advantages over sub-optimal
experimental designs:
#Optimal designs reduce the costs of experimentation by allowing
statistical model
A statistical model is a mathematical model that embodies a set of statistical assumptions concerning the generation of Sample (statistics), sample data (and similar data from a larger Statistical population, population). A statistical model repre ...
s to be estimated with fewer experimental runs.
#Optimal designs can accommodate multiple types of factors, such as process, mixture, and discrete factors.
#Designs can be optimized when the design-space is constrained, for example, when the mathematical process-space contains factor-settings that are practically infeasible (e.g. due to safety concerns).
Minimizing the variance of estimators
Experimental designs are evaluated using statistical criteria.
It is known that the
least squares
The method of least squares is a mathematical optimization technique that aims to determine the best fit function by minimizing the sum of the squares of the differences between the observed values and the predicted values of the model. The me ...
estimator minimizes the
variance
In probability theory and statistics, variance is the expected value of the squared deviation from the mean of a random variable. The standard deviation (SD) is obtained as the square root of the variance. Variance is a measure of dispersion ...
of
mean
A mean is a quantity representing the "center" of a collection of numbers and is intermediate to the extreme values of the set of numbers. There are several kinds of means (or "measures of central tendency") in mathematics, especially in statist ...
-
unbiased
Bias is a disproportionate weight ''in favor of'' or ''against'' an idea or thing, usually in a way that is inaccurate, closed-minded, prejudicial, or unfair. Biases can be innate or learned. People may develop biases for or against an individ ...
estimators
In statistics, an estimator is a rule for calculating an estimate of a given quantity based on observed data: thus the rule (the estimator), the quantity of interest (the estimand) and its result (the estimate) are distinguished. For example, the ...
(under the conditions of the
Gauss–Markov theorem
In statistics, the Gauss–Markov theorem (or simply Gauss theorem for some authors) states that the ordinary least squares (OLS) estimator has the lowest sampling variance within the class of linear unbiased estimators, if the errors in ...
). In the
estimation
Estimation (or estimating) is the process of finding an estimate or approximation, which is a value that is usable for some purpose even if input data may be incomplete, uncertain, or unstable. The value is nonetheless usable because it is d ...
theory for
statistical model
A statistical model is a mathematical model that embodies a set of statistical assumptions concerning the generation of Sample (statistics), sample data (and similar data from a larger Statistical population, population). A statistical model repre ...
s with one
real parameter
A parameter (), generally, is any characteristic that can help in defining or classifying a particular system (meaning an event, project, object, situation, etc.). That is, a parameter is an element of a system that is useful, or critical, when ...
, the
reciprocal of the variance of an (
"efficient") estimator is called the "
Fisher information" for that estimator. Because of this reciprocity, ''minimizing'' the ''variance'' corresponds to ''maximizing'' the ''information''.
When the
statistical model
A statistical model is a mathematical model that embodies a set of statistical assumptions concerning the generation of Sample (statistics), sample data (and similar data from a larger Statistical population, population). A statistical model repre ...
has several
parameter
A parameter (), generally, is any characteristic that can help in defining or classifying a particular system (meaning an event, project, object, situation, etc.). That is, a parameter is an element of a system that is useful, or critical, when ...
s, however, the
mean
A mean is a quantity representing the "center" of a collection of numbers and is intermediate to the extreme values of the set of numbers. There are several kinds of means (or "measures of central tendency") in mathematics, especially in statist ...
of the parameter-estimator is a
vector
Vector most often refers to:
* Euclidean vector, a quantity with a magnitude and a direction
* Disease vector, an agent that carries and transmits an infectious pathogen into another living organism
Vector may also refer to:
Mathematics a ...
and its
variance
In probability theory and statistics, variance is the expected value of the squared deviation from the mean of a random variable. The standard deviation (SD) is obtained as the square root of the variance. Variance is a measure of dispersion ...
is a
matrix
Matrix (: matrices or matrixes) or MATRIX may refer to:
Science and mathematics
* Matrix (mathematics), a rectangular array of numbers, symbols or expressions
* Matrix (logic), part of a formula in prenex normal form
* Matrix (biology), the m ...
. The
inverse matrix of the variance-matrix is called the "information matrix". Because the variance of the estimator of a parameter vector is a matrix, the problem of "minimizing the variance" is complicated. Using
statistical theory
The theory of statistics provides a basis for the whole range of techniques, in both study design and data analysis, that are used within applications of statistics.
The theory covers approaches to statistical-decision problems and to statistica ...
, statisticians compress the information-matrix using real-valued
summary statistics
In descriptive statistics, summary statistics are used to summarize a set of observations, in order to communicate the largest amount of information as simply as possible. Statisticians commonly try to describe the observations in
* a measure of ...
; being real-valued functions, these "information criteria" can be maximized. The traditional optimality-criteria are
invariants of the
information
Information is an Abstraction, abstract concept that refers to something which has the power Communication, to inform. At the most fundamental level, it pertains to the Interpretation (philosophy), interpretation (perhaps Interpretation (log ...
matrix; algebraically, the traditional optimality-criteria are
functionals of the
eigenvalue
In linear algebra, an eigenvector ( ) or characteristic vector is a vector that has its direction unchanged (or reversed) by a given linear transformation. More precisely, an eigenvector \mathbf v of a linear transformation T is scaled by a ...
s of the information matrix.
*A-optimality ("average" or trace)
**One criterion is A-optimality, which seeks to minimize the
trace of the
inverse of the information matrix. This criterion results in minimizing the average variance of the estimates of the regression coefficients.
*C-optimality
**This criterion minimizes the variance of a
best linear unbiased estimator of a predetermined linear combination of model parameters.
*D-optimality (determinant)
**A popular criterion is D-optimality, which seeks to minimize , (X'X)
−1, , or equivalently maximize the
determinant
In mathematics, the determinant is a Scalar (mathematics), scalar-valued function (mathematics), function of the entries of a square matrix. The determinant of a matrix is commonly denoted , , or . Its value characterizes some properties of the ...
of the
information matrix X'X of the design. This criterion results in maximizing the
differential Shannon information content of the parameter estimates.
*E-optimality (eigenvalue)
**Another design is E-optimality, which maximizes the minimum
eigenvalue
In linear algebra, an eigenvector ( ) or characteristic vector is a vector that has its direction unchanged (or reversed) by a given linear transformation. More precisely, an eigenvector \mathbf v of a linear transformation T is scaled by a ...
of the information matrix.
*S-optimality
**This criterion maximizes a quantity measuring the mutual column orthogonality of X and the
determinant
In mathematics, the determinant is a Scalar (mathematics), scalar-valued function (mathematics), function of the entries of a square matrix. The determinant of a matrix is commonly denoted , , or . Its value characterizes some properties of the ...
of the information matrix.
*T-optimality
**This criterion maximizes the discrepancy between two proposed models at the design locations.
Other optimality-criteria are concerned with the variance of
predictions:
*G-optimality
**A popular criterion is G-optimality, which seeks to minimize the maximum entry in the
diagonal
In geometry, a diagonal is a line segment joining two vertices of a polygon or polyhedron, when those vertices are not on the same edge. Informally, any sloping line is called diagonal. The word ''diagonal'' derives from the ancient Greek � ...
of the
hat matrix X(X'X)
−1X'. This has the effect of minimizing the maximum variance of the predicted values.
*I-optimality (integrated)
**A second criterion on prediction variance is I-optimality, which seeks to minimize the average prediction variance ''over the design space''.
*V-optimality (variance)
**A third criterion on prediction variance is V-optimality, which seeks to minimize the average prediction variance over a set of m specific points.
Contrasts
In many applications, the statistician is most concerned with a
"parameter of interest" rather than with
"nuisance parameters". More generally, statisticians consider
linear combination
In mathematics, a linear combination or superposition is an Expression (mathematics), expression constructed from a Set (mathematics), set of terms by multiplying each term by a constant and adding the results (e.g. a linear combination of ''x'' a ...
s of parameters, which are estimated via linear combinations of treatment-means in the
design of experiments
The design of experiments (DOE), also known as experiment design or experimental design, is the design of any task that aims to describe and explain the variation of information under conditions that are hypothesized to reflect the variation. ...
and in the
analysis of variance
Analysis of variance (ANOVA) is a family of statistical methods used to compare the Mean, means of two or more groups by analyzing variance. Specifically, ANOVA compares the amount of variation ''between'' the group means to the amount of variati ...
; such linear combinations are called
contrasts. Statisticians can use appropriate optimality-criteria for such
parameters of interest and for
contrasts.
Implementation
Catalogs of optimal designs occur in books and in software libraries.
In addition, major
statistical systems like
SAS and
R have procedures for optimizing a design according to a user's specification. The experimenter must specify a
model
A model is an informative representation of an object, person, or system. The term originally denoted the plans of a building in late 16th-century English, and derived via French and Italian ultimately from Latin , .
Models can be divided in ...
for the design and an optimality-criterion before the method can compute an optimal design.
Practical considerations
Some advanced topics in optimal design require more
statistical theory
The theory of statistics provides a basis for the whole range of techniques, in both study design and data analysis, that are used within applications of statistics.
The theory covers approaches to statistical-decision problems and to statistica ...
and practical knowledge in designing experiments.
Model dependence and robustness
Since the optimality criterion of most optimal designs is based on some function of the information matrix, the 'optimality' of a given design is ''
model
A model is an informative representation of an object, person, or system. The term originally denoted the plans of a building in late 16th-century English, and derived via French and Italian ultimately from Latin , .
Models can be divided in ...
dependent'': While an optimal design is best for that
model
A model is an informative representation of an object, person, or system. The term originally denoted the plans of a building in late 16th-century English, and derived via French and Italian ultimately from Latin , .
Models can be divided in ...
, its performance may deteriorate on other
models
A model is an informative representation of an object, person, or system. The term originally denoted the plans of a building in late 16th-century English, and derived via French and Italian ultimately from Latin , .
Models can be divided int ...
. On other
models
A model is an informative representation of an object, person, or system. The term originally denoted the plans of a building in late 16th-century English, and derived via French and Italian ultimately from Latin , .
Models can be divided int ...
, an ''optimal'' design can be either better or worse than a non-optimal design. Therefore, it is important to
benchmark the performance of designs under alternative
models
A model is an informative representation of an object, person, or system. The term originally denoted the plans of a building in late 16th-century English, and derived via French and Italian ultimately from Latin , .
Models can be divided int ...
.
Choosing an optimality criterion and robustness
The choice of an appropriate optimality criterion requires some thought, and it is useful to benchmark the performance of designs with respect to several optimality criteria. Cornell writes that
Indeed, there are several classes of designs for which all the traditional optimality-criteria agree, according to the theory of "universal optimality" of
Kiefer. The experience of practitioners like Cornell and the "universal optimality" theory of Kiefer suggest that robustness with respect to changes in the ''optimality-criterion'' is much greater than is robustness with respect to changes in the ''model''.
Flexible optimality criteria and convex analysis
High-quality statistical software provide a combination of libraries of optimal designs or iterative methods for constructing approximately optimal designs, depending on the model specified and the optimality criterion. Users may use a standard optimality-criterion or may program a custom-made criterion.
All of the traditional optimality-criteria are
convex (or concave) functions, and therefore optimal-designs are amenable to the mathematical theory of
convex analysis
Convex analysis is the branch of mathematics devoted to the study of properties of convex functions and convex sets, often with applications in convex optimization, convex minimization, a subdomain of optimization (mathematics), optimization theor ...
and their computation can use specialized methods of
convex minimization. The practitioner need not select ''exactly one'' traditional, optimality-criterion, but can specify a custom criterion. In particular, the practitioner can specify a convex criterion using the maxima of convex optimality-criteria and
nonnegative combinations of optimality criteria (since these operations preserve
convex functions). For ''convex'' optimality criteria, the
Kiefer-
Wolfowitzbr>
equivalence theoremallows the practitioner to verify that a given design is globally optimal. The
Kiefer-
Wolfowitzbr>
equivalence theoremis related with the
Legendre-
Fenchel conjugacy for
convex function
In mathematics, a real-valued function is called convex if the line segment between any two distinct points on the graph of a function, graph of the function lies above or on the graph between the two points. Equivalently, a function is conve ...
s.
If an optimality-criterion lacks
convexity, then finding a
global optimum and verifying its optimality often are difficult.
Model uncertainty and Bayesian approaches
Model selection
When scientists wish to test several theories, then a statistician can design an experiment that allows optimal tests between specified models. Such "discrimination experiments" are especially important in the
biostatistics
Biostatistics (also known as biometry) is a branch of statistics that applies statistical methods to a wide range of topics in biology. It encompasses the design of biological experiments, the collection and analysis of data from those experimen ...
supporting
pharmacokinetics
Pharmacokinetics (from Ancient Greek ''pharmakon'' "drug" and ''kinetikos'' "moving, putting in motion"; see chemical kinetics), sometimes abbreviated as PK, is a branch of pharmacology dedicated to describing how the body affects a specific su ...
and
pharmacodynamics
Pharmacodynamics (PD) is the study of the biochemistry, biochemical and physiology, physiologic effects of drugs (especially pharmaceutical drugs). The effects can include those manifested within animals (including humans), microorganisms, or comb ...
, following the work of
Cox and Atkinson.
Bayesian experimental design
When practitioners need to consider multiple
models
A model is an informative representation of an object, person, or system. The term originally denoted the plans of a building in late 16th-century English, and derived via French and Italian ultimately from Latin , .
Models can be divided int ...
, they can specify a
probability-measure on the models and then select any design maximizing the
expected value
In probability theory, the expected value (also called expectation, expectancy, expectation operator, mathematical expectation, mean, expectation value, or first Moment (mathematics), moment) is a generalization of the weighted average. Informa ...
of such an experiment. Such probability-based optimal-designs are called optimal
Bayesian designs. Such
Bayesian designs are used especially for
generalized linear models (where the response follows an
exponential-family distribution).
The use of a
Bayesian design does not force statisticians to use
Bayesian methods
Bayesian inference ( or ) is a method of statistical inference in which Bayes' theorem is used to calculate a probability of a hypothesis, given prior evidence, and update it as more information becomes available. Fundamentally, Bayesian inferen ...
to analyze the data, however. Indeed, the "Bayesian" label for probability-based experimental-designs is disliked by some researchers. Alternative terminology for "Bayesian" optimality includes "on-average" optimality or "population" optimality.
Iterative experimentation
Scientific experimentation is an iterative process, and statisticians have developed several approaches to the optimal design of sequential experiments.
Sequential analysis
Sequential analysis
In statistics, sequential analysis or sequential hypothesis testing is statistical analysis where the sample size is not fixed in advance. Instead data is evaluated as it is collected, and further sampling is stopped in accordance with a pre-defi ...
was pioneered by
Abraham Wald. In 1972,
Herman Chernoff wrote an overview of optimal sequential designs, while
adaptive designs were surveyed later by S. Zacks. Of course, much work on the optimal design of experiments is related to the theory of
optimal decision
An optimal decision is a decision that leads to at least as good a known or expected outcome as all other available decision options. It is an important concept in decision theory. In order to compare the different decision outcomes, one commonly ...
s, especially the
statistical decision theory of
Abraham Wald.
Response-surface methodology
Optimal designs for
response-surface models are discussed in the textbook by Atkinson, Donev and Tobias, and in the survey of Gaffke and Heiligers and in the mathematical text of Pukelsheim. The
blocking of optimal designs is discussed in the textbook of Atkinson, Donev and Tobias and also in the monograph by Goos.
The earliest optimal designs were developed to estimate the parameters of regression models with continuous variables, for example, by
J. D. Gergonne in 1815 (Stigler). In English, two early contributions were made by
Charles S. Peirce an
Kirstine Smith
Pioneering designs for multivariate
response-surfaces were proposed by
George E. P. Box. However, Box's designs have few optimality properties. Indeed, the
Box–Behnken design requires excessive experimental runs when the number of variables exceeds three.
Box's
"central-composite" designs require more experimental runs than do the optimal designs of Kôno.
System identification and stochastic approximation
The optimization of sequential experimentation is studied also in
stochastic programming and in
systems
A system is a group of interacting or interrelated elements that act according to a set of rules to form a unified whole. A system, surrounded and influenced by its environment, is described by its boundaries, structure and purpose and is exp ...
and
control. Popular methods include
stochastic approximation and other methods of
stochastic optimization
Stochastic optimization (SO) are optimization methods that generate and use random variables. For stochastic optimization problems, the objective functions or constraints are random. Stochastic optimization also include methods with random iter ...
. Much of this research has been associated with the subdiscipline of
system identification
The field of system identification uses statistical methods to build mathematical models of dynamical systems from measured data. System identification also includes the optimal design#System identification and stochastic approximation, optimal de ...
.
In computational
optimal control
Optimal control theory is a branch of control theory that deals with finding a control for a dynamical system over a period of time such that an objective function is optimized. It has numerous applications in science, engineering and operations ...
, D. Judin & A. Nemirovskii an
Boris Polyakhas described methods that are more efficient than the (
Armijo-style)
step-size rules introduced by
G. E. P. Box in
response-surface methodology.
Adaptive designs are used in
clinical trials
Clinical trials are prospective biomedical or behavioral research studies on human subject research, human participants designed to answer specific questions about biomedical or behavioral interventions, including new treatments (such as novel v ...
, and optimal
adaptive designs are surveyed in the ''Handbook of Experimental Designs'' chapter by Shelemyahu Zacks.
Specifying the number of experimental runs
Using a computer to find a good design
There are several methods of finding an optimal design, given an ''a priori'' restriction on the number of experimental runs or replications. Some of these methods are discussed by Atkinson, Donev and Tobias and in the paper by Hardin and
Sloane. Of course, fixing the number of experimental runs ''a priori'' would be impractical. Prudent statisticians examine the other optimal designs, whose number of experimental runs differ.
Discretizing probability-measure designs
In the mathematical theory on optimal experiments, an optimal design can be a
probability measure
In mathematics, a probability measure is a real-valued function defined on a set of events in a σ-algebra that satisfies Measure (mathematics), measure properties such as ''countable additivity''. The difference between a probability measure an ...
that is
supported on an infinite set of observation-locations. Such optimal probability-measure designs solve a mathematical problem that neglected to specify the cost of observations and experimental runs. Nonetheless, such optimal probability-measure designs can be
discretized to furnish
approximately optimal designs.
In some cases, a finite set of observation-locations suffices to
support an optimal design. Such a result was proved by Kôno and
Kiefer in their works on
response-surface designs for quadratic models. The Kôno–Kiefer analysis explains why optimal designs for response-surfaces can have discrete supports, which are very similar as do the less efficient designs that have been traditional in
response surface methodology.
History
In 1815, an article on optimal designs for
polynomial regression was published by
Joseph Diaz Gergonne, according to
Stigler.
Charles S. Peirce proposed an economic theory of scientific experimentation in 1876, which sought to maximize the precision of the estimates. Peirce's optimal allocation immediately improved the accuracy of gravitational experiments and was used for decades by Peirce and his colleagues. In his 1882 published lecture at
Johns Hopkins University
The Johns Hopkins University (often abbreviated as Johns Hopkins, Hopkins, or JHU) is a private university, private research university in Baltimore, Maryland, United States. Founded in 1876 based on the European research institution model, J ...
, Peirce introduced experimental design with these words:
Logic will not undertake to inform you what kind of experiments you ought to make in order best to determine the acceleration of gravity, or the value of the Ohm; but it will tell you how to proceed to form a plan of experimentation.
...Unfortunately practice generally precedes theory, and it is the usual fate of mankind to get things done in some boggling way first, and find out afterward how they could have been done much more easily and perfectly.[Peirce, C. S. (1882), "Introductory Lecture on the Study of Logic" delivered September 1882, published in ''Johns Hopkins University Circulars'', v. 2, n. 19, pp. 11–12, November 1882, see p. 11, ''Google Books']
Eprint
Reprinted in ''Collected Papers'' v. 7, paragraphs 59–76, see 59, 63, ''Writings of Charles S. Peirce'' v. 4, pp. 378–82, see 378, 379, and ''The Essential Peirce'' v. 1, pp. 210–14, see 210–1, also lower down on 211.
Kirstine Smith proposed optimal designs for polynomial models in 1918. (Kirstine Smith had been a student of the Danish statistician
Thorvald N. Thiele and was working with
Karl Pearson
Karl Pearson (; born Carl Pearson; 27 March 1857 – 27 April 1936) was an English biostatistician and mathematician. He has been credited with establishing the discipline of mathematical statistics. He founded the world's first university ...
in London.)
See also
*
Bayesian experimental design
*
Blocking (statistics)
In the statistical theory of the design of experiments, blocking is the arranging of experimental units that are similar to one another in groups (blocks) based on one or more variables. These variables are chosen carefully to minimize the effect ...
*
Computer experiment
A computer experiment or simulation experiment is an experiment used to study a computer simulation, also referred to as an in silico system. This area includes computational physics, computational chemistry, computational biology and other simila ...
*
Convex function
In mathematics, a real-valued function is called convex if the line segment between any two distinct points on the graph of a function, graph of the function lies above or on the graph between the two points. Equivalently, a function is conve ...
*
Convex minimization
*
Design of experiments
The design of experiments (DOE), also known as experiment design or experimental design, is the design of any task that aims to describe and explain the variation of information under conditions that are hypothesized to reflect the variation. ...
*
Efficiency (statistics)
In statistics, efficiency is a measure of quality of an estimator, of an experimental design, or of a hypothesis testing procedure. Essentially, a more efficient estimator needs fewer input data or observations than a less efficient one to achiev ...
*
Entropy (information theory)
In information theory, the entropy of a random variable quantifies the average level of uncertainty or information associated with the variable's potential states or possible outcomes. This measures the expected amount of information needed ...
*
Fisher information
*
Glossary of experimental design
*
Hadamard's maximal determinant problem
*
Information theory
Information theory is the mathematical study of the quantification (science), quantification, Data storage, storage, and telecommunications, communication of information. The field was established and formalized by Claude Shannon in the 1940s, ...
*
Kiefer, Jack
*
Replication (statistics)
In engineering, science, and statistics, replication is the process of repeating a study or experiment under the same or similar conditions. It is a crucial step to test the original claim and confirm or reject the accuracy of results as well as f ...
*
Response surface methodology
*
Statistical model
A statistical model is a mathematical model that embodies a set of statistical assumptions concerning the generation of Sample (statistics), sample data (and similar data from a larger Statistical population, population). A statistical model repre ...
*
Wald, Abraham
*
Wolfowitz, Jacob
Notes
References
*
*
*
*
*
*
*
*
*
*
Further reading
Textbooks for practitioners and students
Textbooks emphasizing regression and response-surface methodology
The textbook by Atkinson, Donev and Tobias has been used for short courses for industrial practitioners as well as university courses.
*
*
Textbooks emphasizing block designs
Optimal
block designs are discussed by Bailey and by Bapat. The first chapter of Bapat's book reviews the
linear algebra
Linear algebra is the branch of mathematics concerning linear equations such as
:a_1x_1+\cdots +a_nx_n=b,
linear maps such as
:(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n,
and their representations in vector spaces and through matrix (mathemat ...
used by Bailey (or the advanced books below). Bailey's exercises and discussion of
randomization
Randomization is a statistical process in which a random mechanism is employed to select a sample from a population or assign subjects to different groups.Oxford English Dictionary "randomization" The process is crucial in ensuring the random alloc ...
both emphasize statistical concepts (rather than algebraic computations).
* Draft available on-line. (Especially Chapter 11.8 "Optimality")
* (Chapter 5 "Block designs and optimality", pages 99–111)
Optimal
block designs are discussed in the advanced monograph by Shah and Sinha and in the survey-articles by Cheng and by Majumdar.
Books for professional statisticians and researchers
*
*
*
*
*
*
* Republication with errata-list and new preface of Wiley (0-471-61971-X) 1993
*
Articles and chapters
*
*
**
**
**
**
**
**
**
*
Historical
*
*
* (Appendix No. 14)
NOAA PDF Eprint Reprinted in paragraphs 139–157, and in
*
{{Statistics, collection, state=collapsed
Design of experiments
Regression analysis
Statistical theory
Optimal decisions
Mathematical optimization
Industrial engineering
Systems engineering
Statistical process control
Management cybernetics