Watch
Watching this resources will notify you when proposed changes or new versions are created so you can keep track of improvements that have been made.
Favorite
Favoriting this resource allows you to save it in the “My Resources” tab of your account. There, you can easily access this resource later when you’re ready to customize it or assign it to your students.
Degrees of Freedom
The number of degrees of freedom is the number of values in the final calculation of a statistic that are free to vary.
Learning Objective

Outline an example of "degrees of freedom"
Key Points
 The degree of freedom can be defined as the minimum number of independent coordinates which can specify the position of the system completely.
 A parameter is a characteristic of the variable under examination as a whole; it is part of describing the overall distribution of values.
 As more degrees of freedom are lost, fewer and fewer different situations are accounted for by a model since fewer and fewer pieces of information could, in principle, be different from what is actually observed.
 Degrees of freedom can be seen as linking sample size to explanatory power.
Terms

residual
The difference between the observed value and the estimated function value.

vector
in statistics, a set of realvalued random variables that may be correlated
Full Text
The number of independent ways by which a dynamical system can move without violating any constraint imposed on it is known as "degree of freedom. " The degree of freedom can be defined as the minimum number of independent coordinates that completely specify the position of the system.
Consider this example: To compute the variance, first sum the square deviations from the mean. The mean is a parameter, a characteristic of the variable under examination as a whole, and a part of describing the overall distribution of values. Knowing all the parameters, you can accurately describe the data. The more known (fixed) parameters you know, the fewer samples fit this model of the data. If you know only the mean, there will be many possible sets of data that are consistent with this model. However, if you know the mean and the standard deviation, fewer possible sets of data fit this model.
In computing the variance, first calculate the mean, then you can vary any of the scores in the data except one. This one score left unexamined can always be calculated accurately from the rest of the data and the mean itself.
As an example, take the ages of a class of students and find the mean. With a fixed mean, how many of the other scores (there are N of them remember) could still vary? The answer is N1 independent pieces of information (degrees of freedom) that could vary while the mean is known. One piece of information cannot vary because its value is fully determined by the parameter (in this case the mean) and the other scores. Each parameter that is fixed during our computations constitutes the loss of a degree of freedom.
Imagine starting with a small number of data points and then fixing a relatively large number of parameters as we compute some statistic. We see that as more degrees of freedom are lost, fewer and fewer different situations are accounted for by our model since fewer and fewer pieces of information could, in principle, be different from what is actually observed.
Put informally, the "interest" in our data is determined by the degrees of freedom. If there is nothing that can vary once our parameter is fixed (because we have so very few data points, maybe just one) then there is nothing to investigate. Degrees of freedom can be seen as linking sample size to explanatory power.
The degrees of freedom are also commonly associated with the squared lengths (or "sum of squares" of the coordinates) of random vectors and the parameters of chisquared and other distributions that arise in associated statistical testing problems.
Notation and Residuals
In equations, the typical symbol for degrees of freedom is
In fitting statistical models to data, the random vectors of residuals are constrained to lie in a space of smaller dimension than the number of components in the vector. That smaller dimension is the number of degrees of freedom for error. In statistical terms, a random vector is a list of mathematical variables each of whose value is unknown, either because the value has not yet occurred or because there is imperfect knowledge of its value. The individual variables in a random vector are grouped together because there may be correlations among them. Often they represent different properties of an individual statistical unit (e.g., a particular person, event, etc.).
A residual is an observable estimate of the unobservable statistical error. Consider an example with men's heights and suppose we have a random sample of n people. The sample mean could serve as a good estimator of the population mean. The difference between the height of each man in the sample and the observable sample mean is a residual. Note that the sum of the residuals within a random sample is necessarily zero, and thus the residuals are necessarily not independent.
Perhaps the simplest example is this. Suppose X_{1},...,X_{n} are random variables each with expected value μ, and let
be the "sample mean. " Then the quantities
are residuals that may be considered estimates of the errors X_{i} − μ. The sum of the residuals is necessarily 0. If one knows the values of any n − 1 of the residuals, one can thus find the last one. That means they are constrained to lie in a space of dimension n − 1, and we say that "there are n − 1 degrees of freedom for error. "
Degrees of Freedom
This image illustrates the difference (or distance) between the cumulative distribution functions of the standard normal distribution (Φ) and a hypothetical distribution of a standardized sample mean (Fn). Specifically, the plotted hypothetical distribution is a t distribution with 3 degrees of freedom.
Assign just this concept or entire chapters to your class for free.
Key Term Reference
 correlation
 Appears in these related concepts: Benefits of Globalization, Controlling for a Variable, and Descriptive and Correlational Statistics
 datum
 Appears in these related concepts: Change of Scale, One and TwoTailed Tests, and Summary for inference of the difference of two means
 degree of freedom
 Appears in these related concepts: Comparing Two Populations: Paired Difference Experiment, Overview of Temperature and Kinetic Theory, and Estimating a Population Variance
 degrees of freedom
 Appears in these related concepts: tTest for One Sample, Structure of the ChiSquared Test, and Specific Heat and Heat Capacity
 deviation
 Appears in these related concepts: Variance, Introduction to Bivariate Data, and Basic properties of point estimates
 distribution
 Appears in these related concepts: Application of Knowledge, Monte Carlo Simulation, and Selling to Consumers
 error
 Appears in these related concepts: Estimation, Precise Definition of a Limit, and Introduction to confidence intervals
 expected value
 Appears in these related concepts: What Does the Law of Averages Say?, Expected Values of Discrete Random Variables, and Expected Return
 independent
 Appears in these related concepts: Fundamentals of Probability, Unions and Intersections, and Party Identification
 mean
 Appears in these related concepts: Mean, Variance, and Standard Deviation of the Binomial Distribution, Averages, and Understanding Statistics
 population
 Appears in these related concepts: Applications of Statistics, The Functionalist Perspective on Deviance, and Quorum Sensing
 random sample
 Appears in these related concepts: Random Samples, Comparing Two Independent Population Proportions, and Confidence Interval, Single Population Mean, Standard Deviation Unknown, Student'st
 random variable
 Appears in these related concepts: Chance Processes, The Sample Average, and Expected Value
 residuals
 Appears in these related concepts: Two Regression Lines, Models with Both Quantitative and Qualitative Variables, and Inferences of Correlation and Regression
 sample
 Appears in these related concepts: Identifying Product Benefits, Surveys, and Basic Inferential Statistics
 sample mean
 Appears in these related concepts: Which Standard Deviation (SE)?, Lab 1: Confidence Interval (Home Costs), and Type I and II Errors
 standard deviation
 Appears in these related concepts: Using the Normal Curve, Interpreting the Standard Deviation, and Variance
 statistics
 Appears in these related concepts: Communicating Statistics, Population Demography, and What Is Statistics?
 variable
 Appears in these related concepts: What is a Linear Function?, Math Review, and Introduction to Variables
 variance
 Appears in these related concepts: Comparing Nested Models, Testing a Single Variance, and Variance Estimates
Sources
Boundless vets and curates highquality, openly licensed content from around the Internet. This particular resource used the following sources:
Cite This Source
Source: Boundless. “Degrees of Freedom.” Boundless Statistics. Boundless, 26 May. 2016. Retrieved 30 Sep. 2016 from https://www.boundless.com/statistics/textbooks/boundlessstatisticstextbook/measuresofvariation6/describingvariability26/degreesoffreedom1364427/