digplanet beta 1: Athena
Share digplanet:

Agriculture

Applied sciences

Arts

Belief

Business

Chronology

Culture

Education

Environment

Geography

Health

History

Humanities

Language

Law

Life

Mathematics

Nature

People

Politics

Science

Society

Technology

In statistics, the method of moments is a method of estimation of population parameters such as mean, variance, median, etc. (which need not be moments), by equating sample moments with unobservable population moments and then solving those equations for the quantities to be estimated.

## Methodology

Suppose that the problem is to estimate $p$ unknown parameters $\theta_{1}, \theta_{2}, \dots, \theta_{p}$ characterizing a distribution $f_{W}(w; \theta)$. Suppose $p$ of the moments of the true distribution can be expressed as functions of the $\theta$s:

$\mu_{1} \equiv E[W^1]=g_{1}(\theta_{1}, \theta_{2}, \dots, \theta_{p}) ,$
$\mu_{2} \equiv E[W^2]=g_{2}(\theta_{1}, \theta_{2}, \dots, \theta_{p}) ,$
$\vdots$
$\mu_{p} \equiv E[W^p]=g_{p}(\theta_{1}, \theta_{2}, \dots, \theta_{p}) .$

Let $\hat{\mu_{j}}=(\Sigma_{i=1}^{n} w_{i}^{j})/n$ be the j-th sample moment corresponding to the population moment $\mu_{j}$. The method of moments estimator for $\theta_{1}, \theta_{2}, \dots, \theta_{p}$ denoted by $\hat{\theta}_{1}, \hat{\theta_{2}}, \dots, \hat{\theta}_{p}$ is defined by the solution (if there is one) to the equations:[citation needed]

$\hat \mu_{1} = g_{1}(\hat{\theta}_{1}, \hat{\theta_{2}}, \dots, \hat{\theta}_{p}) ,$
$\hat \mu_{2} = g_{2}(\hat{\theta}_{1}, \hat{\theta_{2}}, \dots, \hat{\theta}_{p}) ,$
$\vdots$
$\hat \mu_{p} = g_{p}(\hat{\theta}_{1}, \hat{\theta_{2}}, \dots, \hat{\theta}_{p}) .$

## Example

Suppose X1, ..., Xn are independent identically distributed random variables with a gamma distribution with probability density function

${x^{\alpha-1} e^{-x/\beta} \over \beta^\alpha\, \Gamma(\alpha)} \,\!$

for x > 0, and 0 for x < 0.

The first moment, i.e., the expected value, of a random variable with this probability distribution is

$\operatorname{E}(X_1)=\alpha\beta\,$

and the second moment, i.e., the expected value of its square, is

$\operatorname{E}(X_1^2)=\beta^2\alpha(\alpha+1).\,$

These are the "population moments".

The first and second "sample moments" m1 and m2 are respectively

$m_{1} = {X_1+\cdots+X_n \over n} \,\!$

and

$m_{2} = {X_1^2+\cdots+X_n^2 \over n}.\,\!$

Equating the population moments with the sample moments, we get

$\alpha\beta = m_{1} \,\!$

and

$\beta^2\alpha(\alpha+1) = m_{2}.\,\!$

Solving these two equations for α and β, we get

$\alpha={ m_{1}^2 \over m_{2} - m_{1}^2}\,\!$

and

$\beta={ m_{2} - m_{1}^2 \over m_{1}}.\,\!$

We then use these 2 quantities as estimates, based on the sample, of the two unobservable population parameters α and β.

## Advantages and disadvantages of this method

In some respects, when estimating parameters of a known family of probability distributions, this method was superseded by Fisher's method of maximum likelihood, because maximum likelihood estimators have higher probability of being close to the quantities to be estimated.

However, in some cases, as in the above example of the gamma distribution, the likelihood equations may be intractable without computers, whereas the method-of-moments estimators can be quickly and easily calculated by hand as shown above.

Estimates by the method of moments may be used as the first approximation to the solutions of the likelihood equations, and successive improved approximations may then be found by the Newton–Raphson method. In this way the method of moments and the method of maximum likelihood are symbiotic.

In some cases, infrequent with large samples but not so infrequent with small samples, the estimates given by the method of moments are outside of the parameter space; it does not make sense to rely on them then. That problem never arises in the method of maximum likelihood. Also, estimates by the method of moments are not necessarily sufficient statistics, i.e., they sometimes fail to take into account all relevant information in the sample.

When estimating other structural parameters (e.g., parameters of a utility function, instead of parameters of a known probability distribution), appropriate probability distributions may not be known, and moment-based estimates may be preferred to Maximum Likelihood Estimation.

## See also

Original courtesy of Wikipedia: http://en.wikipedia.org/wiki/Method_of_moments_(statistics) — Please support Wikipedia.
A portion of the proceeds from advertising on Digplanet goes to supporting Wikipedia.
 607532 videos foundNext >
 (ML 4.1) Maximum Likelihood Estimation (MLE) (part 1)Definition of maximum likelihood estimates (MLEs), and a discussion of pros/cons. A playlist of these Machine Learning videos is available here: http://www.y... What is Generalized Method of Moments? by Alastair HallThe full length methods@manchester video can be watched for free at http://www.methods.manchester.ac.uk/methods/gmm/index.shtml Generalized Method of Moments... Introductory Statistics - Chapter 7: EstimationA video summary of chapter 7 in Perdisco's Introductory Statistics 360Textbook. To find out more, visit www.perdisco.com/introstats. Transformations Of Multiple r.v.s - Method of MomentsThis is Part 1 of Lecture 12, on "Transformations of Multiple Random Variables". This segment defines the "method of moments" procedure to find the CDF of a ... ECE6340 Lecture 20-1: Introduction to the Method of MomentsIntro to the method of moments (MOM) for solving integral equations. As an example, we consider the charge distribution on a thin wire. www.ece.utah.edu/~cfu... Estimating Insurance Premiums Using Exceedance Data and the Method of Momentshttp://demonstrations.wolfram.com/EstimatingInsurancePremiumsUsingExceedanceDataAndTheMethodOf/ The Wolfram Demonstrations Project contains thousands of free... ECE6340 Lecture 20-4: Weight functions with the Method of MomentsExplanation of the weighting functions when used in the method of moments. The example is applied to finding the charge distribution on a thin wire. Written ... Maximum Likelihood Example: BernoulliThe use of maximum likelihood estimation to estimate the parameter of a Bernoulli random variable. Trendkill Method - In Moments Of Sadness (Riga, 31.08.2012)Latvian metal band opens show with Korn. "The Path of Totality" tour 2012. Arena Riga, Latvia. Drawing Shear and Moment Diagrams Example- Mechanics of Materials and Staticsthis is a detailed example of shear and moment diagrams, i recommend skipping around to the sections shown below if you already have a feel for the subject: ...
 607532 videos foundNext >

We're sorry, but there's no news about "Method of moments (statistics)" right now.

 Limit to books that you can completely read online Include partial books (book previews) .gsc-branding { display:block; }

Oops, we seem to be having trouble contacting Twitter

# Talk About Method of moments (statistics)

You can talk about Method of moments (statistics) with people all over the world in our discussions.

#### Support Wikipedia

A portion of the proceeds from advertising on Digplanet goes to supporting Wikipedia. Please add your support for Wikipedia!
 Copyright © 2009-2013 Digparty. All rights reserved.