Name: Towards AI Legal Name: Towards AI, Inc. Description: Towards AI is the world's leading artificial intelligence (AI) and technology publication. Read by thought-leaders and decision-makers around the world. Phone Number: +1-650-246-9381 Email: [email protected]
228 Park Avenue South New York, NY 10003 United States
Website: Publisher: https://towardsai.net/#publisher Diversity Policy: https://towardsai.net/about Ethics Policy: https://towardsai.net/about Masthead: https://towardsai.net/about
Name: Towards AI Legal Name: Towards AI, Inc. Description: Towards AI is the world's leading artificial intelligence (AI) and technology publication. Founders: Roberto Iriondo, , Job Title: Co-founder and Advisor Works for: Towards AI, Inc. Follow Roberto: X, LinkedIn, GitHub, Google Scholar, Towards AI Profile, Medium, ML@CMU, FreeCodeCamp, Crunchbase, Bloomberg, Roberto Iriondo, Generative AI Lab, Generative AI Lab Denis Piffaretti, Job Title: Co-founder Works for: Towards AI, Inc. Louie Peters, Job Title: Co-founder Works for: Towards AI, Inc. Louis-François Bouchard, Job Title: Co-founder Works for: Towards AI, Inc. Cover:
Towards AI Cover
Logo:
Towards AI Logo
Areas Served: Worldwide Alternate Name: Towards AI, Inc. Alternate Name: Towards AI Co. Alternate Name: towards ai Alternate Name: towardsai Alternate Name: towards.ai Alternate Name: tai Alternate Name: toward ai Alternate Name: toward.ai Alternate Name: Towards AI, Inc. Alternate Name: towardsai.net Alternate Name: pub.towardsai.net
5 stars – based on 497 reviews

Frequently Used, Contextual References

TODO: Remember to copy unique IDs whenever it needs used. i.e., URL: 304b2e42315e

Resources

Take our 85+ lesson From Beginner to Advanced LLM Developer Certification: From choosing a project to deploying a working product this is the most comprehensive and practical LLM course out there!

Publication

Minimizing the Mean Square Error: Frequentist Approach
Latest   Machine Learning

Minimizing the Mean Square Error: Frequentist Approach

Last Updated on June 4, 2024 by Editorial Team

Author(s): Varun Nakra

Originally published on Towards AI.

Statistical Inference and Mean Square Error

In a utopian world, we would have access to unlimited data, i.e., the entire population, so there wouldn’t be a need for an β€˜inference’, as we would know for certain about what we are interested in. However, in all practical settings, we don’t have access to the β€˜population’ but only have access to observations that comprise a sample. This results in the need to make β€˜inferences’ about the population using the sample. Statistical inference is drawing some type of conclusions about one or more parameters (population characteristics), and point estimation is selecting a single number, based on sample data, that represents a sensible value for those parameters. This gives rise to two concepts β€” the true value of the parameter, the β€˜ground truth’, which is hidden from us as we don’t have access to the population, and the point estimate of the parameter.

The Frequentist considers the true value of the parameter ΞΈ as fixed but unknown. The Bayesian considers the true value of the parameter ΞΈ as an β€˜observed’ value of a random variable Θ (which does not get observed by us, but is only a realization of the true value)

Again, under ideal settings, we could find an estimator that is exactly equal to the true value of the parameter, always! However, since the estimate of the true value of the parameter is a function of the sample, it is a random variable per se. This is because different samples would result in different values of the estimate. For some samples, the true parameter will be overestimated and for others underestimated. This leads us to the idea of an β€˜error’ in our estimation.

Now, there could be multiple ways, such as squared error, absolute error, etc., in which the error of estimation mentioned in 1 can be quantified. However, we will stick to only squared error in this article and explore other β€˜error functions’ in related articles in the future.

It makes sense to β€˜average’ the squared error defined in 2 to measure the performance of the estimator on an β€˜average’. This gives rise to the concept of β€˜Mean Square Error’, which is defined as follows

Minimizing Mean Square Error: Frequentist approach

As mentioned above, under the frequentist approach, we will consider the true parameter as fixed but unknown and attempt to minimize it.

Consider a simple example of estimating the population mean using the sample mean for n observations x1, . . . , xn. The sample mean is the estimator of the population mean and is defined as follows

Instead of using 4, we use a modified version of sample mean that is multiplied by a constant k. The objective is to solve for the optimal value of the constant for which the mean square error will be minimum.

As evident from 11, the optimal value of k depends on the unknown parameter ΞΈ. In other words, the estimator of minimum mean square error is not realizable because, being dependent on ΞΈ, it depends on the bias term. Therefore, finding the minimum mean square error estimator becomes an unrealized task for this simple example and also for most practical cases. It is for this very reason that we constrain the bias to zero and find the estimator that minimizes the variance, leading to the concept of Minimum Variance Unbiased Estimator(MVUE). However, unfortuntately, even the MVUE estimator does not always exist. Are there alternative ways of finding estimators with minimum variance and minimum bias? Stay tuned for the answer in the sequel article β€˜Minimizing the Mean Square Error: Bayesian approach’

Join thousands of data leaders on the AI newsletter. Join over 80,000 subscribers and keep up to date with the latest developments in AI. From research to projects and ideas. If you are building an AI startup, an AI-related product, or a service, we invite you to consider becoming aΒ sponsor.

Published via Towards AI

Feedback ↓