Bias-Variance Tradeoff

Published

December 29, 2023

TAGS:

Recently, I have begun working through some of the simpler ML/algorithms/models, such as KNN, logistical regression. The bias-variance tradeoff regularly appears in the materials I am reading.

I do have a vague notion about overfitting, for example “perfecting” a training set by addding addition parameters (dimensions?). When model is run on testing data, of course, it does not work as well.

Seems plausible to me, but something bothers me. I have worked through the mathematical derviations of bias-tradeoff a few times (here, here, and here), understand some not all of the math, yet continue to walk about realizing I still do understand this at an intutive level.

Is bias-variance tradeoff like the Uncertainty Principle ?

\[ \Delta{x}*\Delta{p} = \frac{h}{2*pi} \]

This article greatly helped:

There are b_hat people and y_yat people. This shed some light.

This diagram from wikipedia turned the lightbulb on: https://en.wikipedia.org/wiki/Bias_of_an_estimator

sketch2

vim:linebreak:nospell:nowrap:cul tw=78 fo=tqlnrc foldcolumn=1 cc=+1