Listen "Bias, Variance, and the Bias-Variance Tradeoff"
Episode Synopsis
The bias-variance trade-off is a key problem in your model search. While bias represents how well your model can capture the salient details of a problem, and generally correlates with more complex algorithms, it comes at the trade off of variance. Variance is the degree to which on individual predictions your estimators stray from the mean output on those values. High variance means that a model has overfit, and incorrectly or incompletely learned the problem from the training set. Most commonly, high bias = underfitting, high variance = overfitting.
Please consider joining the conversation on Twitter. I also blog from time to time. You can find me at erikpartridge.com.
For more academic sources, consider reading the slides from this fantastic Carnegie Mellon lecture.
Please consider joining the conversation on Twitter. I also blog from time to time. You can find me at erikpartridge.com.
For more academic sources, consider reading the slides from this fantastic Carnegie Mellon lecture.
More episodes of the podcast Machine Learning Bytes
K-Fold Cross Validation
31/07/2019
Stratified Sampling
30/07/2019
Boosting
26/07/2019
Bagging
24/07/2019
Empirical Risk Minimization
20/07/2019
ZARZA We are Zarza, the prestigious firm behind major projects in information technology.