0% found this document useful (0 votes)
12 views8 pages

Lec 24

The document discusses the concepts of bias and variance in machine learning, explaining how they affect model performance and generalization error. It highlights the tradeoff between bias and variance, emphasizing that reducing bias often increases variance and vice versa. Strategies for overcoming bias and variance include using more powerful models, better algorithms, and additional data, while also noting the challenges of detecting overfitting and underfitting.

Uploaded by

Ashutosh Anand
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
12 views8 pages

Lec 24

The document discusses the concepts of bias and variance in machine learning, explaining how they affect model performance and generalization error. It highlights the tradeoff between bias and variance, emphasizing that reducing bias often increases variance and vice versa. Strategies for overcoming bias and variance include using more powerful models, better algorithms, and additional data, while also noting the challenges of detecting overfitting and underfitting.

Uploaded by

Ashutosh Anand
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
You are on page 1/ 8

Bias and Variance

Bias enters as is not a good approximation for


e.g. trying to learn a linear model to predict a
Where to put blame transformer’s output
• Variance enters as we were not able to learn
We have regression the bestdata where
parameters for , due
sayto poor training
etc.
We have learnt a model
• Bayes with params
error is unavoidable to approximate
– perhaps labels are
depends on That(train set,
is neat
inherently– inotherorrandom
practical
noisy elseterms, choices suchthat
this is saying
the features are as
initialization etc)
errors are either
insufficient duethe
to solve to calling
task the incorrect
Let – will write just to simplify
constructor in sklearnnotation
(or passing incorrect
arguments to the constructor) or issues in how the
Its average testfit() error is not
function alljob,
did its that nice issue
or some – whatin is to
blame? features/labels themselves!
The expectation over means
Variance aka Bias akathe bad performance was
that Bayes
As and
Estimation Approximation
not due to unluckyError
choices e.g.
Error Error train-test split etc.
Bias vs Variance
Similar to the distinction between accuracy and
precision
Variance comes in due to dependence of on random
choices
Training set, choice of mini-batches in (MB)SGD, random
initialization, etc

Low Bias Low Bias High Bias High Bias


Low Variance High Variance Low Variance High Variance
Bias Variance Tradeoff
Easiest way to reduce bias
Test is to make model more
powerful
Error
Non-linear models, universal
models, NN with lots of
layers
Bias
Powerful models have their
Varian own problems 
ce Parameter estimation is
often (NP)hard
Low Medium High Brittle – changing training
Model Complexity data slightly changes
parameters a lot
More prone to overfitting too
Overcoming Bias, Variance,
ChallengingBayes
detect Bayes
to

Error error as it may


hide inside bias

Test error can be blamed on


Bias: lowest error allowed by model (cant get better without changing
model)
Variance: how well are we able to achieve the lowest error our model
allows
Bayes Error: something wrong with how problem itself is setup
Low bias, low variance models usually exist only for specific
domains
To lower bias, make model more powerful (variance may go up)
To lower variance, use more data or better learning algorithm
To lower Bayes error, add more (informative) features
Can also lower bias but can also increase variance
The presence of an (over) abundance

BV Tradeoff in the Age of LLMsof parameters and data seems to


makes it possible to discover good
local minima which not only do well
on training data but on test data as
Test well
Error Still an active area of
research

Bias
Varian
ce

Low Medium High Overparametrized regime


Model Complexity
Generalization error (just like
Generalization Error variance) can usually be brought
down by using more data points or
choosing Gap
modelsbetween train and
that are simpler test
error
Test Error Difference between variance and gen error is
How well is model able to
subtle. Model complexity and train set
“generalize” to size
unseen data
affect both. However, variance can alsooften
Underfitting: be high
due to high
because of using an bias,
improper
poorlearning
training
algorithm or not optimizing properly e.g.
Overfitting: oftensmall
due to high
Underfitting Overfitting
mini-batches
variance, over training
Gen Gen error small if small bias
and small variance
Training Error
Error Note: large bias may have
good gen error yet high
Low Mediu High
Model Complexity
m
test error
Test error may be close to
train error even if both are
Adding more data cannot
Detecting Over/underfitting decrease bias. The chosen
model just sucks  Adding more
data can decrease variance
Low training error but high test error??
thoughis memorizing
You may have overfit (high variance) – your model
training data Sometimes may need to iterate
Your model is clearly powerfulthrough
enoughthe
– does notexperiences
above seem to be a bias
problem (experience high bias, reduce it only
Use more data/better optimizer/simpler model (or
to increase variance, all) decrease
then to decrease
variance
variance etc) before reaching a sweet
High training error and high test error??spot
You may have underfit (high bias) – model incapable of handling learning
task
Increase model class complexity, add better features, to decrease bias
Use more data, better ML algo to address any underlying variance issues
Low training error and low test error
Show-off !!! 
High training error and low test error

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy