[ML] 2. Descending into ML

Daniel Seo·2021년 7월 10일
0
post-custom-banner

Linear Regression

Figure 1. Chirps per Minute vs Temperature in Celsius

The plot shows the temperature rising with the number of chirps. Is this relationship between chirps and temperature linear? Yes, you could draw a single straight line like thw following approximate this relationship:

Figure 2. A linear relationship

True, the line doesn't pass through every dot, but the line does clearly show the relationship between chirps and temperature. Usint the equation for a line, you could write down the relationship as follows:

y=mx+by = mx + b

where:

  • y is the temperature in Celsius-the value we're trying to predict
  • m is the slope of the line.
  • x is the number of chirps per minute - the value of our input feature
  • b is the y-intercept

By convention in machine learning, you'll write the equation for a model slightly differently:

y=b+w1x1{y'} = b+ w_1x_1

where:

  • yy' is the predicted label (a desired output)
  • bb is the bias (the y-intercept), sometimes referred to as w0w0.
  • w1w1 is the weight of feature 1. Weight is the same concept as the "slope" m in the traditional equation of a line.
  • x1x1 is a feature (a known input).

To infer(predict) the temperature yy' for a new chirps-per-minute value x1x1, just subsititue the x1x1 value into this model

Although this model uses only one feature, a more sophisticated model might rely on multiple features, each having a seerate weight. For example, a model that relies on three features might look as follows:

y=b+w1x1+w2x2+w3x3y'= b+w_1x_1 +w_2x_2 + w_3x_3

Training and Loss

Training a model means learning (determining) good values for all the weights and the bias from labled examples.

Loss is the penalty for a bad prediction. That is, loss is a number indicating how bad the model's prediction was on a single example.

Loss Function

Squared Loss (also know as L2L_2 Loss)

= the square of the difference between the label and the prediction
= (obaservation - prediction(x))^2
= (y-y')^2

Mean squre error (MSE)

the everate squared loss per example over the whole datasheet. To calculate MSE, sum up all the squared losses for individual examples and then divide by the number of examples:

MSE=1/Ni=1(yprediction(x))2MSE = 1/N \displaystyle\sum_{i=1} (y-prediction(x))^2

where:

  • (x,y)(x,y) is an example in which
  • xx is the set of features (for example, chirps/minute, age, gender) that the model uses to make predictions.
  • yy is the example's label (for example, temperature).
  • prediction(x)prediction(x) is a function of the weights and bias in combination with the set of features xx.
  • DD is a data set containing many labeled examples, which are (x,y)(x,y)pairs.
  • NN is the number of examples in DD.
profile
배움을 나누는 개발자입니다
post-custom-banner

0개의 댓글