EdwardRaff / Inside-Deep-Learning

Inside Deep Learning: The math, the algorithms, the models
231 stars 74 forks source link

some typos in ch 2 #7

Open murphyk opened 1 year ago

murphyk commented 1 year ago

-p47 you write W_{d,c} instead of W^{d,c}

murphyk commented 1 year ago

Also on p56-p57 there is an inconsistency between y-hat meaning logits (colored equation for loss(sm(yhat), y)) and y-hat meaning probabilities (code snippet, y_hat = F.softmax(logits)). Maybe call the latter p_hat?

yebangyu commented 1 year ago

Wow,murphyk is also reading this book ? so cool

I am confused about the notation, too. on P56 there is a:

loss(yhat,y) = -log(sm(yhat), y))

accordinng to the text, yhat is a vector and y is an index. so yhat is a vector and y is a scalar ?

it is so weird to say a loss between a vector and a scalar because loss often means to the things which are of same level?

And, what does sm(yhat), y) mean ? the prob of yth element ?

if it is , this is confusing. what does yth mean?

if not, there is an inconsistency (since On P53, sm(x)i means the prob of ith element )

@EdwardRaff