Open donovaly opened 4 days ago
side note: I noticed that this line https://github.com/JuliaNLSolvers/LsqFit.jl/blob/8f8c18a08abfe4d182bb13ac612a68a205513b3b/src/curve_fit.jl#L261 can most probably removed since it is not used.
I just tested with the KDE program LabPlot: https://labplot.kde.org/ and their covariance matrix is the one I get when I use the MSE. So they use
Q, R = qr(J)
Rinv = inv(R)
covar = Rinv * Rinv' * mse(fit)
also for weighted fits.
Thank you, this is certainly the right place to report the issue. Thank you for taking the time.
I am the maintainer of this package and I also developed some of it. Most of it is something I moved out of Optim because I didn't think it was good to have all the statistics part of Optim. I think now that the LM method should maybe have stayed but I am digressing.
I think you may be right. I also think we had this discussion many times before in what are now closed issues. I will have to go back and consult. I think there are some nuances here that depend on assumptions such was where the weights are put. Are you weighting the residuals or are you weighting the squared residuals. That matters in the end. So I am open to this change, but I want to make sure that we have references and the math in order, because as I said I think this discussion has been had many times before with diverging opinions and I honestly never really use NLS myself :)
Thanks for the reply. However, I am bit confused. I don't know about a package named Optim. I just googled how to perform a NLSF with Julia, found LsqFit, installed it and gave it a try. As a physicist I have to perform NLSFs all the time.
I hope I made claer where I see the bug: when one has weights (meaning data errors) LsqFit does not take the MSE into account. If it would, it would deliver the same results I get with other programs.
To reproduce, here is my test data
xVals = [1, 2, 4, 5, 8]
yVals = [3, 4, 6, 11, 20]
yError = (0.05 .* yVals)
this is the fit model:
f(x) = p1 * exp(p2*x)
Here is a Julia script to reproduce: Goodness-of-fits.zip
Yes, I understand. My imperfect memory just told me that it was set up in that way on purpose. I just can't remember the details :) So I have to read up on it. I remember it as such that the weights are included in the residuals so they carry over to the Jacobian so they indeed not ignored, but just incorporated into J. But I have to check.
Okay, I looked at your code example. One thing that does strike me is that you input the weights as inverse variances, but I believe they should be inverse standard deviations. If you input a matrix instead it is interpreted as a variance-covariance matrix. So vector weights -> standard deviation interpretation and matrix weights -> variance-covariance interpretation.
Some past discussion: #69 #70 #85 #103 #158
Okay, I looked at your code example. One thing that does strike me is that you input the weights as inverse variances, but I believe they should be inverse standard deviations
but I set
wt = @. 1/yError^2
So the inverse std deviation. This is the default weighting when the data is normally distributed, thus also the default in Origin and LabPlot.
If you input a matrix instead it is interpreted as a variance-covariance matrix.
But I want the variance-covariance matrix to be computed. I don't know it yet.
I just have data and know the error of the data and perform a fit. As result I need the errors of the found fit parameters. I think this is the most standard use case.
Some past discussion: #69 #70 #85 #103 #158
from #85
Since the weights are already included in the function which we want to minimize, the correct covarinace should be: `inv(J' * J).
This statement is misleading. Sure, the Jacobian must contain the weighting and LsqFit does this correctly. The point is why the MSE is not used. Think about it:
MSE * inv(J' * J)
for the case of no weights. Why should now, with the presence the MSE become 1, so that you can write 1 * inv(J' * J)
?You can give this a try:
jacobianMatrix = Zygote.jacobian((params, x, error) -> getyValsWithParams(x, params)./error, fitParams, xVals, yError)[1]
to
jacobianMatrix = Zygote.jacobian((params, x) -> getyValsWithParams(x, params), fitParams, xVals)[1]
In https://github.com/JuliaNLSolvers/LsqFit.jl/issues/103#issuecomment-511611397 he says the same.
I think I made it clear:
r = fit.resid
can be omittedI will make a PR with the way I think the fix should be then you can give this to e.g. a mathematician.
@gerlachs is one of the main LabPlot (https://github.com/KDE/labplot) developers, maybe he can give feedback since my PR would change LsqFit so that it delivers the same output as LabPlot.
OK, the proof that there is a bug and that my PR fixes it is simple: when you make a non-weighted fit you must get the same result than when making a weighted fit but setting the weights to 1.0. (This was already mentioned here: https://github.com/JuliaNLSolvers/LsqFit.jl/issues/103#issuecomment-511611397)
Here is a small Julia script with the proof: LsqTest.zip
I do understand the issue (and thank you for the PR). What I was trying to state above is that as far as I remember, the conclusion in the above links were that the behavior is intentional (you can argue what you want the software to do of course), because if you tell me what the uncertainty is at each collected sample, then there is no variance component to estimate and implicitly the variance is set to 1 and your weights will change the variance (sigma_i) accordingly in the calculations. I suppose it has to do with the way you wish to use the weights.
Happy to hear from @gerlachs on this.
I am not opposed to your PR or your comments I am just trying to explain what I seem to remember was the guiding principle way-back-when... and again, I am not a user of NLS I just happened to become the owner of the code in 2017 :)
but I set
wt = @. 1/yError^2
So the inverse std deviation. This is the default weighting when the data is normally distributed, thus also the default in Origin and LabPlot.
How is that the inverse standard deviation? What is yError? Is that not the standard deviation? Then this is the inverse variance. I do not suppose yError is the square root of the standard deviation?
Edit: and to clarify, my point is that if you have a series of observations and a model function and you estimate parameters unweighted then MSE is indeed an estimate of the variance of the residual distribution. However, if you come to me and say "I know this variance" and we weight the residuals by the inverse of the standard deviation derived from it then your transformed residuals should have variance 1 because what you have done is that you have standardized your normal residuals by weighting.
We should focus on the bug. Weighting with a vector containing 1.0 as entries must deliver the same result than no weighting because this is mathematically the same. In my last post I provided a script to check that this is not the case. So there is a clear bug.
My PR fixes the bug. Where do you see my PR is wrong?
I also think there should be a test to assure that in future weighting with 1.0s deliver the same result than no weighting.
I am a beginner with Julia and tried it out by fitting data. I got unexpected results for vcov(fit) and had a look in the source code.
There: https://github.com/JuliaNLSolvers/LsqFit.jl/blob/8f8c18a08abfe4d182bb13ac612a68a205513b3b/src/curve_fit.jl#L256
I noticed 2 issues:
The second point is my main problem because the errors of the fit model parameters were larger than expected. I think it is a bug to omit the MSE. When there are weights, the MSE includes them already, meaning the MSE does not become 1 because of the presence of weights. And therefore it cannot just be omitted.
I the tutorial I see there correctly: https://julianlsolvers.github.io/LsqFit.jl/latest/tutorial/#Weighted-Least-Squares-1
Cov(γ∗)=σ²[J′WJ]^−1
and σ² is the MSE there you also writeBut this is incorrect because the dimensions won't fit and I get the error:
DimensionMismatch
and looking at the source code the calculations iscovar = covar = inv(J' * J)
so the weights are ignored.I use Julia 1.11.1 and LsqFit v0.15.0
p.s. as a newbie I might report the error at the wrong place, please point me then to the correct place.
p.s. the LsqFit docs: https://julianlsolvers.github.io/LsqFit.jl/latest/tutorial I see the command ''estimate_var'' but this is marked deprecated.