Open FuZixin opened 3 months ago
Hi @FuZixin, thanks for the detailed description of the issue.
Thanks, Arjun
Thank you for your reply:
I've been able to run non-zero predictions by modifying _nonnegative_predsamples=False in the fine-tuning code; I'm speculating that it may be due to the fact that I'm using samples with all negative values?
Sorry I can't provide the dataset directly, here's a sample of the data, I hope it helps: our fine-tuning training set is a time series with 4 thousand data points, with a 1 second acquisition interval, and with 3-bit precision.
Q: However, the predictions based on the fine-tuned model show a large offset from the true value, we are currently working on this issue.
If you have any comments or opinions, please feel free to communicate with us :)
I see, that's very useful to know.
We have updated the repo with best practices for finetuning. Maybe that could help.
Hello @FuZixin , have you maybe set nonnegative_pred_samples=True
when you called LagLlamaEstimator? By setting this True
you only allow the model to give positive predictions.
I'm trying to fine-tune the model using my own data. When using the fine-tuned model to make predictions, I find that lag-llama gives all zero value predictions, which does not reproduce the fine-tuning effect in the demo. I followed the following steps:
Call _from_longdataframe to convert the data from the dataframe format to the standard format.
![20240315-152206(WeLinkPC)](https://github.com/time-series-foundation-models/lag-llama/assets/47811160/916d04c1-6205-46c7-bfbe-01c442b4707f)
The data contains 7000 second-level data points. I use the first 4000 points for fine-tuning training, and the remaining 3000 data points as model input. The task is to predict the last 60 data points.![20240315-152011(WeLinkPC)](https://github.com/time-series-foundation-models/lag-llama/assets/47811160/020cf201-4e01-4a34-8617-49dbccb43a65)
The prediction result is shown in the following figure. The result provided by lag-llama is 60 zero values.![20240315-151958(WeLinkPC)](https://github.com/time-series-foundation-models/lag-llama/assets/47811160/dfef286c-2571-4e73-b460-724daf5c0793)
What is the cause of this, or Which step did I do wrong?