-
i already do
`finetune_forecast_trainer.evaluate(test_dataset)`
but how i use it to make inferention ?
-
# Challenge 16 - Air quality dashboard
> **Stream 1 - Data Visualization and visual narratives for Earth Sciences applications**
### Goal
Develop an online dashboard for displaying and monitori…
-
The UMASEP model is essentially a triggered nowcasting model in that it will only produce a positive forecast in response to either a sufficiently strong solar flare or a detected rise in proton flux.…
-
### Describe the issue linked to the documentation
link: https://auto.gluon.ai/stable/tutorials/timeseries/forecasting-metrics.html
> AutoGluon will use the provided metric to tune model hyperpara…
-
Ref. https://github.com/sktime/sktime/actions/runs/7932291556/job/21661667361#step:3:1533
> FAILED sktime/forecasting/model_evaluation/tests/test_evaluate.py::test_evaluate_error_score[scores0-back…
-
Eventually, it could be great to have strong "evaluation bodies" that could be forecasted upon.
One first step would just be to set up a few evaluation bodies, and have them make a few group evaluati…
-
**Is your feature request related to a problem? Please describe.**
In the forecasting module, the evaluate function currently lets a user use timeseries cross-validation to compare different forecast…
-
## Description
Evaluation multiple windows took too long(25Min!!!). I am tiered from this package. It is real garbage!
## To Reproduce
```python
from gluonts.dataset.common import DataEntry,…
-
"I read the blog post (https://huggingface.co/blog/zh/time-series-transformers) and I'm really confused about the evaluation metrics. I would be extremely grateful if I could get some help. Specifical…
-
# Time series Forecasting in Python & R, Part 2 (Forecasting ) | Sandeep Pawar
In the second part of this blog series on forecasting I discuss forecasting steps, evaluation of forecasting methods, mo…