-
Currently, the evaluate_model function focuses primarily on accuracy and F1-score for classification models, and MSE and R² for regression models. We could enhance this by including additional evaluat…
-
Recently, [Sakai (2021)](https://aclanthology.org/2021.acl-long.214) compared several class, numeric, and proposed "ordinal" performance measures/metrics on ordinal classification tasks. This raises t…
-
Hi, I'm using keras framework train a model, and tracing the parameters and metrics with MLFLOW like this:
import mlflow
import tensorflow as tf
import mlflow.tensorflow
mlflow.tenso…
-
@sanchit-gandhi
Where can I find faster-whisper model evaluation metrics? I don't see them on ASR leaderboard. Thanks!
-
### What is it?
Currently we have these legacy `code_metrics_by_project`, and `onchain_metrics_by_project` computed in dbt.
With the new metrics factory on sqlmesh, we should create new models. …
-
-
### What is it?
The second part of a set of issues for getting Dagster + SQLMesh Metrics running. This is to get sqlmesh running the python models on trino. If, somehow, this is performant enough wit…
-
Hello,
I'm wondering if the models supported in this project have been evaluated to see if their performance replicates the reported metrics in the original papers (primarily FID and CLIP score f…
gtuzi updated
1 month ago
-
Each time an LLM responds, it also outputs some info about its performance.
```
llama_print_timings: load time = 4732.44 ms
llama_print_timings: sample time = 86.82 ms / 48…
-
I’ve also noticed that the evaluation of the regression model includes classification metrics such as accuracy, precision, recall, F1 score, and confusion matrix. These metrics are specifically desig…