-
Hi,
I like your paper. As for the evaluation metrics, I can see your are using recall@k to track the performance. May I ask have you tried to implement other metrics like NMI and MAP@R? Thanks!
-
We need some methods/scripts to evaluate parsing performance. We probably want to do two things: a) replicate previous work that uses parseval so that we can easily report previous results (see table…
-
Is the code of the CDDB-Hard for 'forgetting' calculation provided in this code base?
-
Thank u for your grand project. It really helps a lot. But now I'm in trouble evaluating the task-specific motionGPT to tackle with motion translation. It seems that when the model is training on m2t …
-
**Describe the Feature**
Can you could provide the human assessment data collected for bechmarking RAGAS metrics against human evaluations in your [paper](https://arxiv.org/pdf/2309.15217)?
**Why …
-
Is the evaluation metric public?
Please share how the evaluation metric is computed
-
Thanks for the release of this great repo!
I am interested in your proposed quantitative evaluation metrics (i.e., D-VL, D-A, D-LL, D-Rot/Pos). A standard evaluation metrics will make the following…
-
### Issues Policy acknowledgement
- [X] I have read and agree to submit bug reports in accordance with the [issues policy](https://www.github.com/mlflow/mlflow/blob/master/ISSUE_POLICY.md)
### Where…
-
Let's discuss evaluation metrics for `spikefinder`!
The data for doing evaluation will be, for each neuron, a fluorescence trace and an estimated spike rate. Presumably the rate will be either binar…
-
Several information retrieval "tasks" use a few common evaluation metrics including mean average precision (MAP) [1] and recall@k, in addition to what is already supported (e.g. ERR, nDCG, MRR). Somet…