-
A bunch of issues are a bit stale, and @SunMarc + @muellerzr are a bit short on bandwidth!
Thus we would love to have community support to solve the following:
### Help needed
- [ ] #27830
- […
-
I'm testing the pass rate evaluation, could you offer the reproduction data like Toolbench?
Thanks for your reply
-
Hi, thank you for yours for providing the code and the evaluation metrics on your GitHub repository.
When I reproduced the work of your grabnet, the other metrics were similar. However, the simulati…
-
## Rubric Score
### Criteria 1: Valid Python Code
* _Score Level:_ 4/4
* _Comment(s):_ Code included runs without any errors.
### Criteria 2: Exploration of Data
* _Score Level:_ 3/4
* _Co…
-
Hi,
Results are not yet shown on the account. What may be probable issues?
For early submissions, I had some naming differences in the zipped folder than what is expected. I resubmitted the file…
-
Hello @lartpang, thanks for this great lib.
I would like to be able, on a big dataset, to store image-level metrics.
## Context
For example
Image1 :
* WeightedFmeasure : 92%
* Emeasure …
-
Hello, I wrote a script based on the demo_sample.ipynb to generate 50,000 samples and tested them using OpenAI's FID evaluation toolkit. However, I found that the metrics did not align. Could you help…
LiCHH updated
3 weeks ago
-
### Feature Request
Tests are running in a batch but they are not being evaluated against the GT answer.
The evaluation could be done using similarity metrics, LLM or even mathematics.
### M…
-
The beginning of the cluster evaluation metric section should list all the metrics and list whether they are supervised or unsupervised, similarly to how listings are done for the other metrics.
Mayb…
-
First of all, thank you very much for your great paper and code sharing. While working on my paper, I would like to measure the metric presented in the paper to compare my work with wav2lip results.
…