-
Hi @AlexeyAB et al
Is there any code snippet to evaluate a yolov3 trained model with the following metrics as in Coco evaluation
![screenshot from 2019-01-07 14-34-32](https://user-images.gith…
-
## Is there an existing issue for this?
I have searched the existing issues but can't find one
## Is your feature request related to a problem? Please describe the problem.
I would like that the eva…
-
- [ ] [[2303.16634] G-Eval: NLG Evaluation using GPT-4 with Better Human Alignment](https://arxiv.org/abs/2303.16634)
# [2303.16634] G-Eval: NLG Evaluation using GPT-4 with Better Human Alignment
…
-
Hi Roy, I use https://github.com/abdulfatir/gan-metrics-pytorch as suggested from previous issue to calculate FID and KID, but I cannot generate similar evaluation number as in the paper. I sampled 5k…
-
can you give the code for the evaluation metrics?
-
Hi,
I'm currently trying to replicate the performance of Qwen2-Audio on the AIR Bench. However, I noticed that the repository at [AIR-Bench](https://github.com/OFA-Sys/AIR-Bench/blob/main/score_cha…
-
Hi,
Thanks for sharing your great work! we wonder about some details of the evaluation metrics, i.e., which task is evaluated on Y channel.
Best,
Josh
-
How can we objectively evaluate our model?
Some random thoughts below:
1. **Intrusive v.s. non-intrusive metrics**
In speech generation, we generally have 2 kind of metrics: intrusive and non-…
-
![image](https://user-images.githubusercontent.com/6770853/34157250-ca186822-e4fb-11e7-8ea4-8ef9f91aabc2.png)
+ MOTA
+ MOTP
+ MT
+ ML
+ IDS
+ FRAG
![image](https://user-images.githubusercontent…
-
- [x] Precision-recall curve
- [x] Confusion matrix
> https://github.com/Durant35/paper_tracking/blob/master/2005CVPR_Evaluating%20Multi-Object%20Tracking/README.pdf