Closed zhimin-z closed 6 months ago
Are the benchmarks supported by LAMM and ChEF different by design? @shepnerd @orashi @double125 @yinzhenfei
Sorry for the late response. ChEF did not design any evaluation pipeline for 3D benchmark. Please refer to LAMM paper for 3D leaderboard. We will update the 3D leaderboard on the website sooner. Thanks for your feedback.
Sorry for the late response. ChEF did not design any evaluation pipeline for 3D benchmark. Please refer to LAMM paper for 3D leaderboard. We will update the 3D leaderboard on the website sooner. Thanks for your feedback.
Thanks, what datasets are supported in CheF, only those appearing in the official leaderboards?
Yes, and you can also add other datasets for evaluation in the dataset class.
Yes, and you can also add other datasets for evaluation in the dataset class.
Is there a comprehensive list of the supported benchmarks in CHeF?
As should in the leaderboard, the supported benchmarks are MME, SeedBench and MMBench. The supported datasets are CIFAR, Flickr, VOC, Omnibenchmark , FSC147 and ScienceQA.
As should in the leaderboard, the supported benchmarks are MME, SeedBench and MMBench. The supported datasets are CIFAR, Flickr, VOC, Omnibenchmark , FSC147 and ScienceQA.
Thanks for your reply, does ChEF benchmark support MSCOCO as well? @Coach257
ChEF supports the POPE benchmark, which is built on MSCOCO dataset. You can refer to Evaluating Object Hallucination in Large Vision-Language Models for more details.
The claimed supportive tasks fail to show in the leaderboards...