Closed JunyaoHu closed 7 months ago
Thank you for your question. We will upload our supplementary materials in the next few days.
For Emo-A, we used the average of eight emotion accuracy. As for other outputs, we wanted to observe the performance differences in generating various emotional images.
For Sem-C, we did take the maximum probability between the scene classifier and object classifier for each picture. This is a new evaluation metric we propose for the task, without reference to other papers.
For Sem-D, we used pixel-level MSE as the evaluation metric. Since the cosine distance and Euclidean distance of normalized vectors are equivalent and the difference between them is not significant in practical calculations, we chose the latter. The 10 image pairs are implemented from the supplementary material of one of the papers we follow. You can see the paper we follow in the LPIPS section of our main text.
Thank you very much for your patience!
First of all, thank you for your work, which has made an important contribution to the combination of image generation and emotion calculation, which is very rare and commendable.
Could you please provide supplementary materials? I would like to know more details about the quality of emotion generation metrics. I believe these questions should have been explained in detail in the supplementary materials.
I have read the relevant codes for three metrics.
https://github.com/JingyuanYY/EmoGen/blob/f560012bf56ff68f5c6edc3dfb9728e9c856ad91/training/inference.py#L285-L292
https://github.com/JingyuanYY/EmoGen/blob/f560012bf56ff68f5c6edc3dfb9728e9c856ad91/metrics/other_metrics.py#L151-L153
https://github.com/JingyuanYY/EmoGen/blob/f560012bf56ff68f5c6edc3dfb9728e9c856ad91/metrics/other_metrics.py#L107-L114
https://github.com/JingyuanYY/EmoGen/blob/f560012bf56ff68f5c6edc3dfb9728e9c856ad91/metrics/other_metrics.py#L160-L164
Thank you!