Hello! I am a university student in Beijing, China, and I am also doing research on multi-modal large model anomaly detection. I am very interested in your work! You gave the "rigion-division" method here, I managed to reproduce it and it's great!
However, I don't know how to calculate "metrics" with GPT-4V. I refer to the No.15 reference(WinClip) of your article. He used "WinClip model" to calculate "score"(score = model(data)). Then he successfully obtain "metrcis"(result_dict = metric_cal(np.array(scores)), gt_list, gt_mask_list, cal_pro=cal_pro)).
But I truly don't know how to use "GPT-4V+Prompt+rigion-division" or other methods to calculate the "score"(I know that as long as I can calculate the "score", "metrics" will come naturally). Can you give me some guidance?
This is my first time on github to submit issue, please forgive me if I'm offended. Thank you very much indeed!
Hello! I am a university student in Beijing, China, and I am also doing research on multi-modal large model anomaly detection. I am very interested in your work! You gave the "rigion-division" method here, I managed to reproduce it and it's great!
However, I don't know how to calculate "metrics" with GPT-4V. I refer to the No.15 reference(WinClip) of your article. He used "WinClip model" to calculate "score"(score = model(data)). Then he successfully obtain "metrcis"(result_dict = metric_cal(np.array(scores)), gt_list, gt_mask_list, cal_pro=cal_pro)).
But I truly don't know how to use "GPT-4V+Prompt+rigion-division" or other methods to calculate the "score"(I know that as long as I can calculate the "score", "metrics" will come naturally). Can you give me some guidance?
This is my first time on github to submit issue, please forgive me if I'm offended. Thank you very much indeed!