①[ICLR2024 Spotlight] (GPT-4V/Gemini-Pro/Qwen-VL-Plus+16 OS MLLMs) A benchmark for multi-modality LLMs (MLLMs) on low-level vision and visual quality assessment.
Good job!
I'm wondering whether I can test my own model with q-bench offline.
I've checked the A1 annotation json file and found an unknown field (which may have some relation with the sub-type distortion/others/in-context distortion/...).
Also, the golden descriptions of A2 have not yet been released.
I'm wondering what is the difference between the two subset llvisionqa_dev and llvisionqa_dev? I can't find any clues from the paper and your project page.