Is your feature request related to a problem? Please describe.
Currently, CLIP-IQA uses ["Good photo.", "Bad photo."] as anchors.
However, a more fine-grained assessment should be possible according to the paper.
If you refer to the table 3 of the paper, you will see pairs like:
["Clean photo.", "Noisy photo."], ["Sharp photo.", "Blurry photo."], etc.
Would it be possible for such prompts to be supported?
Describe the solution you'd like
In the clip_iqa.py file, pre-computed tokens are downloaded for "Good photo." and "Bad photo.".
Would it be possible to add pre-computed tokens for other prompts as well?
Or, can you suggest a quick way for us to generate those tokens by ourselves?
Thank you!
Is your feature request related to a problem? Please describe.
Currently, CLIP-IQA uses ["Good photo.", "Bad photo."] as anchors.
However, a more fine-grained assessment should be possible according to the paper.
If you refer to the table 3 of the paper, you will see pairs like:
["Clean photo.", "Noisy photo."], ["Sharp photo.", "Blurry photo."], etc.
Would it be possible for such prompts to be supported?
Describe the solution you'd like
In the clip_iqa.py file, pre-computed tokens are downloaded for "Good photo." and "Bad photo.".
Would it be possible to add pre-computed tokens for other prompts as well?
Or, can you suggest a quick way for us to generate those tokens by ourselves?
Thank you!