SFA
[official] No reference image quality assessment based Semantic Feature Aggregation, published in ACM MM 2017, TMM 2019
Install / Use
/learn @lidq92/SFAREADME
Which Has Better Visual Quality: The Clear Blue Sky or a Blurry Animal?
Description
SFA code for the following papers:
- Li, Dingquan, Tingting Jiang, Weisi Lin, and Ming Jiang. "Which Has Better Visual Quality: The Clear Blue Sky or a Blurry Animal?." IEEE Transactions on Multimedia, vol. 21, no. 5, pp. 1221-1234, May 2019.
- Li, Dingquan, Tingting Jiang, and Ming Jiang. "Exploiting High-Level Semantics for No-Reference Image Quality Assessment of Realistic Blur Images." Proceedings of the 2017 ACM on Multimedia Conference. ACM, 2017.
Requirement
Framework: Caffe 1.0 + MATLAB 2016b Interface
The PLSR model uesd in the test code is trained on LIVE gblur images with DMOS (the larger the worse). w and best_layer in the journal extension are determined by five cross-validation (See TMMinter.m).
The ResNet-50-model.caffemodel is downloaded from KaimingHe/deep-residual-networks and it should be pasted into the directory models/ before you run the code!
It's about 100MB which is too large to upload to this repo.
If you have difficulty, you can also download the ResNet-50-model.caffemodel in my sharing on BaiduNetDisk with password u8sd.
New! We provide the PyTorch implementation of the method in SFA-pytorch
Notes
Note for training
All we need to train is a PLSR model, where the training function is plsregress in MATLAB. The features are extracted from the DCNN models pre-trained on the image classification task.
Update: remember to change the value of "im_dir" and "im_lists" in data info.
Note for datasets
You can download the datasets used in the papers from their owners for research purpose. If you have difficulty, you can refer to my sharing on BaiduNetDisk with password cu9j. We only consider the blur related images in this work.
Note for cross dataset evaluation
The reported Spearman correlation (SROCC) is multiplied by -1 when the training and testing datasets have different forms of subjective scores, i.e., one is MOS and the other is DMOS. This is to make sure that the prediction monotonicity is better when SROCC is closer to 1.
Citation
Please cite our papers if it helps your research:
<pre><code>@arcticle{li2018which, title={Which Has Better Visual Quality: The Clear Blue Sky or a Blurry Animal?}, author={Li, Dingquan and Jiang, Tingting and Lin, Weisi and Jiang, Ming}, booktitle={IEEE Transactions on Multimedia}, volume={21}, number={5}, pages={1221-1234}, month={May}, year={2019}, doi={10.1109/TMM.2018.2875354} }</code></pre>[Paper]
<pre><code>@inproceedings{li2017exploiting, title={Exploiting High-Level Semantics for No-Reference Image Quality Assessment of Realistic Blur Images}, author={Li, Dingquan and Jiang, Tingting and Jiang, Ming}, booktitle={Proceedings of the 2017 ACM on Multimedia Conference}, pages={378--386}, year={2017}, organization={ACM} }</code></pre>Contact
Dingquan Li, dingquanli AT pku DOT edu DOT cn.
Related Skills
qqbot-channel
347.0kQQ 频道管理技能。查询频道列表、子频道、成员、发帖、公告、日程等操作。使用 qqbot_channel_api 工具代理 QQ 开放平台 HTTP 接口,自动处理 Token 鉴权。当用户需要查看频道、管理子频道、查询成员、发布帖子/公告/日程时使用。
docs-writer
100.1k`docs-writer` skill instructions As an expert technical writer and editor for the Gemini CLI project, you produce accurate, clear, and consistent documentation. When asked to write, edit, or revie
model-usage
347.0kUse CodexBar CLI local cost usage to summarize per-model usage for Codex or Claude, including the current (most recent) model or a full model breakdown. Trigger when asked for model-level usage/cost data from codexbar, or when you need a scriptable per-model summary from codexbar cost JSON.
Design
Campus Second-Hand Trading Platform \- General Design Document (v5.0 \- React Architecture \- Complete Final Version)1\. System Overall Design 1.1. Project Overview This project aims t
