Skip to main content
Log in

A multimodal dense convolution network for blind image quality assessment

一种针对盲图像质量评估的多模态密集卷积网络

  • Research Article
  • Published:
Frontiers of Information Technology & Electronic Engineering Aims and scope Submit manuscript

Abstract

Technological advancements continue to expand the communications industry’s potential. Images, which are an important component in strengthening communication, are widely available. Therefore, image quality assessment (IQA) is critical in improving content delivered to end users. Convolutional neural networks (CNNs) used in IQA face two common challenges. One issue is that these methods fail to provide the best representation of the image. The other issue is that the models have a large number of parameters, which easily leads to overfitting. To address these issues, the dense convolution network (DSC-Net), a deep learning model with fewer parameters, is proposed for no-reference image quality assessment (NR-IQA). Moreover, it is obvious that the use of multimodal data for deep learning has improved the performance of applications. As a result, multimodal dense convolution network (MDSC-Net) fuses the texture features extracted using the gray-level co-occurrence matrix (GLCM) method and spatial features extracted using DSC-Net and predicts the image quality. The performance of the proposed framework on the benchmark synthetic datasets LIVE, TID2013, and KADID-10k demonstrates that the MDSC-Net approach achieves good performance over state-of-the-art methods for the NR-IQA task.

摘要

科技进步不断扩大通信行业的潜力. 图像在加强交流中发挥着重要作用, 已被广泛应用. 因此, 图像质量评估(IQA)对优化传递给终端用户的内容至关重要. 在IQA中使用卷积神经网络面临两个常见难题. 一是这些方法难以提供图像最佳表示, 另一个问题是模型具有大量参数, 容易导致过拟合. 为解决这些问题, 提出一种参数更少的深度学习模型——密集卷积网络(DSC-Net), 用于无参考图像质量评估(NR-IQA). 此外, 将多模态数据用于深度学习明显改进各种应用的性能. 多模态密集卷积网络(MDSC-Net)融合了灰度共生矩阵(GLCM)方法提取的纹理特征和DSC-Net方法提取的空间特征, 并对图像质量进行预测. 所提框架在基准合成数据集LIVE、 TID2013和KADID-10k的性能表明, MDSC-Net方法在NR-IQA任务中表现出良好性能, 超过了当前最先进的方法.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Data availability

The data that support the findings of this study are available from the corresponding author upon reasonable request.

References

Download references

Author information

Authors and Affiliations

Authors

Contributions

Nandhini CHOCKALINGAM designed the research, processed the data, and drafted the paper. Brindha MURUGAN helped organize and revise the paper. Nandhini CHOCKALINGAM and Brindha MURUGAN finalized the paper.

Corresponding author

Correspondence to Brindha Murugan.

Ethics declarations

Nandhini CHOCKALINGAM and Brindha MURUGAN declare that they have no conflict of interest.

Additional information

List of supplementary materials

1 Description of the dataset

2 Details of evaluation metrics

3 Local contrast normalization (LCN)

4 Importance of GLCM features for influential performance

5 Visualization of distortion specific prediction

6 Effect of patch size

7 Performance of pre-trained DenseNet

8 Visualization of patchwise training strategy

9 Convergence analysis

Supplementary materials for

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Chockalingam, N., Murugan, B. A multimodal dense convolution network for blind image quality assessment. Front Inform Technol Electron Eng 24, 1601–1615 (2023). https://doi.org/10.1631/FITEE.2200534

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1631/FITEE.2200534

Key words

关键词

CLC number

Navigation