Scientific Image Quality Assessment Challenge (SIQA)

ICME 2026 · Advancing AI’s Ability to Evaluate Scientific Integrity in Visual Data

Introduction

Multimodal large language models (MLLMs) are revolutionizing image quality assessment (IQA) by bringing semantic understanding and world knowledge into evaluation. Yet nearly all existing IQA benchmarks ignore scientific imagery, a cornerstone of research, education, and discovery.

Unlike general photos, where quality is judged by blur or noise, scientific images must also be correct, complete, clear, and conventional. A high-quality scientific visualization must: accurately reflect scientific facts (Validity); include all necessary labels, scales, and context (Completeness); be instantly interpretable by experts (Clarity); and follow field-specific norms in style and notation (Conformity).

To address this gap, the SIQA Challenge introduces two tasks that push models beyond pixels—to evaluate scientific integrity through visual reasoning and domain-aware judgment.

Task Example

Challenge Tasks

🔍 SIQA Understanding (SIQA-U)

Evaluates a model’s ability to reason about scientific image quality through structured visual question answering aligned with the four SIQA dimensions.

Question Types:

  • Yes/No: Binary verification of factual, structural, or representational conditions.
  • What: Multiple-choice comprehension of scientific entities, relationships, or context.
  • How: Quality judgment on completeness, clarity, and disciplinary conventions.

Measures factual verification, semantic understanding, and scientific reasoning.

Evaluation: Weighted accuracy:
Score₁ = 0.2 × ACCyes/no + 0.3 × ACCwhat + 0.5 × ACChow

📈 SIQA Scoring (SIQA-S)

Predicts continuous quality scores along two complementary dimensions:

  • Factual Correctness (objective): Alignment with ground-truth science (validity + completeness).
  • Perceptual Quality (subjective): Human-expert judgments on clarity and discipline-specific usability (clarity + conformity).

Models predict scores directly from images—no text input required.

Evaluation: For each dimension d ∈ {Factual, Perceptual}:
Score(d) = max( (SRCC(d) + PLCC(d)) / 2 , 0 ) × 100
Final Score₂ = (ScoreFactual + ScorePerceptual) / 2

Task Example

📊 Dataset Overview

The SIQA dataset is built around two complementary tasks: SIQA-U for structured reasoning and SIQA-S for continuous quality scoring. These are grounded in four scientific dimensions and two orthogonal evaluation axes.

SIQA Dataset Overview

🔍 Key Design Principles

1. SIQA-U: 4 Dimensions × 3 Question Types

Each image is evaluated through three question types across four scientific dimensions:

  • Yes/No: Binary verification of factual correctness (e.g., "Does this diagram show a correct bond?")
  • What: Multiple-choice comprehension of scientific elements (e.g., "Which labeling violates convention?")
  • How: Qualitative judgment on completeness, clarity, and disciplinary alignment.

2. SIQA-S: Perception vs. Knowledge — Orthogonal Evaluation Axes

SIQA-S scores images along two orthogonal dimensions:

  • Knowledge-Driven: Factual correctness (validity, completeness) — does the image convey accurate science?
  • Perception-Driven: Cognitive clarity (clarity, conformity) — is it easy to understand and follow domain conventions?

This design enables models to be evaluated on both what they know and how well they communicate it.

Timeline

February 7, 2026
Registration opens at, and delever confirm mail at ; Training dataset released at huggingface
March 10, 2026
Validation dataset released huggingface
April 1, 2026
Registration deadline
April 25, 2026
Code & model submission deadline, submission your code at github.com//; Internal evaluation begins
May 4, 2026
Paper submission deadline at official mail//
May 15, 2026
Final results announced (to be posted on results page)

Top-performing teams will be invited to extend their solutions into a full paper for publication in Displays journal.

Organizers

Wenzhe Li
Shanghai Artificial Intelligence Laboratory
Liang Chen
Shanghai Artificial Intelligence Laboratory
Junying Wang
Shanghai Artificial Intelligence Laboratory
Farong Wen
Shanghai Artificial Intelligence Laboratory
Yijing Guo
Shanghai Artificial Intelligence Laboratory
Ye Shen
Shanghai Artificial Intelligence Laboratory
Qihang Yan
Shanghai Artificial Intelligence Laboratory

In partnership with universities and open-data initiatives worldwide.

Advisory Committee

Zicheng Zhang*
Shanghai Artificial Intelligence Laboratory
Chunyi Li
Shanghai Artificial Intelligence Laboratory
Wenlong Zhang
Shanghai Artificial Intelligence Laboratory
Wei Zhou
Cardiff University
Xiaohong Liu
Shanghai Jiao Tong University
Xiongkuo Min
Shanghai Jiao Tong University
Guangtao Zhai
Shanghai Artificial Intelligence Laboratory

* means the project lead