Science-T2I

Addressing Scientific Illusions in Image Synthesis

Science-T2I Icon
Science-T2I: An expert-annotated dataset of over 20k adversarial image pairs and 9k prompts across 16 scientific domains, with an isolated test set of 454 prompts for benchmarking image generation models.
Generalization Analysis Icon
SciScore: A reward model fine-tuned from CLIP-H that captures fine-grained scientific phenomena, surpassing GPT-4o and human evaluators by roughly 5 points.
Tuning Icon
Reality Alignment: A two-stage framework combining supervised fine-tuning with masked online fine-tuning that improves FLUX.1[dev] by over 50% relative improvement.
Teaser Image
Science-T2I
Fig 1: Given prompts requiring scientific knowledge, image generation models produce scientifically incorrect outputs, and LMMs like GPT-4o fail to identify the errors. We introduce Science-T2I (dataset + benchmark), SciScore (reward model surpassing human evaluators), and a two-stage alignment framework that substantially improves scientific reasoning in image generation.

Current image generation models produce visually compelling but scientifically implausible images, exposing a fundamental gap between visual fidelity and physical realism. We introduce Science-T2I, an expert-annotated dataset comprising over 20k adversarial image pairs and 9k prompts across 16 scientific domains, along with an isolated test set of 454 challenging prompts. Using this benchmark, we evaluate 18 recent image generation models and find that none scores above 50 out of 100 under implicit scientific prompts, while explicit prompts yield scores roughly 35 points higher, confirming a systematic composition–reasoning disconnect. To address this, we develop SciScore, a reward model fine-tuned from CLIP-H that surpasses GPT-4o and human evaluators by roughly 5 points. We further propose a two-stage alignment framework combining supervised fine-tuning with masked online fine-tuning. Applying this framework to FLUX.1[dev] yields a relative improvement exceeding 50%, demonstrating that scientific reasoning in image generation can be substantially improved through targeted data and alignment.

Dataset Dataset Benchmark Benchmark SciScore SciScore Alignment Alignment

Click to jump to each section.


Science-T2I: Bridging Visual Imagination and Scientific Realism

Current image generation models produce visually plausible but scientifically incorrect outputs when prompted with descriptions that require physical reasoning. We identify two root causes: (1) existing training data rarely pairs scientific concepts with their correct visual manifestations, and (2) standard evaluation protocols do not test whether a model understands the science behind a prompt. To address both issues, we introduce Science-T2I, a dataset that challenges models to perform implicit scientific reasoning.

Task overview. As shown below, Science-T2I consists of 16 tasks spanning physics, chemistry, and biology. Each task requires the model to infer or visualize a concept not explicitly stated in the prompt but rooted in an underlying scientific principle.

Fig 2: (Left) Science-T2I is organized into three scientific fields, each divided into specific categories. (Right) Word cloud of structured prompts in Science-T2I.

Task classification. Beyond scientific discipline, the tasks naturally fall into two categories that reveal distinct reasoning demands:

1. Subject-oriented Tasks (ST) require reasoning about how inherent differences between subjects lead to varying visual features under identical conditions. For example, different metals produce different flame colors.

2. Condition-oriented Tasks (CT) focus on how a single condition affects the visual appearance of various subjects. Here, scientific reasoning centers on the applied condition rather than the subject's individual properties. For example, all objects float in the absence of gravity.

Fig 3: Subject-oriented vs. condition-oriented task classification. Ten tasks are condition-oriented and six are subject-oriented.

Prompt design. A central design choice is the three-tier prompt structure, which disentangles a model's scientific reasoning ability from its compositional rendering ability. For each task, we construct a tuple of three prompts:

1. Implicit Prompt (IP) contains terms that imply certain visual characteristics requiring interpretative reasoning based on scientific knowledge. For example, "an unripe apple" suggests greenness without explicitly stating it.

2. Explicit Prompt (EP) reformulates the implicit prompt into a clear, descriptive statement that directly conveys the intended visual outcome. For instance, "a green apple" makes the expected appearance explicit.

3. Superficial Prompt (SP) provides a plausible but scientifically incorrect interpretation, focusing only on surface-level associations. For example, "a red apple" interprets "unripe" based on the default visual prototype rather than the scientific implication.

Together, the three prompt types serve complementary roles: the IP tests implicit reasoning, the EP establishes an upper bound on what the model can render when given direct instructions, and the SP provides a hard negative for preference-based training.

Fig 4: Data curation pipeline. GPT-4o generates structured templates, then expands each implicit prompt into explicit and superficial counterparts, guiding the synthesis of corresponding images.

Data curation. We leverage GPT-4o to generate structured templates and corresponding prompts for each task. The pipeline first produces implicit prompts from task-specific templates, then expands each into its explicit and superficial counterparts. These prompts are used to generate image pairs via image generation models. All generated data undergoes manual verification by domain experts, who cross-reference each tuple against established scientific knowledge.

Science-T2I Examples


Benchmarking Scientific Image Synthesis

Test Splits. We construct two manually annotated test sets covering the same scientific domains as the training data but containing no overlapping samples:

1. Science-T2I-S (Simple), containing 227 tuples. This split uses minimalist compositions with simple backgrounds. By isolating the scientific subject from environmental distractions, it provides a controlled test of whether the model can generate the correct scientific visual features.

2. Science-T2I-C (Complex), containing 227 tuples. This split embeds the same scientific tasks within diverse real-world scenes, adding contextual elements such as "in a bedroom" or "on the street." It tests whether the model can maintain scientific accuracy when the background is visually complex.

Science-T2I-S&C Leaderboard

🤩 Submit your LMM/VLM scores now and watch the leaderboard refresh with your achievements! Email us at .

Click the button below to view different results!

#ModelOverallPhysicsChemistryBiology
AccGRSOMEABBUDIELEVLIAvgRUIMFRAvgLRWRSCRIAvg
154.778.340.525.057.163.971.447.626.777.855.116.754.277.852.462.231.181.534.655.9
255.047.544.156.938.150.050.052.420.033.350.442.953.176.743.176.738.958.338.559.9
357.278.345.244.457.158.383.347.663.358.359.623.860.462.253.246.733.383.353.955.9
463.883.342.926.440.547.270.277.468.384.760.073.866.752.267.057.867.895.434.668.8
565.192.556.036.138.145.875.077.410095.868.259.555.248.957.851.172.278.746.264.7
687.093.086.198.266.774.665.995.610082.187.792.977.881.075.996.999.690.794.695.3
770.896.752.441.747.655.663.172.691.790.367.869.156.352.262.284.490.084.375.084.4
870.871.335.736.133.356.977.482.110076.462.095.265.658.973.896.783.397.253.986.8
993.198.390.510071.466.797.610010010094.910068.897.881.0100100100100100

Image Generation Benchmark

Evaluation Setup. We evaluate 18 image generation models on the Science-T2I test set using implicit prompts. Each model generates two images per prompt, scored by Qwen3.5-27B using per-tuple grading criteria. We report the normalized Reality Score (0–100) across Physics, Chemistry, and Biology.

#ModelOverallPhysicsChemistryBiology
1SDXL19.6016.1120.9225.56
2FLUX.1[schnell]23.7223.1938.3013.33
3LongCat-Image23.8621.9439.0115.83
4GLM-Image24.2322.2241.8414.44
5SANA26.1423.8946.4514.72
6Lumina-Image-226.5120.2847.8722.22
7SD 3.5 Medium26.7323.0639.0124.44
8Z-Image26.7326.5332.9822.22
9FLUX.1[dev]26.8722.6450.0017.22
10SD 3.5 Large28.7127.6444.6818.33
11Bagel29.0025.2845.7523.33
12FLUX.2[klein-4B-Base]29.0827.2243.9721.11
13Z-Image-Turbo29.2226.8136.5328.33
14FLUX.2[klein-4B]30.8428.7550.3619.72
15FLUX.2[klein-9B-Base]31.5730.2845.7523.06
16FLUX.2[klein-9B]32.6029.4457.4519.44
17Qwen-Image33.9934.5846.4523.06
18FLUX.2[dev]47.8053.1953.5532.50

Current models lack scientific reasoning. Even the best model (FLUX.2[dev]) scores only 47.80 out of 100, and the majority cluster between 20 and 35. Biology poses the greatest challenge: no model exceeds 33%. When evaluated with explicit prompts that directly describe the intended scientific outcome, scores increase by roughly 35 points on average, confirming that the bottleneck is not in visual rendering but in scientific reasoning.

SciScore: A Reward Model for Scientific Image Assessment

CLIP aligns textual and visual data effectively for general purposes, but struggles with implicit scientific prompts: it tends to embed implicit prompts closer to their superficial counterparts than their explicit counterparts. We introduce SciScore, a reward model fine-tuned on the Science-T2I training set that extends CLIP-H to assess whether a generated image reflects the scientific principles implied by a prompt.

Tab 1: (Left) SciScore surpasses all VLMs, LMMs, and human evaluators on both test splits. (Right) Nearly all failures concentrate in subject-oriented tasks.

SciScore surpasses human evaluators. SciScore achieves 93.14 on Science-T2I-S and 91.19 on Science-T2I-C, surpassing human evaluators (87.01 and 86.02) by roughly 6 and 5 points, respectively. The strong performance on Science-T2I-C is particularly notable, as this split embeds scientific reasoning tasks within diverse environmental contexts not present during training.

Subject-oriented tasks remain the primary challenge. Nearly all failure cases concentrate in subject-oriented tasks, which demand subject-specific knowledge (e.g., which metals produce which flame colors). Condition-oriented tasks, by contrast, involve generalizable visual patterns that transfer more easily across subjects.

Reality Alignment: Two-stage Fine-tuning Framework

We combine the Science-T2I training set with SciScore as a reward signal to close the reasoning gap. Our framework proceeds in two stages: SFT exposes the model to scientific visual phenomena it has never encountered during pretraining, and OFT then optimizes for the implicit reasoning ability measured by SciScore.

Supervised Fine-tuning. Pre-trained models have never been exposed to images depicting scientific phenomena, so no amount of preference optimization over existing outputs can teach them what these phenomena look like. We therefore begin with supervised fine-tuning on the Science-T2I training set using FLUX.1[dev] as our base model.

Masked Online Fine-tuning. SFT teaches the model what scientific images look like, but does not directly optimize for the ability to infer the correct visual outcome from an implicit prompt. We apply online fine-tuning with SciScore as the reward signal and a subject-based masking strategy that restricts gradients to the scientifically relevant region.

Fig 5: Online fine-tuning uses SciScore and subject-based masking to align generation with scientific principles. For each prompt, two images are generated and scored by SciScore to determine preference. GroundingDINO extracts subject masks from each image, restricting gradient propagation to the scientifically relevant regions.

Results and Ablation

Relative improvement metric. Raw SciScore values are useful for comparing models but do not reveal how close a fine-tuned model is to its own ceiling. We observe that SciScore under explicit prompts consistently surpasses that under implicit prompts, providing a natural upper bound. To quantify progress toward this ceiling, we define the Relative Improvement (RI) metric: \[ \begin{equation} RI = \frac{\text{SciScore}_F^{IP}-\text{SciScore}_B^{IP}}{\text{SciScore}_B^{EP}-\text{SciScore}_B^{IP}} \end{equation} \] An RI of 100% would indicate that fine-tuning has fully closed the gap between implicit and explicit prompts.

SFT and OFT together exceed the baseline by over 50%. The combined two-stage framework increases the score from 23.56 to 28.52 on Science-T2I-S (RI = 53.39%) and from 27.26 to 30.11 on Science-T2I-C (RI = 38.31%). SFT contributes the larger share of the improvement: it provides the foundational knowledge of what scientific images look like, while OFT refines the model's ability to activate that knowledge from implicit prompts.

Tab 2: The two-stage framework improves FLUX.1[dev] by over 50% in RI.

Generalization to Complex Scenes. While the training set primarily contains images in straightforward scenarios, the fine-tuned model shows clear improvement on Science-T2I-C. This indicates that the model has internalized underlying scientific principles rather than memorizing training examples.

Fig 6: SFT is essential before OFT, and masking stabilizes training. Without prior SFT (purple), OFT fails to improve SciScore. Without masking (yellow/red), performance becomes erratic or stalls.

OFT without SFT fails. Without the scientific knowledge base provided by SFT (purple curve), the model receives two scientifically incorrect images; the preference signal between two poor samples provides insufficient gradient information for meaningful learning.

Masking suppresses noise from irrelevant features. Without masking (yellow curve), the model collapses because it tries to match all visual features of the preferred image. Halving the learning rate (red) prevents collapse but fails to improve SciScore. The masked configuration (blue) produces stable and consistent improvement.

Fig 7: The two-stage framework corrects scientific errors while preserving visual quality. Upper row: base FLUX.1[dev]. Lower row: our fine-tuned model. Each pair uses an identical random seed. Displayed prompts are simplified summaries.

Conclusion

We investigate the gap between visual fidelity and scientific realism in image generation. We construct Science-T2I, an expert-annotated dataset of over 20k adversarial image pairs spanning 16 scientific domains, and use its test set to benchmark 18 recent models. We find that no model scores above 50 under implicit prompts, confirming that the bottleneck is not in visual rendering but in scientific reasoning. To address this, we develop SciScore, a reward model that surpasses LMMs and human evaluators by roughly 5 points. Building on SciScore, we propose a two-stage alignment framework that yields over 50% relative improvement on FLUX.1[dev]. These results demonstrate that scientific reasoning in image generation is not an inherent limitation but a consequence of missing data and misaligned training objectives.

BibTeX

@misc{li2025sciencet2iaddressingscientificillusions,
  title={Science-T2I: Addressing Scientific Illusions in Image Synthesis},
  author={Jialuo Li and Wenhao Chai and Xingyu Fu and Haiyang Xu and Saining Xie},
  year={2025},
  eprint={2504.13129},
  archivePrefix={arXiv},
  primaryClass={cs.CV},
  url={https://arxiv.org/abs/2504.13129},
}