Riverflow logo

ResearchHype Edit 1 Benchmark

Hype-Edit-1 introduces a new reliability benchmark for frontier image editing models, helping teams compare instruction-following and edit consistency.

  • Sourceful Research
  • Image Editing
  • Benchmarks
  • Reliability
Richard Allen avatar

Richard Allen

February 2, 2026

Hype-Edit-1 benchmark
arXiv paperGitHub dataset

Sourceful Research is introducing our first image editing benchmark, Hype-Edit-1, designed to measure reliability in frontier image editing models. As editing capabilities advance, the gap between impressive demos and dependable, repeatable edits in real workflows keeps getting wider. Hype-Edit-1 aims to close that gap by providing a shared way to evaluate reliability across models.


What Hype-Edit-1 focuses on

Hype-Edit-1 is built to evaluate whether models can follow editing instructions precisely while preserving everything that should remain unchanged. It emphasizes practical editing reliability over purely aesthetic outcomes, giving teams a clearer signal on how a model behaves in real-world editing workflows.

Crucially, it measures the difference between the capability of the model on a task versus its ability to do it reliably.


Results

Combined results (generated at 2026-01-22). 100 tasks, 10 images per task, 1,000 total images.

ModelPass ratePass@4Pass@10Expected attemptsEffective cost/successHype gap
riverflow-2-b182.7%90.5%93.0%1.40$0.666.0%
gemini-3-pro-preview63.8%79.9%87.0%1.85$0.9517.0%
gpt-image-1.561.2%70.3%77.0%2.04$1.3016.0%
flux-2-max45.7%63.8%75.0%2.38$1.4125.0%
qwen-image-edit-251145.4%57.4%66.0%2.48$1.3325.0%
seedream-4.035.6%57.4%72.0%2.64$1.4238.0%
seedream-4.534.4%59.9%77.0%2.63$1.3937.0%

Benchmark design

Hype-Edit-1 combines a curated task set with a consistent evaluation protocol so results can be compared across labs, models, and release cycles. The benchmark is designed to be model-agnostic and to avoid overfitting to any single editing style or capability.

Collage of 10 attempts for each model for the prompt: 'Change the color of the background to match the tshirt exactly. Keep everything else the same.' Model A is consistent, whereas Model B is wildly unreliable.
Collage of 10 attempts for each model for the prompt: 'Change the color of the background to match the tshirt exactly. Keep everything else the same.' Model A is consistent, whereas Model B is wildly unreliable.

Evaluation protocol

Each task is well-defined and is focused on targeted edits instead of ambiguous, creative instructions. We evaluated using a panel of 5 humans who judged pass/fail without seeing which model generated the edit, and used majority voting to determine the result per task instance. We also provide example code for a VLM-based judge using Gemini 3 Flash Preview, which has around 80% agreement with humans. We've found that the VLM-based judge tends to be more strict than humans, picking up detail changes or color shifts that are barely perceptible to most humans.


Why reliability matters

For brand, packaging, and marketing assets, small inaccuracies can introduce costly rework. Reliability is what turns an image editing model from a creative toy into a dependable production tool. Hype-Edit-1 gives researchers and developers a common language to compare model behavior and to track progress over time.

Without a greater focus on reliability, the relationship between humans and AI remains uneasy due to a feeling of being an inconsistent work partner from image to image.


Example tasks

We split tasks into 4 categories: Change (50%), Remove (21%), Restructure (17%) and Enhance (12%).

Example of a Change Task
Example of a Change Task.
Example of a Remove Task
Example of a Remove Task.
Example of a Restructure Task
Example of a Restructure Task.

Metrics

Reliability and effective cost

Let yt,k{0,1}y_{t,k} \in \{0, 1\} denote whether the kk-th attempt on task tt is successful. With TT tasks and KK repeats, overall reliability is:

R=1TKt=1Tk=1Kyt,k.R = \frac{1}{T K} \sum_{t=1}^{T} \sum_{k=1}^{K} y_{t,k}.

We report Pass Rate (P@1) as the probability of success on the first attempt and Pass@10 (P@10) as the probability of at least one success across the 10 attempts.

We report expected attempts and effective cost using the same retry-cap model implemented in our internal analysis script; the equations here allow reproduction.

Cost model

The benchmark cost model combines (1) the model's per-candidate generation cost, (2) an estimated human review cost per candidate, and (3) the expected number of attempts required for a task. For model mm, the per-attempt cost is:

Cattempt=Cmodel(m)+Creview.C_{attempt} = C_{model}(m) + C_{review}.

We estimate review cost with a default hourly rate of $50 and 20 seconds of inspection time per image:

Creview=(50/3600)200.278.C_{review} = (50 / 3600) \cdot 20 \approx 0.278.

For each task tt, we estimate a pass rate ptp_t from the full K=10K=10 candidates and compute the expected number of attempts with a maximum retry budget A=4A=4 (reflecting a practical user retry limit). The expected attempts and success probability under this cap are:

Et={1(1pt)Aptif pt>0,Aif pt=0,St=1(1pt)A.E_t = \begin{cases} \frac{1 - (1 - p_t)^A}{p_t} & \text{if } p_t > 0, \\ A & \text{if } p_t = 0, \end{cases} \quad S_t = 1 - (1 - p_t)^A.

We then aggregate across tasks to compute pass@4 and effective cost per success using the retry-cap model:

P@4=1TtSt,E=1TtEt.\text{P@4} = \frac{1}{T} \sum_t S_t, \quad E = \frac{1}{T} \sum_t E_t. Ceff=ECattemptP@4.C_{eff} = \frac{E \cdot C_{attempt}}{\text{P@4}}.

Hype Gap

We report the Hype Gap (Best-of-10 Uplift) as the difference, in percentage points, between Pass@10 and Pass@1. Let P@1\text{P@1} be the success rate on the first attempt and P@10\text{P@10} be the probability of achieving at least one success within 10 attempts. Then:

HypeGap=P@10P@1.\text{HypeGap} = \text{P@10} - \text{P@1}.

Lower Hype Gap indicates consistent performance between a model's typical outputs and its best-of-10 results, while a larger Hype Gap implies that strong edits are possible but less reliable in practical use.


Read the paper and code

If you benchmark with Hype-Edit-1, we would love to hear your feedback and results.