Haozhe Zhao 1,3⋆, Xiaojian Ma 2⋆, Liang Chen 1,5, Shuzheng Si 4, Rujie Wu 5, Kaikai An 1,3,
Peiyu Yu 6, Minjia Zhang 7, Qing Li , Baobao Chang
1 National Key Laboratory for Multimedia Information Processing, Peking University
2 State Key Laboratory of General Artificial Intelligence, BIGAI
3 School of Software and Microelectronics, Peking University
4 Department of Computer Science and Technology, Tsinghua University
5 School of Computer Science, Peking University, China, 6 UCLA, 7 UIUC
Equal contribution ✉ Corresponding author
mimazhe55360@gmail.com,{maxiaojian,liqing}@bigai.ai,chbb@pku.edu.cn
ultra-editing.github.io

Abstract

This paper presents UltraEdit, a large-scale (~4M editing samples), automatically generated dataset for instruction-based image editing. Our key idea is to address the drawbacks in existing image editing datasets like InstructPix2Pix 1 and MagicBrush 2, and provide a systematic approach to producing massive and high-quality image editing samples. UltraEdit offers several distinct advantages: 1) It features a broader range of editing instructions by leveraging the creativity of large language models (LLMs) alongside in-context editing examples from human raters; 2) Its data sources are based on real images, including photographs and artworks, which provide greater diversity and reduced bias compared to datasets solely generated by text-to-image models; 3) It also supports region-based editing, enhanced by high-quality, automatically produced region annotations. Our experiments show that canonical diffusion-based editing baselines trained on UltraEdit set new records on MagicBrush and Emu-Edit benchmarks. Our analysis further confirms the crucial role of real image anchors and region-based editing data. The dataset, code, and models are available in github.com/pkunlp-icler/UltraEdit.

Refer to caption

Figure 1: Examples of UltraEdit. Free-form ( left ) and region-based ( right ) image editing.

1 Introduction

We present a new dataset, UltraEdit, for instruction-based image editing at scale (see Figure 1 for examples, more in Appendix B.2). Compared to several prior works on the same front 1 3 2, we venture into tackling some of their drawbacks, which prevent the model from robustly interpreting and executing instructions. We summarize our observations on their downsides below:

To address these limitations, we propose a systematic approach to curate massive and high-quality image editing data automatically. An overview of our method can be found in Figure 2. Specifically, we begin by using LLMs along with in-context human-written instruction examples to ensure diverse editing instructions. Next, we use prompt-to-prompt (P2P) 4 control with off-the-shelf T2I diffusion models to produce source and target (edited) images from captions and editing instructions. However, to avoid the biases within the T2I models, we collect high-quality image-caption pairs from diverse real image datasets like COCO 5 and use these images as anchors to guide the T2I models when producing source images given the corresponding caption. Additionally, we employ an automatic region generation approach to produce editing regions from the instruction and utilize such region annotations in a modified inpainting diffusion pipeline to produce region-based editing samples. As a result, the final dataset, UltraEdit, comprises ~4M editing samples with ~750K unique instructions across 9+ editing types. To the best of our knowledge, UltraEdit is the largest instruction-based image editing dataset to be released to the public. With UltraEdit, canonical diffusion-based editing baselines enjoy new records on challenging MagicBrush and Emu-Edit benchmarks, respectively. Our analysis further confirms the crucial role of real image anchors and region-based editing data. To sum up, our contributions can be summarized as follows:

  • We propose a novel automatic pipeline to generate image editing data, mitigating the issues of existing datasets: instruction diversity, image biases, and missing region-based editing data.
  • With our pipeline, we curate UltraEdit, a large-scale and high-quality image editing dataset with diverse instructions, real images as anchors, and additional editing region annotations.
  • We conduct extensive studies on how canonical diffusion-based image editing model can benefit from UltraEdit and provide insights and analysis on the key design principles.

2 The UltraEdit Dataset

Refer to caption

Figure 2: The construction of UltraEdit. (Upper) We use LLM with in-context examples to produce editing instructions and target captions given the collected image captions; (Middle) For free-form editing, we use the collected images as anchors, and invoke regular diffusion followed by prompt-to-prompt (P2P) 24 control to produce source and target images; (Bottom) For region-based editing, we first produce an editing region based on the instruction, then invoke a modified inpainting diffusion pipeline 44 5 to produce the images.

2.1 Dataset Formation

In UltraEdit, we consider two editing settings: 1) free-form editing, where the editing could happen at any area of the input image; 2) region-based editing, where the editing is expected to be about a certain region of the image, allowing more fine-grained editing. In all, our dataset can be formulated as a set of , where and denote source and target (edited) image, respectively, is the editing instruction, denotes the additional editing region (mask). The model is expected to map editing input , and mask to . We also need the captions of source and target image , for image generation and evaluation purposes. Below, we detail how to collect UltraEdit. An overview of our pipeline can be found in Figure 2.

Stage I: Instruction Generation. As we mentioned in Section 1, our key idea to address the issue of limited instruction diversity in existing datasets is to obtain high-quality editing instructions by combining LLM creativity and human raters. Specifically, we begin by manually creating hundreds of editing instructions with our human raters. The raters are first given images and captions from the COCO dataset 5 as context and asked to write appropriate editing instructions for these scenarios. We then invoke LLM to expand these human-written instructions into more diverse examples for editing in terms of semantics and editing tasks. We end up with ~10K instruction examples after the expansion. Details on the prompt and examples can be found in Appendix A.2.

Stage II: Free-form Editing Samples. We follow the main idea of prior work and invoke off-the-shelf T2I models to generate source and target images , in editing samples. However, instead of synthesizing all images using T2I models as in prior works 1 3, UltraEdit adopts real images as anchors to mitigate the biases within these T2I models. Specifically, we first collect ~1.6M high-quality and diverse image-caption paired data from real image datasets like COCO 5, NoCaps 6, etc. (full details on the mix can be found in Appendix A.1). For each image-caption pair , we sample in-context examples from the instruction pool obtained in Stage I and prompt LLM to produce the editing instruction and the resulting target caption after the editing. We then invoke a regular Img2Img diffusion pipeline with the noise-perturbed latent embedding of as (similar to SDEdit 7) and the source caption as a condition; therefore, the produced source image will resemble the anchor , i.e., we use diverse real images to guide the generation of T2I models, mitigating their biases. After producing , we invoke prompt-to-prompt (P2P) control 4 with the target caption to produce the target image on the same from real image anchor . Due to space limits, we refer the readers to the P2P paper 4 for details. We utilize SDXL-Turbo 8 as our diffusion backbone, which allows for high-quality generation with just 2-4 diffusion steps, maintaining a generation quality comparable to SDXL 9.

[Uncaptioned image]

Table 1: Comparison of different image editing datasets. Both EditBench and MagicBrush are manually annotated but are limited in size. InstructPix2Pix and HQ-Edit are large datasets automatically generated using T2I models like Stable Diffusion 51 and DALL-E 50, though they present notable biases from the generative models leading to failure cases. UltraEdit offers large-scale samples with rich editing tasks and fewer biases.

Stage III: Region-based Editing Samples. Upon the free-form editing samples, we create additional region-based editing data using an automatic editing region extraction method. Given an image-instruction pair , we first detect all objects within using recognize-anything 10 and prompt LLM with the object list, soruce caption , target caption and editing instruction to identify the object to be edited. Then we employ GroundingDINO 11 and SAM 12 to obtain bounding boxes and fine-grained mask of this target object. For edits involving transformation over the entire image (e.g., “turn this into an oil paint”), we mark the whole image as the editing region. We save an expanded version of the original mask (which becomes a contour) as the editing region . Finally, the bounding box and fine-grained mask will be fused into a soft mask (see Figure 2) to help smooth the transition between inpainting area and the rest of the image. While producing the source image is the same as in Stage II, we adopt a modified inpainting pipeline to produce the target image to take the editing region into consideration:

where is in the size of latent space, denotes the diffusion model. In a nutshell, we alternate between regular diffusion and inpainting only within the mask region to guide the generation within the given region while avoiding edge artifacts, as illustrated in Figure 16. This pipeline is compatible with P2P control and the SDXL-Turbo backbone and it takes 3-7 diffusion steps to produce a target image. Further details can be found in Appendix A.3.

Misc. To ensure high-quality image generation, for each editing sample, we run the diffusion pipeline 100 times and filter out the deficient generations using a mixture of automatic metrics following the prior practices 1 13 (detailed in Section 2.3). Thanks to the efficient SDXL-Turbo diffusion backbone and our implementation, our pipeline is ~100 times faster than prior work like 1.

2.2 Characteristics and Statistics

Our dataset contains a total of 4,108,262 instruction-based image editing data (757,879 unique edits), where free-form image editing (without region annotation ) consists of 4,000,083 instances and region-based editing includes 108,179 samples. To the best of our knowledge, this is the largest dataset to be released to the public (a comparison to other datasets can be found in Table 1). As illustrated in Figure 3, UltraEdit encompasses over 9 distinct editing instruction types, detailed in Table 3. The distribution of image editing instances across these types can be found in Appendix B.

Refer to caption

Figure 3: Distribution of edit types and keywords in the instructions of UltraEdit. The inner ring illustrates the various types of edit instructions, while the outer ring presents the frequency of instruction keywords. This visualization highlights the rich diversity found within our instructions.

2.3 Quality Assessment

To ensure the quality of our dataset, we employ several automatic metrics to filter out substandard images during generation following prior practies 13 1. First, we utilize DINOv2 similarity, CLIP image similarity, and SSIM between source and target (edited) images and to guarantee that semantic similarity and pixel-level coherence can be maintained, which is crucial for image editing. Second, to ensure the generated images accurately reflect the editing instructions, we assess the alignment between the images ,, and their corresponding captions , using CLIP similarity. Finally, to verify the dataset’s instruction-following capability, we employ CLIP Directional Similarity 2, which measures the alignment between changes in images and changes in corresponding captions. We provide these scores on UltraEdit in Table 3. Our dataset maintains a high standard of image quality and instruction alignment across both free-form and region-based editing data. Notably, region-based data are significantly better in terms of SSIM, confirming the benefits brought by region-based guidance (via our modified inpainting pipeline). Further quality assessments can be found in Appendix B and D.

3 Experiments

Our experiments evaluate the quality of UltraEdit in following user instructions faithfully while preserving the visual fidelity of the source image. First, we evaluate the performance of canonical diffusion-based editing models trained on our dataset across different instruction-based image editing benchmarks. Second, we dive into insights and analysis on the design principles of our dataset.

3.1 Setup

Settings. We follow the settings of Emu Edit 3 to train a diffusion model using various scales and types of data from our dataset, then evaluate the trained model across multiple benchmarks to assess the effectiveness of our dataset in advancing its image editing capabilities. For a fair comparison, we adopt the editing diffusion model introduced in InstructPix2Pix 1, which uses Stable Diffusion v1.5 14 as the backbone diffusion model. We also employ training data volumes identical to that used in 1. To support region-based editing, we augment the model’s U-Net to take in extra channels for region (mask) input. More details can be found in Appendix A.4.

Table 4: Results on the MagicBrush test set. We include both single-turn and multi-turn settings. We evaluate the models trained on UltraEdit without region-based editing data and full data (“Ours”). Models trained with full data are either evaluated with or without editing region as input.

SettingsMethodsL1 \downarrow L2 \downarrow CLIP-I \uparrow DINO \uparrow
Single-turnGlobal Description-guided
SD-SDEdit0.10140.02780.85260.7726
Null Text Inversion0.07490.01970.88270.8206
GLIDE3.4973115.83470.94870.9206
Blended Diffusion3.5631119.28130.92910.8644
Instruction-guided
HIVE0.10920.03800.85190.7500
InstructPix2Pix (IP2P)0.11410.03710.85120.7437
IP2P w/ MagicBrush0.06250.02030.93320.8987
Ours, trained w/o region data0.06890.02010.89860.8477
Ours, eval w/o region0.06140.01810.91970.8804
Ours, eval w/ region0.05750.01720.93070.8982
Multi-turnGlobal Description-guided
SD-SDEdit0.16160.06020.79330.6212
Null Text Inversion0.10570.03350.84680.7529
GLIDE11.74871079.59970.90940.8494
Blended Diffusion14.54391510.22710.87820.7690
Instruction-guided
HIVE0.15210.05570.80040.6463
InstructPix2Pix (IP2P)0.13450.04600.83040.7018
IP2P w/ MagicBrush0.09640.03530.89240.8273
Ours, trained w/o region data0.08830.02760.86850.7922
Ours, eval w/o region0.07800.02460.89540.8322
Ours, eval w/ region0.07450.02360.90450.8505

Baselines. We set the following models as baselines, categorized into instruction-guided image editing methods (the same setting as ours) and global description-guided methods, the latter requires descriptions of the target image to perform editing. The instruction-guided image editing methods include: InstructPix2Pix 1, HIVE 15, MagicBrush 2, and Emu Edit 3. The global description-guided image editing methods include: Null Text Inversion 16, SD-SDEdit 7, GLIDE 17, and Blended Diffusion 18. Notably, GLIDE and Blended Diffusion require a region mask for editing.

Benchmark and Metrics. We evaluate the model trained on our dataset across two popular benchmarks: MagicBrush 2 and Emu Edit Test 3. MaigicBrush benchmark evaluates the model by comparing the edited images with ground truth images and corresponding captions across different metrics. Following the MagicBrush 2, we chose the L1 distance, L2 distance, CLIP image similarity, and DINO similarity as metrics. Emu Edit Test benchmark compares the edited images with the source image and target captions for evaluation. Consistent with the Emu Edit 3, we use L1 distance, CLIP image similarity, DINO similarity, CLIP text-image similarity, and CLIP text-image direction similarity as metrics. These metrics generally measure how the edited image both preserves the original’s style and content and reflects modifications according to instructions. Details of the benchmark and metrics can be found in Appendix C.

3.2 Main Result I: General Image Editing on MagicBrush

We present the results on MagicBrush benchmark in Table 4. Here are our main observations: 1) Compared to canonical image editing baselines like HIVE, SD-SDEdit, and IP2P (zero-shot), merely training an editing diffusion model on the free-form editing data of UltraEdit (downsampled to ~450K to match the training set size of IP2P, denoted as Ours, trained w/o region data) already attains significant improvement over the baseline, confirming the advantages brought by our dataset to general image editing; 2) When also considering the relatively small-scale region-based editing data (~100K region-based + ~350K free-form, denoted as Ours) and evaluate on the same setting without editing region input, the general editing performance can be boosted considerably, verifying the effectiveness of region-based editing data to image editing in general; 3) Finally, when being evaluated using region input, the model trained on both free-form and region-based editing data (still downsampled to ~450K in total) sets the new record on MagicBrush especially on the challenging multi-turn setting, demonstrate that our region-based editing data can indeed help with the emergence of region-based editing capability, while existing approaches that also utilize masks (GLIDE, Blended Diffusion) are poor at effectively guiding their editing with region input.

Note: As also reported in 3, we found the MagicBrush benchmark itself introduces biases towards its training set, leading to unfairly high results of models trained on its training set, i.e. IP2P w/ MagicBrush. It effectively overfits on MagicBrush and loses general editing ability on other datasets. Poor results of the same model on Emu Edit (MagicBrush in Table 5) verify this.

Table 5: Results on Emu Edit Test. We present the benchmark results for various methods and models trained on different scales of data.

MethodCLIPdir CLIPout L1 CLIPimg DINO
InstructPix2Pix (450K)0.07840.27420.12130.85180.7656
MagicBrush (450+20K)0.06580.27630.06520.91790.8924
Emu Edit(10M)0.10660.28430.08950.86220.8358
Ours (450k, w/o region data)0.08230.27780.06260.86170.8190
Ours (1M w/o region data)0.08620.28040.05150.89150.8656
Ours (1.5M, w/o region data)0.09520.28080.06000.86590.8243
Ours (2M, w/o region data)0.09600.28110.06080.86890.8269
Ours (2.5M, w/o region data)0.09970.28220.08540.84070.7814
Ours (3M, w/o region data)0.10760.28320.07130.84460.7937

3.3 Main Result II: Scaling Effect of UltraEdit on Emu Edit Test

Refer to caption

Figure 4: Qualitative comparison of images generated by our data generation pipeline with other zero-shot image editing methods.

In Table 5, we present results on how progressively scaling the size of UltraEdit could affect the performance of general image editing on the challenging Emu Edit Test. We made the following observations: 1) In general, models trained on UltraEdit attain better results than IP2P, and the advantages expand significantly when the scale of the dataset increases; 2) Scaling effect can be more significant on metrics indicating editing, i.e., CLIPdir (measuring the if the editing is consistent with the changes between the captions of source and target images) and CLIPout (measure if the edited image is consistent with the caption of the target image), and we ultimately set a new record on Emu Edit than baseline trained on a proprietary 10M dataset; 3) When it comes to content preserving metrics, i.e., L1, CLIPimg, and DINO, our hypothesis to the trend is: as the data scale increases, the model is gradually learning to make hard edits, therefore it only starts to edit more after a certain scale of data is reached; 4) Compared to Emu Edit baseline, our model can produce both accurate edits (hight CLIPdir and CLIPout) while preserving the content of the original image (lower L1), indicating that our model is mostly editing the image while considering the context, not just creating new contents. Our qualitative results (Figure 5) provide further evidence on this.

3.4 Qualitative Evaluation

Refer to caption

Figure 5: Qualitative evaluation. (Top) On four distinct tasks on Emu Edit, listed from top to bottom: background, color, global, style; (Bottom) Multi-turn editing on MagicBrush.

Data Generation.

In Figure 4, we present qualitative examples generated by our data generation pipeline alongside various image editing methods for comparison. We evaluate our pipeline against several baselines, including image inversion methods like Null-text Inversion 16 and PnP Inversion 19; inpainting methods such as BrushNet 20 and PowerPaint 21; and zero-shot image editing methods like InfEdit 22 and MasaCtrl 23. Our data generation pipeline, leveraging SDXL-turbo, requires only 1-4 steps to produce a sample, which is significantly faster than other methods by magnitudes. Importantly, this efficiency does not compromise the quality. Our pipeline demonstrates competitive performance against other baselines, highlighting the quality of UltraEdit.

Image Editing.

In Figure 5, we provide some qualitative examples of different editing tasks on Emu Edit Test and multi-turn editing on MagicBrush generated by the Stable Diffusion v1.5 trained on UltraEdit. More examples can be found in Appendix D. Our main observations are: 1) most baselines, especially Emu Edit, tend to overedit the image, i.e. creating content while ignoring the context of the source image. For example, Emu Edit makes crude modifications by setting the entire person’s body as blank. We attribute these shortcomings to the biases introduced during the construction of Emu Edit’s training data; 2) Many baselines also fail to generalize to novel instructions, e.g., blurring, adding special style, etc.; 3) For multi-turn editing, even the MagicBrush baselines cannot complete the long-term editing coherently, while our model, even without editing region input, can strictly follow the instruction, e.g. one person, one bench, etc, and reaches the best results with region information, which confirms the effectiveness of region-based data in UltraEdit.

Table 6: Ablation study on real image anchors. The first three rows present results for models trained on different scales from the UltraEdit. The last three rows show results for models trained on data from the same generation pipeline but without using the real image as an anchor.

Data TypeData VolumeCLIPdir \uparrow CLIPimg \uparrow CLIPout \uparrow L1 \downarrow DINO \uparrow
UltraEditing450k0.08230.86170.27780.06260.8190
1M0.09250.86960.28070.05990.8307
1.5M0.09520.86590.28080.06000.8243
w/o image anchor450k0.07280.87160.27960.08480.8154
1M0.06380.88370.27700.06740.8353
1.5M0.07200.86430.27810.07140.8105

3.5 Insights and Analysis

In this section, we investigate how two of our designs when curating UltraEdit: real image anchors and region-based editing could affect the performances. All results are conducted on Emu Edit Test.

Real Image Anchors.

We conduct an ablation study to verify the effectiveness of incorporating real images as anchors during data generation. We train the editing model using two distinct datasets: free-form image editing data from UltraEdit and the editing data generated using the same pipeline of UltraEdit without using real images as anchors, i.e. identical to the data creation pipeline of InstructPix2Pix 1. The models were trained on data volumes of 450K, 1M, and 1.5M for both datasets. The results can be found in Table 6. Our key observations: 1) Dataset generated with real image anchors generally leads to better models across all three scales; 2) The scaling effect only presents when real image anchors are adopted. We hypothesize that datasets without them could suffer from more severe image biases and therefore hinder the effect of further scaling up. In Appendix D.3, we demonstrate more qualitative results comparing the image editing generation method with and without using real images as anchors.

Refer to caption

Figure 6: Ablations on region-based editing. We report CLIPout as the main metric.

Free-from vs. Region-based Editing.

We then explore the impact of incorporating region-based editing data during model training. We experiment with varying amounts of free-form editing data, ranging from 200K to 400K instances, and region-based editing data, ranging from 30K to 90K instances. We demonstrate the performances on Emu Edit Test of models trained on the possible volume combinations of these two types of data in Figure 6. It can be seen that: 1) Region-based editing data, despite its relatively smaller scale, can help with free-form editing tasks; 2) While scaling free-form editing data has a significant impact on the performance of region-based editing, we need to ensure a considerable volume of region-based data to ensure peak results. In Appendix D.4, we demonstrate more qualitative results with region input and the model exhibits significantly more precise operations for background and localized edits, validating the effectiveness of the incorporation of the region-based editing.

Image Editing via Generation.

Editing real photos according to specific instructions has long been a notable task in image processing 24 25 26 27 28 29. Powerful Large-scale diffusion models have significantly facilitated text-based image editing 30 31 32 33 34. SDEdit 7 applies noise to the guidance image at an intermediate diffusion step and then denoises using the target description. Prompt-to-Prompt 4 injecting the input caption attention maps to the target ones. Null-Text Inversion 16 inverts the source image to the null-text embedding for editing, eliminating the need for original captions. Plug-and-Play 35 incorporates spatial features besides attention maps for better global image editing. Imagic 30 supports complex textual instructions via text embedding optimization and model fine-tuning. GLIDE 17 and Imagen Editor 36 fine-tuning the model to take channel-wise concatenation of the input image and mask. Blended Diffusion 18 blends the input image in the unmasked regions in the diffusion step. Meanwhile, instruction-based image editing has been introduced as a user-friendly method for image editing. Instructpix2pix 1 and HIVE 15 both are trained on generated editing data to handle user-written instructions during inference. MagicBrush 2 creates a manually-annotated dataset for fine-tuning Instructpix2pix. Emu Edit 3, trained on 10 million proprietary multi-task data, demonstrates state-of-the-art performance.

Image Editing Dataset.

Table 1 compares various image editing datasets, revealing that high-quality data are scarce and challenging to obtain. The largest dataset only contains around 300,000 samples. EditBench 37 is manually curated with only 240 examples. MagicBrush 2 manually annotated by hearing online labors using DALL-E 2 38 for crafting dataset, also limited in dataset size. For automatically annotated datasets, InstructPix2Pix utilizes prompt-to-prompt 4 to generate data pairs based on the captions form LAION-Aesthetics 39 and edited captions generated by GPT-3 40, but biases in the generative model and insufficient information in web-crawled captions make the generated image samples fail to represent the editing instructions, as shown in the Table 1. The the limited scope of the image instructions types of InstructPix2Pix further limits the utility of the dataset. HQ-Edit 41 uses advanced models like GPT-4 42 43 and DALL-E 3 44 for generating image editing pairs, but may fail to preserve fine-grained details and realism in the target image.

5 Conclusion

We’ve presented UltraEdit, a large-scale, high-quality dataset for instruction-based image editing. We mitigate the issues in existing editing datasets with a systematic approach for automatic data generation: combining LLM creativity and in-context examples from human raters for more diverse editing instructions; the use of real images as anchors for more balanced generations; support of region-based editing via automatic region extraction. Experiments on challenging MagicBrush and EmuEdit benchmarks confirm the effectiveness of training on our dataset. Possible future work includes further expanding the region-based editing data and bootstrapped training for image editing.

Acknowledgement

We extend our heartfelt gratitude to the anonymous reviewers whose dedication and insightful feedback have significantly enhanced the caliber of this paper. Their constructive critiques and valuable suggestions were instrumental in refining our work. Additionally, we are deeply appreciative of the Program Chairs and Area Chairs for their meticulous handling of our submission and for their comprehensive and invaluable feedback. Their guidance has been pivotal in elevating the quality of our research.

This work is supported by the National Science Foundation of China under Grant No.61936012 and 61876004.

References

Table 7: Datasets used to form the high-quality image caption dataset.

DatasetSamplesLicenseAnnotator
MS COCO 45164,000CC BY 4.0Human
Flickr 4631,783CustomHuman
NoCaps 4745,000CC BY 2.0Human
VizWiz Caption 4823,431CC BY 4.0Human
TextCaps 4928,408CC BY 4.0Human
Localized Narratives 50849,000CC BY 4.0Human
ShareGPT4V 511,200,000CC BY-NC 4.0GPT-4V
LAION-LVIS 52220,000Apache-2.0GPT-4V

Table 8: Examples of instructions and their corresponding captions after Expansion.

Original CaptionEdit InstructionEdited Caption
Two Asian dolls with big noses, fancy purple dresses, and golden hats.Replace the dolls with miniature elephants in colorful traditional Indian clothTwo miniature elephants wearing colorful traditional Indian cloth.
A man is watching TV in bed with only his foot showing.Remove the bed and place the man in a deserted beach scene.A man is watching TV on a deserted island beach with only his foot showing.
a person throwing a red frisbee, which is currently in mid-air, slightly to the right and above the person’s hand.Change the color of the frisbee to bluea person throwing a blue frisbee, which is currently in mid-air, slightly to the right and above the person’s hand.
A slipper near the edge of a concrete floor near small rocks.Transform the slipper into a glass oneA glass slipper near the edge of a concrete floor near small rocks.
A woman wearing a shirt for the Religious Coalition for Reproductive Choice.Change the background to a bustling cityscape at nightA woman wearing a shirt for the Religious Coalition for Reproductive Choice in front of a bustling cityscape at night.
A pot and some trays are in a kitchen.Add a warm, inviting atmosphere to the imageThe warm glow highlights a pot and some trays in a cozy kitchen.
a person that is jumping his skateboard doing a trick.Turn the skateboard into a flying carpeta person that is jumping his flying carpet doing a trick.
A stuffed bear is hanging on a fence.Make it a snowy winter landscapeA stuffed bear is hanging on a snowy winter landscape fence.
A police dog wearing his bullet proof vest.replace the background with a city skylineA police dog wearing his bullet proof vest in a city skyline.
A baseball game is going with children playing a runner is about to hit the base.Turn the baseball field into a magical forestChildren playing a runner is about to hit the base in a magical forest.
A small horse carries a women in a sled.Turn the horse and sled into a spaceship traveling through outer spaceA futuristic spaceship travels through outer space.
a person wearing a life jacket participating in water sports like water skiing.Add a family of dolphins swimming around the person in the watera person wearing a life jacket participating in water sports like water skiing, with a family of dolphins swimming around him in the water.

Appendix A Implementation Details

A.1 Collection of High-Quality Image Caption Data

To ensure the diversity and quality of the image editing pairs in UltraEdit, our data generation pipeline relies on high-quality oracle data to mitigate biases within the generation models. Consequently, we focus on building the image editing dataset based on real images and their captions, enhancing the dataset’s reliability in real-world scenarios and providing more comprehensive guidance for data generation than text-only data.

Like previous works 53 51 2 54, we gathered source data from various public data sources with image caption data, as illustrated in Table 7, which includes diverse images with either manually annotated captions or detailed captions generated by advanced image captioning models 51 52. After filtering out images with excessively long or short captions, our collection amounted to 1.6 million high-quality image-caption pairs, which will be used for edit instruction generation and subsequent image generation.

A.2 Instruction and Caption Generation

To obtain high-quality editing instructions, we introduce a pipeline that combines human raters and language models for generating edit instructions and corresponding captions for subsequent image generation. Large language models have demonstrated remarkable abilities in various areas, such as agents 55 56 57 58 and tool uage 59. In our practice, we utilize the LLM to generate suitable image editing instructions. Firstly, we use language models to expand manually crafted edit examples to a set of 100,000 examples, as shown in Table 8. These examples serve as in-context learning examples to help the language models grasp the understanding of editing styles and requirements, enabling them to generate suitable editing instructions and corresponding edited caption. The query prompt is illustrated in Table 9.

We sample 50 editing instructions and 10 edit examples as in-context learning examples for querying the language model. The language model then generates appropriate editing instructions and corresponding edited captions for the given image captions from the collection. Leveraging the in-context-learning capabilities and generalization ability of the language model, we ultimately generate 4.16 million text-only data comprising creative yet sensible edit instructions and corresponding captions. Each case consists of a high-quality image caption for a real image, an editing instruction, and an edited caption corresponding to a target image.

Table 9: Query prompts for LLMs to write edit instructions and corresponding captions.

Prompt for Writing Edit Instruction
ElementContent
IntroI will present a series of image editing instruction examples essential for mastering and understanding a variety of editing styles and requirements. Here are some sample instructions:
Instruction Examples{instruction_str}
Task DescriptionI will provide one image caption corresponding to a specific image. You are required to apply the learned editing techniques to form suitable, detailed, and accurate editing instructions for the image defined by the caption. Note that your editing instructions should be distinct from the examples provided. Then, produce a description corresponding to the revised image after applying the editing instruction. Only necessary amendments should be made for the new image caption.
Output FormatThe output format should be "original image caption; edit instruction; new image caption". Maintain the given format for the result. Please ensure to deliver solely the result, without incorporating any additional titles.
Image CaptionThe image caption is: {caption}
Produce InstancesProduce three suitable instances based on the caption and return the list.
ExamplesHere are some output examples for your reference: {example_str}
ResponseResponse:

Table 10: Query prompts for LLMs to capture objects that need editing.

Prompt for Capturing Editing Object
ElementContent
IntroThe following prompt provides an instruction for image editing, an original image caption, a revised image caption that reflects the given edit instruction, and a set of objects detected by an object detection algorithm.
Edit InstructionEdit Instruction: "{edit_instruction}"
Original CaptionOriginal Image Caption: "{input_text}"
Revised CaptionRevised Image Caption: "{output_text}"
Object ListSet of Objects Identified by the Recognition Model: {object_list}
Task DescriptionYour task is to identify the objects most likely to be modified based on the information provided above. Consider this from a comprehensive perspective; note that some objects might not be explicitly mentioned in the instructions or the Identified Objects list, but their appearance could still be affected. Please use precise words or phrases in your response.
NoteNote:
1. If you can’t identify any specific edited object (e.g., a style transfer involving the entire image instead of a single object; add/move an object, which does not fit object-oriented editing instructions), please respond with "NONE".
2. Your response should exclusively identify the objects requiring edits, excluding any extra context or details.
3. Please list the objects to be edited, separating each one with a comma. The number of objects identified in the answer should not exceed 2.
ResponseResponse:

A.3 Region-based Data Generation

For generating region-based image editing data, we first employ the recognize-anything 10 to identify objects in the source images. We then query the language model with the obtained object lists, edit instructions, and corresponding image captions to determine the target objects of the editing instruction using the query prompt in Table 10. If the editing instruction is object-oriented, the language model identifies the objects involved in the editing; otherwise, the entire image is considered as the editing target.

To generate region-based edited images, we use the Grounding DINO 11 to obtain bounding boxes of editing areas in real images, serving as coarse-grained masks. Subsequently, we perform SAM 60 on these bounding boxes to derive fine-grained object masks, expanding them to create contour masks that define the editing region. During image generation, we have observed irregular color boundaries between the editing region and the rest of the image. To ensure smooth transitions and high image quality, we fuse the fine-grained and bounding box masks to create a soft mask guiding the image generation. Specifically, for the editing region latent and bounding box latent , we fuse the two masks, making the region between them a soft mask region . The generation pipeline can be formulated as follows:

Additionally, we define as:

where is the editing region latent, is the bounding box latent, and is the hyperparameter that determines the inpainting rate. During the generation, During the generation, we set to range from 0.2 to 0.8.

A.4 An Improved Baseline for Free-form and Region-based Image Editing

We fine-tune the Stable Diffusion 1.5 model 61 using the Diffusers library 62 with data from UltraEdit. We maintain the hyperparameters as set in 1. Specifically, we train the model on 8 80GB NVIDIA A100 GPUs with a total batch size of 256. Following prior works 1 2, we use an image resolution of 256 256 for training and 512 512 for generation.

To incorporate additional guidance from region masks, we concatenate the latent of the Region Mask with the noisy latent and the latent of the source image to form the input to the diffusion model. We add four additional channels to the UNet of the diffusion model to accommodate the latent of the region mask . The weights of the UNet are initialized with the pretrained diffusion model, while the extra eight channels (four for the source image latent and four for the mask latent ) in the convolutional layers of the diffusion UNet are randomly initialized. The model is then trained using a mixture of free-form and region-based image editing data from UltraEdit. For free-form image editing data, the model takes a blank mask as input to implicitly indicate that the editing should affect the entire image.

When training the model exclusively with Free-form Image Editing data, we strictly follow the settings of 1 without making any additional modifications.

Appendix B Statistics of UltraEdit

Table 11: Statistics of Free-form and Region-based Image Editing Data. The table shows the instance numbers, number of unique instructions, and their respective proportions for different instruction types.

Data TypeStatisticChangeTransformAddReplaceTurnOthersTotal
ColorGlobalLocalGlobalLocal
Free-formInst. No.111,563204,294500,108150,851597,165909,065683,529490,219353,2894,000,083
Proportion (%)2.795.1112.503.7714.9322.7317.0912.268.83/
Unique Inst.27,43624,02092,89126,587117,063114,647133,222102,28086,180724,326
Proportion (%)3.793.3212.823.6716.1615.8318.3914.1211.90/
Region-basedInst. No.2,9123,51515,7742,79621,80711,91825,74916,6287,080108,179
Proportion (%)2.693.2514.582.5820.1611.0223.8015.376.54/
Unique Inst.1,0567275,0327626,8353,2018,0645,2562,62033,553
Proportion (%)3.152.1715.002.2720.379.5424.0315.667.81/

In this section, we dive into the characteristics and statistics of UltraEdit. We present UltraEdit, a large-scale, diverse, and high-quality real-image-based image editing dataset designed to advance the capabilities of image editing models. UltraEdit comprises over 4,000,000 instruction-based free-form image editing instances and 100,000 region-based image editing instances, making it the largest open-source image editing dataset. Notably, it is also the first large-scale dataset focused on region-based image editing. Table 11 illustrates the statistics of the image editing data in UltraEdit. It shows the numbers and proportions of the different instruction types and data types of UltraEdit. Moreover, the Figure 7 shows the distribution of keywords in the instructions of UltraEdit for various instruction types.

Refer to caption

(a) Add

B.1 Comparison with other dataset

In this section, we compare our dataset with the InstructPix2Pix (IP2P) dataset in Free-form image editing. We report the results of automatic metrics to evaluate the data quality of each dataset.

As illustrated in Table 12, our dataset outperforms the InstructPix2Pix (IP2P) dataset across all tasks. Notably, the higher CLIPimg scores observed in some tasks for the IP2P dataset suggest that while the image pairs in the dataset may exhibit semantic similarity, it does not achieve the desired visual similarity, which is crucial for successful image editing. This highlights a fundamental shortcoming in the IP2P dataset and underscores the superior of our dataset.

Table 12: Comparison of dataset quality between UltraEdit and InstructPix2Pix (IP2P) using automatic metrics. We evaluates the data quality across different tasks types.

TaskDatasetCLIPinCLIPoutCLIPdirCLIPimgSSIMDINOV2
ChangeOurs0.28490.30240.29670.84410.63600.7403
IP2P0.26670.26610.23170.85570.56850.6194
TransformOurs0.28510.30050.29020.82890.62510.6875
IP2P0.26460.26800.19740.86670.58530.6972
TurnOurs0.28460.30180.29220.83210.62550.6949
IP2P0.26540.26980.20150.85750.55260.6419
AddOurs0.27860.31450.29570.86610.66450.7758
IP2P0.26290.27440.19900.88510.63180.7026
OthersOurs0.28430.30380.29810.83740.64200.7048
IP2P0.26570.27060.19290.86290.57340.6847
OverallOurs0.28340.30490.29500.84270.64010.7231
IP2P0.26500.26940.19820.86600.58260.6859

B.2 More Examples of UltraEdit

In this section, we showcase additional examples from UltraEdit to illustrate the versatility and robustness of our dataset in various image editing tasks. The free-form editing data is depicted in the left two columns, while the region-based image editing data examples are in the right column. The examples highlight both Free-form and Region-based editing capabilities. It can be noticed that, due to using real images as anchors, our data shows high diversity in real-world scenarios, including text, natural environments, human figures, abstract objects, and even blurred low-quality images.

In Figure 8 and Figure 9, editing examples not only contain text modification, and abstract object editing, but also multi-step editing within a single instruction and fine-grained editing. Moreover, because of high-quality captions derived from open-source image caption datasets for generating editing instructions, the generated instructions are highly related to the source image. The region-based image editing data demonstrates high image element preservation in the editing examples. For instance, in the examples in the right column, the target images only perform edits within the masked area and keep the rest unchanged, even for highly blurred texts and human facial expressions in the figure.

Refer to caption

Figure 8: More Examples of UltraEdit. Free-form ( left ) and region-based ( right ) image editing.

Refer to caption

Figure 9: More Examples of UltraEdit. Free-form ( left ) and region-based ( right ) image editing.

Appendix C Baseline and Metrics

C.1 Baselines.

We set the following models as baselines, categorized into instruction-based image editing methods and global description-guided image editing methods, the latter requiring global descriptions of the target image to perform zero-shot editing. The instruction-based image editing methods include: InstructPix2Pix 1, HIVE 15, MagicBrush 2, and Emu Edit 3. The global description-guided image editing methods include: Null Text Inversion 16, SD-SDEdit 7, GLIDE 17, and Blended Diffusion 18. Notably, GLIDE and Blended Diffusion require a mask for editing.

Instruction-Based Methods:

  • InstructPix2Pix uses automatically generated instruction-based image editing data to fine-tuning Stable Diffusion 14 and performance image editing based on the instructions during the inference, without any test-time tuning.
  • HIVE is trained with more data similarly to InstructPix2Pix, and is further fine-tuned with a reward model trained on human-ranked data.
  • MagicBrush: is a variant of InstructPix2Pix, which is fine-tuned on the human-annotated dataset, MagicBrush.
  • Emu Edit: is a closed-source model that supports multi-task image editing and achieves state-of-the-art performance. It is trained on a diverse set of tasks using 10 million of training data, including image editing and computer vision tasks.

Global Description-Guided Methods:

  • Null Text Inversion: inverts the source image with DDIM 63 trajectory and then performs editing during the denoising process with text-image cross-attention control 4.
  • SD-SDEdit: noises the guidance image to an intermediate diffusion step, and then denoises it using the target description.
  • GLIDE: is trained with 67M text-image pairs to fill in the masked region of an image conditioned on the local description with CLIP guidance.
  • Blended Diffusion: blends the input image in the unmasked regions with the context in the noisy source image during each denoising timestep to enhance region-context consistency.

C.2 Details on Benchmarks and Metrics

MagicBrush aims for evaluating the single and multi-turn image editing ability of the model. It provides annotator defined instructions and editing masks, as well as the ground truth images generated by DALLE-2 38 for evaluation, allowing for more effective metric assessment of the model’s editing performance. However, this dataset also suffers from inherent bias. During data collection, annotators were directed to use the DALLE-2 image editing platform to generate the edited images. Thus, this benchmark is biased towards generated images and editing instructions that the DALLE-2 editor can successfully follow, which may compromise both its diversity and complexity. Following the setting of MagicBrush 2, we utilize L1 and L2 to measure the pixel-level difference between the generated image and ground truth image. And also adopts the CLIP similarity and DINO similarity to measure the overall similarity with the ground truth. Finally, the CLIP-T is used to measure the text-image alignment between local descriptions and generated images CLIP embedding.

Emu Edit Test aims for reducing bias of the annotator defined dataset and reach higher diversity. It contains the devise relevant, creative, and challenging instructions and high quality captions that capture both important elements in the image for source and target images, without any ground truth images. Consequently, consistent with the Emu Edit 3, we utilize the L1 distance, CLIP image similarity and DINO similarity between the source images and edited images to measure the the model‘s ability of preserving elements from the source image. Also, we use the CLIP text-image similarity between edited image and output caption and the CLIP text-image direction similarity(CLIPdir) to measure the instruction following ability of the model. Specifically, the CLIPdir measures agreement between change in caption embedding and the change in image embedding.Since the Emu Edit 3 does not specify the versions of the CLIP and DINO models used for the metric, we adopted the settings utilized by MagicBrush to maintain alignment with other benchmarks. Specifically, the versions are ViT-B/32 for CLIP and dino_vits16 for DINO embeddings. We ensure consistency by rerunning all results of different methods on Emu Edit benchmark. Additionally, there are known issues with the quality of the benchmark, wherein some image-caption pairs appear incorrect. These issues include placeholder captions (e.g., ’a train station in city’) or instances where source and target captions are identical. To address these problems, we simply remove the incorrect cases prior to evaluation. Despite the Emu Edit Test eliminating bias and overfitting at the image level by not providing ground-truth images, the evaluation metrics still implicitly measure the model’s editing ability.

Appendix D Qualitative and Human Evaluations

D.1 Human Evaluation

We conducted human evaluations to assess the consistency, instruction alignment, and image quality of the edited images generated by our model trained on UltraEdit using the MagicBrush benchmark and Emu test benchmark. We first compared the performance of our model with the MagicBrush 2 and instructPix2Pix 1 models through a comprehensive human evaluation on MagicBrush benchmark. Additionally, we compared the performance of various models trained using our dataset with Emu Edit 3 on the Emu test benchmark. For the two evaluations, we randomly sampled 500 examples from the test sets of the MagicBrush benchmark and the Emu test benchmark, respectively.

For each sample, the evaluators compared the consistency, instruction alignment, and image quality of the edited images generated by the different models. As shwon in Figure 10, the evaluators were asked to determine which edited image was better by selecting between “First Image”, “Second Image”, or “Tie”. The results are evaluated with TrueSkill 64 rating system. The scores of these evaluations are presented in Table 13 and Table 14. Our model (finetuned with our own UltraEdit) can produce more preferable editing results than the baselines, even better than the MagicBrush baseline, which is reported to overfit on its test set.

Refer to caption

Figure 10: The interface of human evaluation on MagicBrush 71 and Emu test 60 benchmark to evaluate generated images by different models.

Ours w/ UltraEditMagicBrush 2InstructPix2Pix 1
TrueSkill score

Table 13: TrueSkill 64 scores of image editing models evaluated by human raters on MagicBrush 2 test set.

SD3 65 w/ UltraEditSDXL 9 w/ UltraEditSD1.5 14 w/ UltraEditEmu Edit 3
TrueSkill score

Table 14: TrueSkill 64 scores of image editing models evaluated by human raters on Emu Test 3 test set.

D.2 Qualitative Evaluation on Different Benchmarks

In Figure 11 and Figure 12, we present the qualitative examples of different editing tasks on single-turn and multi-turn editing on MagicBrush. In Figure 13, we present the qualitative examples on Emu Edit Test across various editing tasks.

Refer to caption

Figure 11: Qualitative evaluation of the model trained on UltraEdit across MagrichBrush benchmark in the Single-turn Setting.

Refer to caption

Figure 12: Qualitative evaluation of the model trained on UltraEdit across MagrichBrush benchmark in the Multi-turn Setting.

Refer to caption

Figure 13: Qualitative evaluation of the model trained with UltraEdit on the Emu Test.

D.3 Qualitative Evaluation on Real Image Anchors

In Figure 14, we present qualitative results comparing the image editing generation method with and without using real images as anchors. Using real images as anchors to guide the data generation significantly enhances the diversity of the generated images and ensures that the generation results are more stable and aligned with the editing instructions. The image anchors provide substantial information for generation that goes beyond what is conveyed by the image captions alone. Specifically, image anchor ensures visual consistency between the generated source and target images in the image editing pairs, as shown in the first three rows of Figure 14. It can also be observed that with real image anchors, the editing process is more controlled, resulting in fine-grained image edits in the generated samples (see the last three rows in Figure 14).

Refer to caption

Figure 14: Qualitative evaluation of using real images as anchors during image generation. We compare qualitative examples between the generation pipeline using real image anchors ( left ) and the generation pipeline without real image anchors ( right ). The real images are presented in the middle column.

D.4 Qualitative Evaluation on Free-form vs. Region-based Editing

In Figure 15, we present qualitative results comparing the model trained with an additional region-based editing task against the model trained solely with free-form image editing data. The comparison highlights that the inclusion of the region-based editing task during training enables the model to perform significantly more precise operations even in the absence of region input during evaluation, especially for background and localized edits.

Refer to caption

Figure 15: Qualitative evaluation comparing free-form and region-based editing task.

D.5 Details of the region-based image editing pipeline

Refer to caption

Figure 16: Qualitative evaluations of the region-based image editing pipeline. Generated images Without our method exhibit noticeable artifacts along the boundaries of the original and edited regions, emphasizing pronounced border effects.

In Stage III, we apply our proposed method for generating region-based images to ensure a seamless transition between inpainted areas and the rest of the image. Initially, we analyze inaccurate masks generated by the segmentation model, as shown in Figure 17. We find these inaccuracies generally fall into a few categories: incorrect identification resulting in overly large masks, masks that are too small for effective editing, fragmented masks from segmentation failures, and fine-grained segment masks that closely resemble the original object, complicating the editing process.

To address these issues, we filter out excessively large, small, or fragmented masks. Fine-grained masks are adjusted using a soft mask, either a bounding box or contour mask. Our data circulation indicates that our methods significantly reduce artifacts and abrupt boundaries between the mask region and the remaining image. Qualitative evaluations shown in Figure 16 demonstrate the effectiveness of our approach. Images generated without our method exhibit noticeable artifacts along the boundaries of the original and edited regions, highlighting the advantages of our method.

Refer to caption

Figure 17: The four main categories of inaccuracies generated masks.

Appendix E Statement on Limitations and Ethical Concerns

E.1 Limitations

While UltraEdit represents a significant advancement in instruction-based image editing, several limitations should be acknowledged. Firstly, although our dataset includes diverse editing instructions and real image anchors, the reliance on automatically generated data may introduce some biases and errors. The quality and relevance of the editing instructions, influenced by current large language models and human raters, may not capture all nuances of creative and artistic editing tasks. Furthermore, despite our efforts to provide high-quality region annotations, there may be occasional inaccuracies or inconsistencies in the automatically produced region-based editing data.

Moreover, while our experiments demonstrate the benefits of using real image anchors and region-based editing data, the improvements shown by our diffusion-based editing baselines are benchmark-specific. They may not generalize across all editing scenarios. Future work should focus on enhancing the precision of region annotations and validating the dataset’s applicability across a broader range of editing tasks.

Despite these limitations, UltraEdit offers a robust and diverse dataset that significantly contributes to the field of image editing, paving the way for future research and development.

E.2 Ethical concerns

While UltraEdit offers substantial advancements in the field of instruction-based image editing, several ethical concerns must be considered:

  • Bias and Fairness. The dataset, while diverse thanks to the efforts on real image anchors, etc., may still contain biases introduced by the automatic generation process and the inherent biases present in the large language models and human raters used. These biases could perpetuate stereotypes or unfair representations in the edited images.
  • Misinformation and Misuse. The powerful image editing capabilities enabled by UltraEdit could be misused to create misleading or deceptive content, contributing to the spread of misinformation. It is crucial to implement safeguards and promote responsible use of the technology to mitigate this risk.
  • Privacy. Real image anchors included in the dataset may contain identifiable information. Although efforts have been made to anonymize and protect personal data, there remains a risk of unintentional breaches of privacy.

To address these ethical concerns, we encourage users of UltraEdit to adhere to ethical guidelines, implement robust checks for bias and fairness, and prioritize transparency and accountability in their work. Additionally, we recommend ongoing dialogue within the research community to continuously refine and improve ethical standards in developing and applying image editing technologies.

Appendix F Datasheet for UltraEdit

We present a Datasheet 66 for documentation and responsible usage of our internet knowledge databases. The required author statement, hosting, licensing, metadata, and maintenance plan can be found in the datasheet.

F.1 Motivation

For what purpose was the dataset created?

We create this large-scale dataset to facilitate research towards image editing based on natural language instructions and regions (masks).

Who created the dataset (e.g., which team, research group) and on behalf of which entity (e.g., company, institution, organization)?

This dataset was created by Haozhe Zhao (Peking University), Xiaojian Ma (BIGAI), Liang Chen (Peking University), Shuzheng Si (Tsinghua University), Rujie Wu (Peking University), Kaikai An (Peking University), Peiyu Yu (UCLA), Minjia Zhang (UIUC), Qing Li (BIGAI), and Baobao Chang (Peking University).

F.2 Distribution

Will the dataset be distributed to third parties outside of the entity (e.g., company, institution, organization) on behalf of which the dataset was created?

Yes, the dataset is publicly available on the internet.

How will the dataset will be distributed (e.g., tarball on website, API, GitHub)?

All datasets can be downloaded from https://huggingface.co/. Please refer to this table of URL, DOI, and licensing. The Croissant metadata can be found on the dataset hosting platform (https://huggingface.co/).

DatasetDOILicense
UltraEdit-full10.57967/hf/2481Creative Commons Attribution 4.0 (CC BY 4.0)
UltraEdit-free-form-500k10.57967/hf/2535Creative Commons Attribution 4.0 (CC BY 4.0)
UltraEdit-region-based-100k10.57967/hf/2534Creative Commons Attribution 4.0 (CC BY 4.0)

Have any third parties imposed IP-based or other restrictions on the data associated with the instances?

No.

Do any export controls or other regulatory restrictions apply to the dataset or to individual instances?

No.

F.3 Maintenance

Who will be supporting/hosting/maintaining the dataset?

The authors will be supporting, hosting, and maintaining the dataset.

How can the owner/curator/manager of the dataset be contacted (e.g., email address)?

Please contact Haozhe Zhao (mimazhe55360@gmail.com), Xiaojian Ma (maxiaojian@bigai.ai) and Qing Li (liqing@bigai.ai).

Is there an erratum?

No. We will make announcements if there is any.

Will the dataset be updated (e.g., to correct labeling errors, add new instances, delete instances)?

Yes. New updates will be posted on https://ultra-editing.github.io.

If the dataset relates to people, are there applicable limits on the retention of the data associated with the instances (e.g., were the individuals in question told that their data would be retained for a fixed period of time and then deleted)?

The images in our dataset might contain human subjects, but they are all synthetic.

Will older versions of the dataset continue to be supported/hosted/maintained?

Yes, old versions will be permanently accessible on huggingface.co.

If others want to extend/augment/build on/contribute to the dataset, is there a mechanism for them to do so?

Yes, please refer to https://ultra-editing.github.io.

F.4 Composition

What do the instances that comprise the dataset represent?

Our data is generally stored in the Apache Parquet format, which is a table with multiple columns. We provide images (as source images and target/edited images), captions of source and target images, editing instructions, objects to be edited, metrics (CLIPimg, DINOv2, SSIM, CLIPin, CLIPout, and CLIPdir), and editing regions (optional), as separate columns.

How many instances are there in total (of each type, if appropriate)?

There are ~4M samples, among which ~100K are region-based editing data, while the rests are free-form editing data.

Does the dataset contain all possible instances or is it a sample (not necessarily random) of instances from a larger set?

We provide all instances in our Huggingface data repositories.

Is there a label or target associated with each instance?

No.

Is any information missing from individual instances?

No.

Are there any errors, sources of noise, or redundancies in the dataset?

Please refer to Appendix E.

The dataset is self-contained.

Does the dataset contain data that might be considered confidential?

No.

Does the dataset contain data that, if viewed directly, might be offensive, insulting, threatening, or might otherwise cause anxiety?

We have made our best efforts to detoxify the contents via an automated procedure. Please refer to Sec. E.

F.5 Collection Process

The collection procedure, preprocessing, and cleaning are explained in detail in Section 2 of the main paper.

Who was involved in the data collection process (e.g., students, crowdworkers, contractors) and how were they compensated (e.g., how much were crowdworkers paid)?

All data collection, curation, and filtering are done by UltraEdit coauthors.

Over what timeframe was the data collected?

The data was collected between Jan. 2024 and May 2024.

F.6 Uses

Has the dataset been used for any tasks already?

Yes, we have used UltraEdit for training our image edit models.

What (other) tasks could the dataset be used for?

Our dataset is primarily for facilitating research in building more capable image editing models that follow natural language instructions and (optionally) editing region input. Our data might also be used to benchmark existing and future image editing models.

Is there anything about the composition of the dataset or the way it was collected and preprocessed/cleaned/labeled that might impact future uses?

No.

Are there tasks for which the dataset should not be used?

We strongly oppose any research that intentionally generates harmful or toxic content using our data.

Checklist

The checklist follows the references. Please read the checklist guidelines carefully for information on how to answer these questions. For each question, change the default [TODO] to [Yes], [No], or [N/A]. You are strongly encouraged to include a justification to your answer, either by referencing the appropriate section of your paper or providing a brief inline description. For example:

  • Did you include the license to the code and datasets? [Yes] See supplementary.
  • Did you include the license to the code and datasets? [Yes] See supplementary.

Please do not modify the questions and only use the provided macros for your answers. Note that the Checklist section does not count towards the page limit. In your paper, please delete this instructions block and only keep the Checklist section heading above along with the questions/answers below.

  1. For all authors…
    1. Do the main claims made in the abstract and introduction accurately reflect the paper’s contributions and scope? [Yes]
    2. Did you describe the limitations of your work? [Yes] See supplementary.
    3. Did you discuss any potential negative societal impacts of your work? [Yes] See supplementary.
    4. Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes]
  2. If you are including theoretical results…
    1. Did you state the full set of assumptions of all theoretical results? [N/A]
    2. Did you include complete proofs of all theoretical results? [N/A]
  3. If you ran experiments (e.g. for benchmarks)…
    1. Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes]
    2. Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] See supplementary.
    3. Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [N/A]
    4. Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] See supplementary.
  4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets…
    1. If your work uses existing assets, did you cite the creators? [Yes]
    2. Did you mention the license of the assets? [Yes] See supplementary.
    3. Did you include any new assets either in the supplemental material or as a URL? [N/A]
    4. Did you discuss whether and how consent was obtained from people whose data you’re using/curating? [Yes] See supplementary.
    5. Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [Yes] See supplementary.
  5. If you used crowdsourcing or conducted research with human subjects…
    1. Did you include the full text of instructions given to participants and screenshots, if applicable? [Yes] See supplementary.
    2. Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A]
    3. Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]

Footnotes

  1. Tim Brooks, Aleksander Holynski, and Alexei A Efros. Instructpix2pix: Learning to follow image editing instructions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 18392–18402, 2023. 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17

  2. Kai Zhang, Lingbo Mo, Wenhu Chen, Huan Sun, and Yu Su. Magicbrush: A manually annotated dataset for instruction-guided image editing. Advances in Neural Information Processing Systems, 36, 2024. 2 3 4 5 6 7 8 9 10 11 12 13 14 15

  3. Shelly Sheynin, Adam Polyak, Uriel Singer, Yuval Kirstain, Amit Zohar, Oron Ashual, Devi Parikh, and Yaniv Taigman. Emu edit: Precise image editing via recognition and generation tasks. arXiv preprint arXiv:2311.10089, 2023. 2 3 4 5 6 7 8 9 10 11 12 13 14

  4. Amir Hertz, Ron Mokady, Jay Tenenbaum, Kfir Aberman, Yael Pritch, and Daniel Cohen-Or. Prompt-to-prompt image editing with cross attention control. CoRR, abs/2208.01626, 2022. URL https://doi.org/10.48550/arXiv.2208.01626. 2 3 4 5 6

  5. Tsung-Yi Lin, Michael Maire, Serge Belongie, Lubomir Bourdev, Ross Girshick, James Hays, Pietro Perona, Deva Ramanan, C. Lawrence Zitnick, and Piotr Dollár. Microsoft coco: Common objects in context, 2015. 2 3

  6. Harsh Agrawal, Karan Desai, Yufei Wang, Xinlei Chen, Rishabh Jain, Mark Johnson, Dhruv Batra, Devi Parikh, Stefan Lee, and Peter Anderson. Nocaps: Novel object captioning at scale. In Proceedings of the IEEE/CVF international conference on computer vision, pages 8948–8957, 2019a.

  7. Chenlin Meng, Yutong He, Yang Song, Jiaming Song, Jiajun Wu, Jun-Yan Zhu, and Stefano Ermon. SDEdit: Guided image synthesis and editing with stochastic differential equations. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=aBsCjcPu_tE. 2 3 4

  8. Axel Sauer, Dominik Lorenz, Andreas Blattmann, and Robin Rombach. Adversarial diffusion distillation, 2023.

  9. Dustin Podell, Zion English, Kyle Lacey, Andreas Blattmann, Tim Dockhorn, Jonas Müller, Joe Penna, and Robin Rombach. Sdxl: Improving latent diffusion models for high-resolution image synthesis, 2023. 2

  10. Youcai Zhang, Xinyu Huang, Jinyu Ma, Zhaoyang Li, Zhaochuan Luo, Yanchun Xie, Yuzhuo Qin, Tong Luo, Yaqian Li, Shilong Liu, et al. Recognize anything: A strong image tagging model. arXiv preprint arXiv:2306.03514, 2023c. 2

  11. Shilong Liu, Zhaoyang Zeng, Tianhe Ren, Feng Li, Hao Zhang, Jie Yang, Chunyuan Li, Jianwei Yang, Hang Su, Jun Zhu, et al. Grounding dino: Marrying dino with grounded pre-training for open-set object detection. arXiv preprint arXiv:2303.05499, 2023. 2

  12. Alexander Kirillov, Eric Mintun, Nikhila Ravi, Hanzi Mao, Chloe Rolland, Laura Gustafson, Tete Xiao, Spencer Whitehead, Alexander C Berg, Wan-Yen Lo, et al. Segment anything. arXiv preprint arXiv:2304.02643, 2023b.

  13. Rinon Gal, Or Patashnik, Haggai Maron, Amit H. Bermano, Gal Chechik, and Daniel Cohen-Or. Stylegan-nada: Clip-guided domain adaptation of image generators. ACM Trans. Graph., 41(4):141:1–141:13, 2022. URL https://doi.org/10.1145/3528223.3530164. 2

  14. Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Björn Ommer. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 10684–10695, June 2022c. 2 3

  15. Shu Zhang, Xinyi Yang, Yihao Feng, Can Qin, Chia-Chih Chen, Ning Yu, Zeyuan Chen, Huan Wang, Silvio Savarese, Stefano Ermon, et al. Hive: Harnessing human feedback for instructional visual editing. arXiv preprint arXiv:2303.09618, 2023b. 2 3

  16. Ron Mokady, Amir Hertz, Kfir Aberman, Yael Pritch, and Daniel Cohen-Or. Null-text inversion for editing real images using guided diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 6038–6047, 2023. 2 3 4

  17. Alexander Quinn Nichol, Prafulla Dhariwal, Aditya Ramesh, Pranav Shyam, Pamela Mishkin, Bob McGrew, Ilya Sutskever, and Mark Chen. GLIDE: towards photorealistic image generation and editing with text-guided diffusion models. In International Conference on Machine Learning, ICML 2022, 17-23 July 2022, Baltimore, Maryland, USA, volume 162 of Proceedings of Machine Learning Research, pages 16784–16804. PMLR, 2022. URL https://proceedings.mlr.press/v162/nichol22a.html. 2 3

  18. Omri Avrahami, Dani Lischinski, and Ohad Fried. Blended diffusion for text-driven editing of natural images. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022, New Orleans, LA, USA, June 18-24, 2022, pages 18187–18197. IEEE, 2022. URL https://doi.org/10.1109/CVPR52688.2022.01767. 2 3

  19. Xuan Ju, Ailing Zeng, Yuxuan Bian, Shaoteng Liu, and Qiang Xu. Direct inversion: Boosting diffusion-based editing with 3 lines of code. arXiv preprint arXiv:2310.01506, 2023.

  20. Xuan Ju, Xian Liu, Xintao Wang, Yuxuan Bian, Ying Shan, and Qiang Xu. Brushnet: A plug-and-play image inpainting model with decomposed dual-branch diffusion, 2024. URL https://arxiv.org/abs/2403.06976.

  21. Junhao Zhuang, Yanhong Zeng, Wenran Liu, Chun Yuan, and Kai Chen. A task is worth one word: Learning with task prompts for high-quality versatile image inpainting, 2024. URL https://arxiv.org/abs/2312.03594.

  22. Sihan Xu, Yidong Huang, Jiayi Pan, Ziqiao Ma, and Joyce Chai. Inversion-free image editing with natural language, 2023. URL https://arxiv.org/abs/2312.04965.

  23. Mingdeng Cao, Xintao Wang, Zhongang Qi, Ying Shan, Xiaohu Qie, and Yinqiang Zheng. Masactrl: Tuning-free mutual self-attention control for consistent image synthesis and editing, 2023. URL https://arxiv.org/abs/2304.08465.

  24. Zhi Gao, Yuntao Du, Xintong Zhang, Xiaojian Ma, Wenjuan Han, Song-Chun Zhu, and Qing Li. Clova: A closed-loop visual assistant with tool usage and update. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 13258–13268, June 2024.

  25. Katherine Crowson, Stella Biderman, Daniel Kornis, Dashiell Stander, Eric Hallahan, Louis Castricato, and Edward Raff. VQGAN-CLIP: open domain image generation and editing with natural language guidance. In Computer Vision - ECCV 2022 - 17th European Conference, Tel Aviv, Israel, October 23-27, 2022, Proceedings, Part XXXVII, volume 13697 of Lecture Notes in Computer Science, pages 88–105. Springer, 2022. URL https://doi.org/10.1007/978-3-031-19836-6_6.

  26. Xihui Liu, Zhe Lin, Jianming Zhang, Handong Zhao, Quan Tran, Xiaogang Wang, and Hongsheng Li. Open-edit: Open-domain image manipulation with open-vocabulary instructions. In Computer Vision - ECCV 2020 - 16th European Conference, Glasgow, UK, August 23-28, 2020, Proceedings, Part XI, volume 12356 of Lecture Notes in Computer Science, pages 89–106. Springer, 2020. URL https://doi.org/10.1007/978-3-030-58621-8_6.

  27. Lvmin Zhang, Anyi Rao, and Maneesh Agrawala. Adding conditional control to text-to-image diffusion models, 2023a.

  28. Nataniel Ruiz, Yuanzhen Li, Varun Jampani, Yael Pritch, Michael Rubinstein, and Kfir Aberman. Dreambooth: Fine tuning text-to-image diffusion models for subject-driven generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 22500–22510, June 2023.

  29. Xichen Pan, Li Dong, Shaohan Huang, Zhiliang Peng, Wenhu Chen, and Furu Wei. Kosmos-g: Generating images in context with multimodal large language models, 2024.

  30. Bahjat Kawar, Shiran Zada, Oran Lang, Omer Tov, Huiwen Chang, Tali Dekel, Inbar Mosseri, and Michal Irani. Imagic: Text-based real image editing with diffusion models. CoRR, abs/2210.09276, 2022. URL https://doi.org/10.48550/arXiv.2210.09276. 2

  31. Chitwan Saharia, William Chan, Saurabh Saxena, Lala Li, Jay Whang, Emily Denton, Seyed Kamyar Seyed Ghasemipour, Raphael Gontijo-Lopes, Burcu Karagol Ayan, Tim Salimans, Jonathan Ho, David J. Fleet, and Mohammad Norouzi. Photorealistic text-to-image diffusion models with deep language understanding. In Advances in Neural Information Processing Systems. URL https://openreview.net/forum?id=08Yk-n5l2Al.

  32. Dongxu Li, Junnan Li, and Steven C. H. Hoi. Blip-diffusion: Pre-trained subject representation for controllable text-to-image generation and editing, 2023.

  33. Wenhu Chen, Hexiang Hu, Yandong Li, Nataniel Ruiz, Xuhui Jia, Ming-Wei Chang, and William W. Cohen. Subject-driven text-to-image generation via apprenticeship learning, 2023b.

  34. Jian Ma, Junhao Liang, Chen Chen, and Haonan Lu. Subject-diffusion: Open domain personalized text-to-image generation without test-time fine-tuning. arXiv preprint arXiv:2307.11410, 2023.

  35. Narek Tumanyan, Michal Geyer, Shai Bagon, and Tali Dekel. Plug-and-play diffusion features for text-driven image-to-image translation, 2022.

  36. Su Wang, Chitwan Saharia, Ceslee Montgomery, Jordi Pont-Tuset, Shai Noy, Stefano Pellegrini, Yasumasa Onoe, Sarah Laszlo, David J. Fleet, Radu Soricut, Jason Baldridge, Mohammad Norouzi, Peter Anderson, and William Chan. Imagen editor and editbench: Advancing and evaluating text-guided image inpainting. CoRR, abs/2212.06909, 2022. URL https://doi.org/10.48550/arXiv.2212.06909.

  37. Su Wang, Chitwan Saharia, Ceslee Montgomery, Jordi Pont-Tuset, Shai Noy, Stefano Pellegrini, Yasumasa Onoe, Sarah Laszlo, David J Fleet, Radu Soricut, et al. Imagen editor and editbench: Advancing and evaluating text-guided image inpainting. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 18359–18369, 2023.

  38. Aditya Ramesh, Prafulla Dhariwal, Alex Nichol, Casey Chu, and Mark Chen. Hierarchical text-conditional image generation with CLIP latents. CoRR, abs/2204.06125, 2022. URL https://doi.org/10.48550/arXiv.2204.06125. 2

  39. Christoph Schuhmann, Romain Beaumont, Richard Vencu, Cade Gordon, Ross Wightman, Mehdi Cherti, Theo Coombes, Aarush Katta, Clayton Mullis, Mitchell Wortsman, Patrick Schramowski, Srivatsa Kundurthy, Katherine Crowson, Ludwig Schmidt, Robert Kaczmarczyk, and Jenia Jitsev. Laion-5b: An open large-scale dataset for training next generation image-text models, 2022.

  40. Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel M. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, and Dario Amodei. Language models are few-shot learners. In Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual, 2020. URL https://proceedings.neurips.cc/paper/2020/hash/1457c0d6bfcb4967418bfb8ac142f64a-Abstract.html.

  41. Mude Hui, Siwei Yang, Bingchen Zhao, Yichun Shi, Heng Wang, Peng Wang, Yuyin Zhou, and Cihang Xie. Hq-edit: A high-quality dataset for instruction-based image editing. arXiv preprint arXiv:2404.09990, 2024.

  42. Josh Achiam, Steven Adler, Sandhini Agarwal, Lama Ahmad, Ilge Akkaya, Florencia Leoni Aleman, Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al. Gpt-4 technical report. arXiv preprint arXiv:2303.08774, 2023.

  43. Gpt-4v(ision) system card. 2023. URL https://api.semanticscholar.org/CorpusID:263218031.

  44. James Betker, Gabriel Goh, Li Jing, † TimBrooks, Jianfeng Wang, Linjie Li, † LongOuyang, † JuntangZhuang, † JoyceLee, † YufeiGuo, † WesamManassra, † PrafullaDhariwal, † CaseyChu, † YunxinJiao, and Aditya Ramesh. Improving image generation with better captions. URL https://api.semanticscholar.org/CorpusID:264403242.

  45. Tsung-Yi Lin, Michael Maire, Serge J. Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C. Lawrence Zitnick. Microsoft COCO: common objects in context. In Computer Vision - ECCV 2014 - 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part V, volume 8693 of Lecture Notes in Computer Science, pages 740–755. Springer, 2014. URL https://doi.org/10.1007/978-3-319-10602-1_48.

  46. Peter Young, Alice Lai, Micah Hodosh, and Julia Hockenmaier. From image descriptions to visual denotations: New similarity metrics for semantic inference over event descriptions. Transactions of the Association for Computational Linguistics, 2, 2014.

  47. Harsh Agrawal, Karan Desai, Yufei Wang, Xinlei Chen, Rishabh Jain, Mark Johnson, Dhruv Batra, Devi Parikh, Stefan Lee, and Peter Anderson. nocaps: novel object captioning at scale. In ICCV, pages 8948–8957, 2019b.

  48. Jeffrey P Bigham, Chandrika Jayant, Hanjie Ji, Greg Little, Andrew Miller, Robert C Miller, Robin Miller, Aubrey Tatarowicz, Brandyn White, Samual White, et al. Vizwiz: nearly real-time answers to visual questions. In Proceedings of the 23nd annual ACM symposium on User interface software and technology, pages 333–342, 2010.

  49. Oleksii Sidorov, Ronghang Hu, Marcus Rohrbach, and Amanpreet Singh. Textcaps: a dataset for image captioning with reading comprehension, 2020.

  50. Jordi Pont-Tuset, Jasper Uijlings, Soravit Changpinyo, Radu Soricut, and Vittorio Ferrari. Connecting vision and language with localized narratives. In ECCV, 2020.

  51. Lin Chen, Jinsong Li, Xiaoyi Dong, Pan Zhang, Conghui He, Jiaqi Wang, Feng Zhao, and Dahua Lin. Sharegpt4v: Improving large multi-modal models with better captions, 2023a. 2 3

  52. Christoph Schuhmann and Peter Bevan. 220k-gpt4vision-captions-from-lvis. https://huggingface.co/datasets/laion/220k-GPT4Vision-captions-from-LIVIS, 2023. 2

  53. Pengxiang Li, Zhi Gao, Bofei Zhang, Tao Yuan, Yuwei Wu, Mehrtash Harandi, Yunde Jia, Song-Chun Zhu, and Qing Li. Fire: A dataset for feedback integration and refinement evaluation of multimodal models, 2024a. URL https://arxiv.org/abs/2407.11522.

  54. Haozhe Zhao, Zefan Cai, Shuzheng Si, Xiaojian Ma, Kaikai An, Liang Chen, Zixuan Liu, Sheng Wang, Wenjuan Han, and Baobao Chang. Mmicl: Empowering vision-language model with multi-modal in-context learning. arXiv preprint arXiv:2309.07915, 2023.

  55. Yue Fan, Xiaojian Ma, Rujie Wu, Yuntao Du, Jiaqi Li, Zhi Gao, and Qing Li. Videoagent: A memory-augmented multimodal agent for video understanding. In Aleš Leonardis, Elisa Ricci, Stefan Roth, Olga Russakovsky, Torsten Sattler, and Gül Varol, editors, Computer Vision – ECCV 2024, pages 75–92, Cham, 2025. Springer Nature Switzerland. ISBN 978-3-031-72670-5.

  56. Sirui Hong, Mingchen Zhuge, Jonathan Chen, Xiawu Zheng, Yuheng Cheng, Ceyao Zhang, Jinlin Wang, Zili Wang, Steven Ka Shing Yau, Zijuan Lin, Liyang Zhou, Chenyu Ran, Lingfeng Xiao, Chenglin Wu, and Jürgen Schmidhuber. Metagpt: Meta programming for a multi-agent collaborative framework, 2023. URL https://arxiv.org/abs/2308.00352.

  57. Shuzheng Si, Wentao Ma, Haoyu Gao, Yuchuan Wu, Ting-En Lin, Yinpei Dai, Hangyu Li, Rui Yan, Fei Huang, and Yongbin Li. Spokenwoz: A large-scale speech-text benchmark for spoken task-oriented dialogue agents, 2024. URL https://arxiv.org/abs/2305.13040.

  58. Yunshui Li, Binyuan Hui, Xiaobo Xia, Jiaxi Yang, Min Yang, Lei Zhang, Shuzheng Si, Ling-Hao Chen, Junhao Liu, Tongliang Liu, Fei Huang, and Yongbin Li. One-shot learning as instruction data prospector for large language models, 2024b. URL https://arxiv.org/abs/2312.10302.

  59. Timo Schick, Jane Dwivedi-Yu, Roberto Dessì, Roberta Raileanu, Maria Lomeli, Eric Hambro, Luke Zettlemoyer, Nicola Cancedda, and Thomas Scialom. Toolformer: Language models can teach themselves to use tools. Advances in Neural Information Processing Systems, 36, 2024.

  60. Alexander Kirillov, Eric Mintun, Nikhila Ravi, Hanzi Mao, Chloe Rolland, Laura Gustafson, Tete Xiao, Spencer Whitehead, Alexander C. Berg, Wan-Yen Lo, Piotr Dollár, and Ross Girshick. Segment anything. arXiv:2304.02643, 2023a.

  61. Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Björn Ommer. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 10684–10695, June 2022b.

  62. Patrick von Platen, Suraj Patil, Anton Lozhkov, Pedro Cuenca, Nathan Lambert, Kashif Rasul, Mishig Davaadorj, Dhruv Nair, Sayak Paul, William Berman, Yiyi Xu, Steven Liu, and Thomas Wolf. Diffusers: State-of-the-art diffusion models. https://github.com/huggingface/diffusers, 2022.

  63. Jiaming Song, Chenlin Meng, and Stefano Ermon. Denoising diffusion implicit models. In 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021. OpenReview.net, 2021. URL https://openreview.net/forum?id=St1giarCHLP.

  64. Ralf Herbrich, Tom Minka, and Thore Graepel. Trueskill™: a bayesian skill rating system. Advances in neural information processing systems, 19, 2006. 2 3

  65. Patrick Esser, Sumith Kulal, Andreas Blattmann, Rahim Entezari, Jonas Müller, Harry Saini, Yam Levi, Dominik Lorenz, Axel Sauer, Frederic Boesel, Dustin Podell, Tim Dockhorn, Zion English, Kyle Lacey, Alex Goodwin, Yannik Marek, and Robin Rombach. Scaling rectified flow transformers for high-resolution image synthesis, 2024. URL https://arxiv.org/abs/2403.03206.

  66. Timnit Gebru, Jamie Morgenstern, Briana Vecchione, Jennifer Wortman Vaughan, Hanna Wallach, Hal Daumé Iii, and Kate Crawford. Datasheets for datasets. Communications of the ACM, 64(12):86–92, 2021.