When Model Knowledge meets Diffusion Model: Diffusion-assisted Data-free Image Synthesis with Alignment of Domain and Class
Abstract
Open-source pre-trained models hold great potential for diverse applications, but their utility declines when their training data is unavailable. Data-Free Image Synthesis (DFIS) aims to generate images that approximate the learned data distribution of a pre-trained model without accessing the original data. However, existing DFIS methods produce samples that deviate from the training data distribution due to the lack of prior knowledge about natural images. To overcome this limitation, we propose DDIS, the first Diffusion-assisted Data-free Image Synthesis method that leverages a text-to-image diffusion model as a powerful image prior, improving synthetic image quality. DDIS extracts knowledge about the learned distribution from the given model and uses it to guide the diffusion model, enabling the generation of images that accurately align with the training data distribution. To achieve this, we introduce Domain Alignment Guidance (DAG) that aligns the synthetic data domain with the training data domain during the diffusion sampling process. Furthermore, we optimize a single Class Alignment Token (CAT) embedding to effectively capture class-specific attributes in the training dataset. Experiments on PACS and ImageNet demonstrate that DDIS outperforms prior DFIS methods by generating samples that better reflect the training data distribution, achieving SOTA performance in data-free applications.
1 Introduction
The widespread availability of open-source pre-trained models have significantly advanced the field of deep learning (Ridnik et al., 2021; Singh et al., 2023; Goldblum et al., 2024). Recently, platforms such as Hugging Face (Wolf, 2019) have further enhanced accessibility of the pre-trained model, enabling researchers to easily apply these models to various tasks like knowledge distillation and model pruning (Tran et al., 2024; Wang et al., 2024; Yin et al., 2020). However, a common requirement for utilizing these models in such applications is access to their training data, either in full or in part, for training student networks or fine-tuning pruned models. Unfortunately, these training datasets are often inaccessible due to various reasons, including data privacy and copyright issues. Consequently, the lack of access to training data presents a challenge in leveraging the potential of pre-trained models across various applications.
To address the above issue, Data-Free Image Synthesis (DFIS) has been proposed as a solution. DFIS aims to synthesize images that approximate the training data distribution by extracting the model’s internal understanding of its training data. This enables the application of these models to tasks like data-free knowledge distillation or pruning, enhancing their utility (Yin et al., 2020; Mordvintsev et al., 2015; Kim et al., 2022; Ghiasi et al., 2022). However, since existing DFIS methods generate images without prior knowledge of natural images, the image search space becomes huge. Consequently, they produce images with unnatural or artificial patterns that deviate from the training data distribution, ultimately limiting the model’s utility.
In this paper, we introduce a novel DFIS paradigm that leverages the inherent natural image understanding of an off-the-shelf Text-to-Image (T2I) diffusion model. Recent breakthroughs in T2I diffusion models trained on large-scale open-world datasets (Rombach et al., 2022; Saharia et al., 2022) provide powerful image priors capable of significantly narrowing the search space on DFIS. However, naively replacing the original training data with images generated by the T2I diffusion model is insufficient. Due to the inaccessibility of information regarding the training dataset, encompassing its domain, class-specific attributes, the input prompts of the T2I diffusion model are necessarily imprecise (e.g., A dog). This vagueness leads to the generation of countless images that do not align with the training set distribution. Consequently, without a guiding mechanism based on the pre-trained model’s knowledge of its training data distribution, the generated images are highly susceptible to deviating from the underlying distribution, ultimately limiting the model’s utility in various applications.

Therefore, we propose a Diffusion-assisted Data-free Image Synthesis (DDIS), which guides the T2I diffusion model to generate images that are closely aligned with the training set distribution. Our approach tackles the misalignment problem that arises when directly substituting the training set with images synthesized by a T2I diffusion model. Specifically, our method focuses on achieving alignment in two key aspects: 1) domain and 2) class-specific detail.
Firstly, to capture the domain of the training data, we introduce a Domain Alignment Guidance (DAG) that guides the image latent during the diffusion sampling process. We are inspired by the understanding that Batch Normalization (BN) layers encode the domain knowledge of the training set (Wu et al., 2024; Wang et al., 2020). Specifically, the running statistics in BN layers, derived from the mean and variance of all training batches, capture the distribution of the entire training set, including its domain knowledge. Consequently, DAG guides the diffusion model in aligning the internal statistics of the synthesized samples with the running statistics within the given model. This alignment effectively induces and strengthens the domain-specific features of the training dataset in the synthesized images.
Secondly, we propose to find a new embedding vector corresponding to a pseudo-word, Class Alignment Token (CAT), that represents specific concepts of the target class. CAT is a token designed to capture the class details of training data that are not revealed on the class label name itself. For example, while a class label might be ‘dog,’ the specific dog breeds within the training set are not specified. We aim to find the embedding vector of CAT, ensuring that the synthetic images accurately align with the target class the model was trained on. Interestingly, CAT can not only capture detailed class features but also resolve the lexical ambiguity of class labels. We observe that CAT generates images precisely associated with the intended class when a class label is a homonym or a compound noun.
Conclusively, our proposed method, leveraging a T2I diffusion model as a strong image prior, synthesizes samples closely aligned with the training set distribution. DDIS enables replacing inaccessible original training data, ultimately enhancing the pre-trained model’s utility. DDIS outperforms existing DFIS methods in capturing the training data distribution across various domains, such as art, cartoons, and manga, as well as across the various 1000 classes within ImageNet-1k. Moreover, DDIS achieves state-of-the-art results in data-free knowledge distillation and pruning, verifying the benefit of our synthesized data in enhancing the given model’s utility. Our key contributions are:
-
•
We propose a DDIS, a novel Data-Free Image Synthesis (DFIS) method that pioneers the use of a Text-to-Image (T2I) diffusion model as a strong image prior to address the issue of existing DFIS methods producing images distant from the training set distribution due to an vast image search space.
-
•
To ensure alignment at both the domain and class levels, we introduce Domain Alignment Guidance (DAG) and Class Alignment Token (CAT). DAG guides diffusion sampling to align the statistics of generated images with the internal statistics of the pre-trained model and CAT embedding encodes class-specific details.
-
•
Experimental results show that DDIS outperforms prior DFIS methods by generating samples closely aligned with the original training set, leading to state-of-the-art performance in various data-free applications.
2 Related Works
2.1 Data-Free Image Synthesis (DFIS)
DFIS aims to approximate the distribution learned by a given classifier by solving an optimization problem in the input space without accessing the training data (Kim et al., 2022). DeepDream (Mordvintsev et al., 2015) visualizes the patterns learned by the model for specific classes by iteratively optimizing noise to maximize the output probability for the desired class. DeepInversion (DI) (Yin et al., 2020) tackles the limited image prior in traditional DFIS methods and introduces a regularizer that ensures the statistics of synthetic images follow the running statistics within the model’s Batch Normalization (BN) layers (Ioffe & Szegedy, 2015). Plug-In-Inversion (PII) (Ghiasi et al., 2022) addresses the issue of low diversity in generated images by combining various data augmentations during optimization. Despite these successes, finding an optimal that closely approximates within a vast search space remains a significant challenge without access to the training data. Consequently, the generated images are distant from the training samples.
2.2 Diffusion Models
Diffusion models (Ho et al., 2020; Sohl-Dickstein et al., 2015; Song et al., 2020) have garnered considerable attention due to their exceptional sampling quality. Initiating the process with a random noise sample , the diffusion model iteratively predicts a denoised sample at each time step . This is known as the reverse process:
(1) |
where is noise scale at time and is given by
(2) |
where is the noise predicted by the U-Net (Ho et al., 2020).
Sampling Guidance for Diffusion Models.
Classifier-assisted guidance methods contribute to the advancement of diffusion models by allowing the adjustment of outputs to generate desired images. From a score function perspective (Song et al., 2020), we can modify the unconditional score function to , controlling image generation to produce results that align with a given condition , such as a text prompt or class label.
Classifier Guidance (CG) (Dhariwal & Nichol, 2021) generates class-conditional images by factorizing to the unconditional score and the gradient derived from the classifier by Bayes’ rule . The classifier guidance can be formulated as the following with a guidance scale :
(3) |
CG requires generating training data at each time step and training a time-dependent external classifier , which is not feasible in DFIS due to the lack of access to labeled datasets.
Classifier-Free Guidance (CFG) (Ho & Salimans, 2022) proposes a method for achieving the effects of classifier guidance without the need for an additional classifier by interpolating between unconditional and conditional outputs.
(4) |
Unfortunately, despite this efficiency, CFG cannot provide guidance on the domain of the training set learned by the classifier, making it insufficient for standalone use in DFIS.
3 Method
First, Section 3.1 introduces the preliminaries of data-free image synthesis and diffusion models. Section 3.2 explains the concept of Diffusion-assisted Data-free Image Synthesis (DDIS). Then, Section 3.3 details Domain Alignment Guidance (DAG), which directs the image latent during diffusion sampling, and Class Alignment Token (CAT) in Section 3.4, which captures class details for accurate class alignment.
3.1 Preliminary
Data-Free Image Synthesis.
The goal of Data-Free Image Synthesis (DFIS) is to find the optimal point within the high-dimensional input space that elicits a maximum response from the classifier for a desired class under data-free conditions. This process involves iteratively updating a random noise into a visually natural image by optimizing the following loss function:
(5) |
where is Cross-Entropy loss, and is an image prior regularizer. Despite the potential of DFIS, finding an optimal point in a vast search space requires huge computational costs under data-free settings. Although DeepDream (Mordvintsev et al., 2015) explores a total variation regularizer to generate visually plausible outputs, the resulting images lack the naturalness and fidelity of the real samples.
Require: Model(Pre-trained unconditional score estimator , text encoder , image decoder , prompt , classifier pre-trained on unknown dataset
Output: Latent
Require: Pre-trained T2I diffusion model Model, classifier pre-trained on unknown dataset
Parameter: Class Alignment token embedding
Output: Sampling images
Text Conditioned Latent Diffusion Models.
Text Conditioned Latent Diffusion models (LDM) (Rombach et al., 2022) operate in the latent space by projecting the image to through the auto-encoder . In the diffusion sampling process, starting from , we subsequently obtain the denoised latent and finally generate the clean image by passing the to the decoder . Expressly, since the text conditioned LDMs focus on generating images guided by the given text prompt encoded by the text encoder as a condition, and the predicted noise is .
Guidance for Text-to-Image Diffusion Models.
As mentioned in Section 2.2, Classifier-Free Guidance (CFG), allows the class label or text to be used as a condition in the diffusion process can be represented in the latent space by modifying Equation 4 as follows:
(6) |
where means a null condition (unconditioned) and is latent variable in time step given by
(7) |
3.2 Diffusion-Assisted Data-free Image Synthesis
In this section, we introduce Diffusion-assisted Data-free Image Synthesis (DDIS), which is a pioneer of the use of a Text-to-Image (T2I) diffusion model as a strong image prior to address the issue of existing DFIS methods producing images distant from the training set distribution due to a vast image search space. We utilize the Stable Diffusion (SD) (Rombach et al., 2022), which has achieved remarkable success in T2I diffusion models. By leveraging the comprehensive knowledge of natural images embedded in this pre-trained T2I diffusion model, DDIS can produce samples that better approximate the learned distribution than current DFIS methods. However, since the images generated by T2I diffusion models still differ from the exact training set distribution, additional guidance is necessary to ensure they align properly with the training set.
3.3 Domain Alignment Guidance
We propose an effectively applicable guidance method for the diffusion model, called Domain Alignment Guidance (DAG), which provides guidance on the domain learned by the classifier during the diffusion sampling process. Inspired by studies suggesting that Batch Normalization (BN) layers encode domain knowledge (Lim et al., 2023; Mirza et al., 2022), Domain Alignment Guidance (DAG) guides the statistics of synthesized images to align with the training set statistics stored in the model’s BN layers. Specifically, we can obtain the channel-wise running mean and running variance of the entire training set from all BN layers in the given model . To guide the image latent so that the statistics of the generated images follow the running statistics, we modify the unconditional score of the image latent to from the perspective of SDE (Song et al., 2020). Using Bayes’ rule, the conditional score can be factorized as below two terms:
(8) |
where is a parameter controlling the guidance strength. We can interpret the second term as the gradient with respect to the latent obtained from an external loss function, which minimizes the difference between the running statistics and those of the synthetic image:
(9) |
where and represent the mean and variance of the feature map at the -th layer of the generated image . In this process, we project the image latent to the pixel space at each time step , by , to compute the gradient. We replace with the obtained gradient w.r.t the latent variable, thereby providing guidance to the image latent in the direction of the training set distribution as below:
(10) |
where is the scaling factor for the gradient influence. Finally, the DAG is integrated with the CFG to guide the latent toward the target distribution, enabling the T2I diffusion model to sample images faithful to the prompt while incorporating domain knowledge within the classifier.
(11) |
In conclusion, starting from noise , we estimate the noise based on guided latent via the DAG. This process iteratively updates the latent to a cleaner one, ultimately sampling images that is aligned with the target domain by modifying Equation 7.
(12) |
The overall algorithm of DAG is described in Algorithm 1.



3.4 Class Alignment Token
To capture the accurate class’s property, we encode the semantic information about class details into the word embedding of the proposed Class Alignment Token (CAT). In a T2I diffusion model, words or sub-words from the input prompt are converted into tokens from a predefined vocabulary, and a pre-trained text encoder maps each token to a unique word embedding. We chose this embedding space for optimization in DDIS because this space encodes rich semantic representations learned by the pre-trained text encoder, which we expect to capture the semantic features of the desired class effectively. Inspired by (Gal et al., 2022; Huang et al., 2024; Kim et al., 2025), which encodes specific concepts into pseudo-word token embeddings, we define as a CAT, to precisely represent the desired class . Then, we construct a simple prompt using the given class label as “A/An {class label}”. We expand the vocabulary by adding the CAT and only optimize its token embedding of the using Cross-Entropy (C.E.) loss. This iterative process aims to find the optimal that captures the class-specific information learned by the model, as follows:
(13) |
where indicates output logits given the pre-trained model and . Specifically, we compute the C.E. loss only for the final image obtained through the DAG-based diffusion sampling process up to the final time step, as this image closely resembles the distribution observed by the classifier. (i.e. ). Thus, even without access to the training set or additional training processes, we can capture the class-specific attribute by optimizing only the CAT embedding. Furthermore, once the optimal CAT embedding is found, we can quickly generate images of the desired class through a sampling process. Consequently, DDIS can generate images that approximate the training set via DAG in the sampling process and CAT optimization in the T2I diffusion model, ensuring significant improvements in image quality compared to existing DFIS methods. The overall framework is described in Figure 1, and the algorithm can be found in Algorithm 2.
Dataset | Domain | DDIS (Ours) | DeepInversion | PlugInInversion | |||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
IS() | FID() | Precision() | Recall() | IS() | FID() | Precision() | Recall() | IS() | FID() | Precision() | Recall() | ||
ImageNet-1k | Photo | 15.92 | 30.31 | 0.8028 | 0.7731 | 9.52 | 187.63 | 0.5038 | 0.0907 | 3.51 | 220.62 | 0.2969 | 0.0528 |
PACS | Art painting | 4.12 | 133.37 | 0.7742 | 0.3213 | 4.00 | 188.53 | 0.4033 | 0.0004 | 2.53 | 208.73 | 0.3442 | 0.0014 |
Cartoon | 4.04 | 85.41 | 0.7541 | 0.3104 | 3.91 | 148.94 | 0.3704 | 0.0038 | 2.81 | 275.86 | 0.0004 | 0.0001 | |
Style-Aligned | Caricature | 3.94 | 139.75 | 0.3552 | 0.5672 | 3.58 | 195.25 | 0.1027 | 0.0618 | 2.51 | 293.58 | 0.0001 | 0.0004 |
Manga | 3.87 | 145.82 | 0.2036 | 0.4396 | 3.32 | 206.57 | 0.0834 | 0.0019 | 2.36 | 295.14 | 0.0001 | 0.0003 |
4 Experiments
4.1 Experimental Settings
Baselines.
We mainly compare our method with existing data-free image synthesis (DFIS) methods such as DeepInversion (DI) (Yin et al., 2020) and PlugInInversion (PII) (Ghiasi et al., 2022), which deal with the large-scale datasets synthesis inverting the CNNs. We do not compare the performance the NaturalInversion (NI) (Kim et al., 2022) since NI only considers small-scale datasets.
Datasets.
We utilize ResNets (He et al., 2016) trained on five domains to generate images. For the photo domain, we use ImageNet-1k (Russakovsky et al., 2015). For the art painting and cartoon domain, we use the PACS dataset (Li et al., 2017). Additionally, we collect artificial datasets via Style Aligned (Hertz et al., 2024) for manga and caricature domains. Each dataset includes 7 classes. Detailed experimental settings and class labels are in the appendix.

4.2 Main Results
Image Synthesis with Various Domain.
DDIS is the first to succeed in generating samples from various domains, not just the photo domain, in the DFIS. Figures 2, 3 and 4 compare images generated with the models pre-trained on diverse domain datasets with baselines. Since we do not know the knowledge of the dataset used for training the classifier, existing DFIS methods must explore an extremely large image search space. This leads to generated images with artifacts that fail to accurately capture the training dataset’s domain properties. As a result, they fail to capture the domain properties of the training dataset and usually generate images with severe artifact effects. Although Stable Diffusion can generate class-faithful images given prompts such as ‘A dog’, it lacks guidance on the domain of the training data, preventing it from capturing the dataset’s domain-specific information. In contrast, the proposed DDIS leverages DAG in the diffusion sampling process and optimizes CAT embeddings to synthesize images that accurately reflect the training dataset’s domain and class attributes, producing images that closely resemble the original data.
Image Synthesis Faithful to the Target Class.
We demonstrate how CAT embedding optimization enhances the capture of target class attributes. Figure 5 presents images generated with a ResNet-34 pre-trained on ImageNet-1k. Surprisingly, DDIS achieves precise class mappings, even resolving the lexical overlap problem, where generators misinterpret class labels. For instance, suppose the classifier has learned the tiger cat, as shown in Figure 5 (a) (though we do not know the training set). However, Stable Diffusion (SD) incorrectly generates a tiger due to word similarity. This issue becomes even more pronounced with compound nouns, such as ‘beach wagon’ or ‘mailbag’ where ambiguity leads to incorrect images. Even for single-word labels (Figure 5 (d)), SD struggles with homonyms, reducing accuracy. In contrast, DDIS overcomes these challenges by optimizing only the CAT embedding, ensuring that the generated images accurately align with the intended class.
Dataset | T. | S. | Original | DI | PII | SD | Ours |
---|---|---|---|---|---|---|---|
PACS-art | ResNet-34 | ResNet-18 | 32.60 | 17.49 | 12.27 | 21.93 | 28.46 |
ResNet-34 | VGG-11 | 26.03 | 12.65 | 9.23 | 18.12 | 21.89 | |
ResNet-50 | ResNet-18 | 37.98 | 18.48 | 15.32 | 31.67 | 35.87 | |
ResNet-50 | ResNet-34 | 28.46 | 16.63 | 13.36 | 19.52 | 24.65 | |
VGG-16 | VGG-11 | 32.89 | 17.37 | 12.47 | 22.61 | 27.43 | |
VGG-16 | ShuffleNetv2 | 54.25 | 30.32 | 23.14 | 41.74 | 48.56 | |
PACS-cartoon | ResNet-34 | ResNet-18 | 51.35 | 28.65 | 17.23 | 38.12 | 47.06 |
ResNet-34 | VGG-11 | 49.95 | 26.92 | 15.62 | 38.94 | 44.92 | |
ResNet-50 | ResNet-18 | 58.15 | 31.48 | 20.32 | 44.67 | 51.45 | |
ResNet-50 | ResNet-34 | 47.34 | 24.05 | 15.66 | 36.73 | 42.93 | |
VGG-16 | VGG-11 | 48.82 | 25.29 | 16.04 | 38.15 | 44.81 | |
VGG-16 | ShuffleNetv2 | 54.24 | 29.94 | 19.32 | 43.63 | 49.32 | |
ImageNet-1k | ResNet-34 | ResNet-18 | 43.30 | 4.67 | 2.01 | 33.02 | 41.68 |
ResNet-34 | VGG-11 | 34.91 | 2.67 | 1.34 | 27.41 | 32.71 | |
ResNet-50 | ResNet-18 | 43.09 | 6.31 | 1.98 | 35.43 | 40.77 | |
ResNet-50 | ResNet-34 | 44.57 | 7.29 | 3.08 | 33.15 | 42.87 | |
VGG-16 | VGG-11 | 34.70 | 2.55 | 1.23 | 26.93 | 32.14 | |
VGG-16 | ShuffleNetv2 | 28.13 | 1.37 | 1.02 | 18.52 | 24.61 |

Quantitative Results.
We compare the image quality with existing DFIS studies to evaluate whether the generated images approximate the distribution of the training set used to train the model. For this evaluation, we use (1) Inception Score (IS) (Reed et al., 2016) and Frechet Inception Distance (FID) (Heusel et al., 2017), and (2) Precision and Recall (P&R) (Sajjadi et al., 2018), to measure the fidelity and diversity of synthetic images strictly. We evaluate the image quality of 10,000 synthetic ImageNet-1k samples, 2,800 synthetic PACS samples, and 1,400 synthetic Style-Aligned samples. As shown in Table 1, DDIS outperforms the baselines across all metrics. Our proposed method generates images that most closely approximate the training set distribution, effectively producing images that align with both the class and domain of the training set.
4.3 Data-Free Applications
DDIS aims to enhance the utility of a given model by generating samples that approximate the distribution of the training data. Accordingly, we conduct experiments on Knowledge Distillation (KD) and Pruning using synthetic images without direct access to the training data. Specifically, we synthesize 2,800 images from the PACS dataset and 100k images from ImageNet-1k for these experiments.
Data-Free Knowledge Distillation.
In this section, we ensure that we can transfer information from a teacher network to a student network without original data and outperform existing data-free knowledge distillation approaches. The experimental setup for knowledge distillation is based on the protocol outlined in (Li et al., 2023). Our method is compared against previous DFIS approaches, including DI and PII, as well as the use of images generated by Stable Diffusion (SD).
Table 2 demonstrates the superior performance of the proposed approach. Our method consistently achieves results closer to the baseline across all datasets compared to prior studies. Notably, it better approximates the training set distribution than directly using SD outputs as training data. This suggests that our method generates samples that are more aligned with the domain and class distributions of the original training data.
Data-Free Pruning.
We demonstrate that DDIS enhances pruned model accuracy without using real data. We apply L1-norm pruning (Liu et al., 2017) to ResNet-34 and VGG-16 on PACS art painting, cartoon and ImageNet-1k with pruning ratios from 50% to 90%. Following (Liu et al., 2018) settings, we locally prune the least important channels in each layer at the specified ratios. Similar to DFKD results, DDIS synthesizes samples closely matching the underlying distribution, enhancing model utility, as shown in Figure 6.
4.4 Further Analysis
Ablation Study.
To understand the impact of each component of DDIS on image synthesis, we evaluate the performance of various component combinations on the PACS art painting dataset, as described in Figure 8 and Table 3. When directly generating images from vanilla SD, the resulting images lack the information of the training set and fail to capture the unique characteristics of each class and domain, leading to significant deviations from the training set. The DAG leads to synthetic images that better reflect the target domain by following the inner statistics of the training set encapsulated in the model. The optimal embedding of the CAT captures class-specific features, generating images with high fidelity to the class label, but if not used with DAG, domain discrepancy issues arise. When all proposed components are used together, the resulting images align with both the domain and class of the training set.

Impact of Optimal CAT Embedding on Synthetic Image.
We evaluate if the optimal Class Alignment Token (CAT) embedding captures the desired class’s high-level semantics and fine details. Figure 7 shows the influence of the CAT embedding vector by scaling the cross-attention map with parameter from 1.0 to 30.0. As increases, the CAT embedding accentuates class-specific features, such as the ‘eye shape’ in the tiger cat images, showing its ability to capture detailed class attributes. Furthermore, strengthening the CAT embedding improves alignment with the target domain, showcasing its capability to represent precise visual details and maintain consistency with the training set distribution.

Method | DAG | CAT | IS() | FID() | Precision() | Recall() |
---|---|---|---|---|---|---|
SD | 2.88 | 193.57 | 0.6429 | 0.2572 | ||
SD w/o DAG | 3.29 | 174.31 | 0.6995 | 0.3074 | ||
SD w/o CAT | 3.95 | 166.22 | 0.6871 | 0.2843 | ||
Ours | 4.12 | 133.37 | 0.7742 | 0.3213 |

5 Limitations
While DDIS excelled across domains, it struggled in the Sketch domain, which has abstract representations of objects and scenes that make image generation particularly challenging, as shown in Figure 9. However, DDIS captures common Sketch characteristics, such as monochrome backgrounds and darker tones. Additionally, since our DAG relies on feature statistics from the BN layer, it can only be applied to models with BN layers. In future work, we plan to explore model-agnostic domain alignment guidance.
6 Conclusion
We introduce DDIS, the first Diffusion-assisted Data-free Image Synthesis method, using a T2I diffusion model as a powerful image prior to narrow the image search space. We introduce Domain Alignment Guidance (DAG) for domain alignment during diffusion sampling and Class Alignment Token (CAT) embedding optimization for desired class alignment. We are also the first to conduct experiments across various domains in DFIS, proving its effectiveness.
Impact Statement
This work aims to enhance the utility of pre-trained models by generating synthetic data in scenarios where access to the original data is unavailable. We acknowledge that data-free image synthesis approaches, designed to recover data following the training dataset’s distribution, can raise concerns regarding privacy leakage and other ethical issues. However, our objective is not to reconstruct individual instances or facilitate unauthorized data access but rather to synthesize a surrogate dataset that can be effectively utilized for data-free applications such as data-free knowledge distillation or data-free pruning. Specifically, our method guides the data generation process by leveraging the running statistics from classifiers, which represent the averaged information of the entire dataset. This design inherently prevents the inclusion of individual instance details in the generated images, making it extremely difficult to recover any specific sample (see Figures 2, 3, 4 and 5 in the main paper). We hope that this work contributes to the responsible development of data-free techniques for scenarios in which data sharing is limited or infeasible.
Acknowledgements
This research was partly supported by the MSIT(Ministry of Science and ICT), Korea, under the ITRC(Information Technology Research Center) support program(IITP-2024-RS-2023-00258649, 50%) supervised by the IITP(Institute for Information & Communications Technology Planning & Evaluation), partly supported by the National Research Foundation of Korea(NRF) grant funded by the Korea government(MSIT) (RS-2025-00562437, 40%), and partly supported by Institute of Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korea government(MSIT) (No.RS-2022-00155911, Artificial Intelligence Convergence Innovation Human Resources Development (Kyung Hee University), 10%).
References
- Dhariwal & Nichol (2021) Dhariwal, P. and Nichol, A. Diffusion models beat gans on image synthesis. Advances in neural information processing systems, 34:8780–8794, 2021.
- Gal et al. (2022) Gal, R., Alaluf, Y., Atzmon, Y., Patashnik, O., Bermano, A. H., Chechik, G., and Cohen-Or, D. An image is worth one word: Personalizing text-to-image generation using textual inversion. arXiv preprint arXiv:2208.01618, 2022.
- Ghiasi et al. (2022) Ghiasi, A., Kazemi, H., Reich, S., Zhu, C., Goldblum, M., and Goldstein, T. Plug-in inversion: Model-agnostic inversion for vision with data augmentations. 2022.
- Goldblum et al. (2024) Goldblum, M., Souri, H., Ni, R., Shu, M., Prabhu, V., Somepalli, G., Chattopadhyay, P., Ibrahim, M., Bardes, A., Hoffman, J., et al. Battle of the backbones: A large-scale comparison of pretrained models across computer vision tasks. Advances in Neural Information Processing Systems, 36, 2024.
- He et al. (2016) He, K., Zhang, X., Ren, S., and Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770–778, 2016.
- Hertz et al. (2024) Hertz, A., Voynov, A., Fruchter, S., and Cohen-Or, D. Style aligned image generation via shared attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4775–4785, 2024.
- Heusel et al. (2017) Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., and Hochreiter, S. Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in neural information processing systems, 30, 2017.
- Ho & Salimans (2022) Ho, J. and Salimans, T. Classifier-free diffusion guidance. arXiv preprint arXiv:2207.12598, 2022.
- Ho et al. (2020) Ho, J., Jain, A., and Abbeel, P. Denoising diffusion probabilistic models. Advances in neural information processing systems, 33:6840–6851, 2020.
- Huang et al. (2024) Huang, Z., Wu, T., Jiang, Y., Chan, K. C., and Liu, Z. Reversion: Diffusion-based relation inversion from images. In SIGGRAPH Asia 2024 Conference Papers, pp. 1–11, 2024.
- Ioffe & Szegedy (2015) Ioffe, S. and Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In International conference on machine learning, pp. 448–456. pmlr, 2015.
- Kim et al. (2025) Kim, H., Kim, D., and Kim, S. Difference inversion: Interpolate and isolate the difference with token consistency for image analogy generation. In Proceedings of the Computer Vision and Pattern Recognition Conference, pp. 18250–18259, 2025.
- Kim et al. (2022) Kim, Y., Park, D., Kim, D., and Kim, S. Naturalinversion: Data-free image synthesis improving real-world consistency. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pp. 1201–1209, 2022.
- Li et al. (2017) Li, D., Yang, Y., Song, Y.-Z., and Hospedales, T. M. Deeper, broader and artier domain generalization. In Proceedings of the IEEE international conference on computer vision, pp. 5542–5550, 2017.
- Li et al. (2023) Li, Z., Li, Y., Zhao, P., Song, R., Li, X., and Yang, J. Is synthetic data from diffusion models ready for knowledge distillation? arXiv preprint arXiv:2305.12954, 2023.
- Lim et al. (2023) Lim, H., Kim, B., Choo, J., and Choi, S. Ttn: A domain-shift aware batch normalization in test-time adaptation. arXiv preprint arXiv:2302.05155, 2023.
- Liu et al. (2017) Liu, Z., Li, J., Shen, Z., Huang, G., Yan, S., and Zhang, C. Learning efficient convolutional networks through network slimming. In Proceedings of the IEEE international conference on computer vision, pp. 2736–2744, 2017.
- Liu et al. (2018) Liu, Z., Sun, M., Zhou, T., Huang, G., and Darrell, T. Rethinking the value of network pruning. arXiv preprint arXiv:1810.05270, 2018.
- Mirza et al. (2022) Mirza, M. J., Micorek, J., Possegger, H., and Bischof, H. The norm must go on: Dynamic unsupervised domain adaptation by normalization. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 14765–14775, 2022.
- Mordvintsev et al. (2015) Mordvintsev, A., Olah, C., and Tyka, M. Deepdream-a code example for visualizing neural networks. Google Research, 2(5), 2015.
- Paszke et al. (2019) Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., Antiga, L., et al. Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems, 32, 2019.
- Reed et al. (2016) Reed, S., Akata, Z., Yan, X., Logeswaran, L., Schiele, B., and Lee, H. Generative adversarial text to image synthesis. In International conference on machine learning, pp. 1060–1069. PMLR, 2016.
- Ridnik et al. (2021) Ridnik, T., Ben-Baruch, E., Noy, A., and Zelnik-Manor, L. Imagenet-21k pretraining for the masses. arXiv preprint arXiv:2104.10972, 2021.
- Rombach et al. (2022) Rombach, R., Blattmann, A., Lorenz, D., Esser, P., and Ommer, B. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 10684–10695, 2022.
- Russakovsky et al. (2015) Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Ma, S., Huang, Z., Karpathy, A., Khosla, A., Bernstein, M., et al. Imagenet large scale visual recognition challenge. International journal of computer vision, 115(3):211–252, 2015.
- Saharia et al. (2022) Saharia, C., Chan, W., Saxena, S., Li, L., Whang, J., Denton, E. L., Ghasemipour, K., Gontijo Lopes, R., Karagol Ayan, B., Salimans, T., et al. Photorealistic text-to-image diffusion models with deep language understanding. Advances in neural information processing systems, 35:36479–36494, 2022.
- Sajjadi et al. (2018) Sajjadi, M. S., Bachem, O., Lucic, M., Bousquet, O., and Gelly, S. Assessing generative models via precision and recall. arXiv preprint arXiv:1806.00035, 2018.
- Singh et al. (2023) Singh, M., Duval, Q., Alwala, K. V., Fan, H., Aggarwal, V., Adcock, A., Joulin, A., Dollár, P., Feichtenhofer, C., Girshick, R., et al. The effectiveness of mae pre-pretraining for billion-scale pretraining. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 5484–5494, 2023.
- Sohl-Dickstein et al. (2015) Sohl-Dickstein, J., Weiss, E., Maheswaranathan, N., and Ganguli, S. Deep unsupervised learning using nonequilibrium thermodynamics. In International conference on machine learning, pp. 2256–2265. PMLR, 2015.
- Song et al. (2020) Song, Y., Sohl-Dickstein, J., Kingma, D. P., Kumar, A., Ermon, S., and Poole, B. Score-based generative modeling through stochastic differential equations. arXiv preprint arXiv:2011.13456, 2020.
- Tran et al. (2024) Tran, M.-T., Le, T., Le, X.-M., Harandi, M., Tran, Q. H., and Phung, D. Nayer: Noisy layer data generation for efficient and effective data-free knowledge distillation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 23860–23869, 2024.
- Wang et al. (2020) Wang, D., Shelhamer, E., Liu, S., Olshausen, B., and Darrell, T. Tent: Fully test-time adaptation by entropy minimization. arXiv preprint arXiv:2006.10726, 2020.
- Wang et al. (2024) Wang, Y., Yang, D., Chen, Z., Liu, Y., Liu, S., Zhang, W., Zhang, L., and Qi, L. De-confounded data-free knowledge distillation for handling distribution shifts. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 12615–12625, 2024.
- Wolf (2019) Wolf, T. Huggingface’s transformers: State-of-the-art natural language processing. arXiv preprint arXiv:1910.03771, 2019.
- Wu et al. (2024) Wu, Y., Chi, Z., Wang, Y., Plataniotis, K. N., and Feng, S. Test-time domain adaptation by learning domain-aware batch normalization. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 38, pp. 15961–15969, 2024.
- Yin et al. (2020) Yin, H., Molchanov, P., Alvarez, J. M., Li, Z., Mallya, A., Hoiem, D., Jha, N. K., and Kautz, J. Dreaming to distill: Data-free knowledge transfer via deepinversion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 8715–8724, 2020.
Appendix A Experiment Settings
A.1 Hyper parameters
We generate a 512512 resolution image using Stable Diffusion 2.1 (Rombach et al., 2022) with 30 diffusion time steps. For applying Domain Alignment Guidance (DAG), we define the gradient scaling factor as the product of and the guidance scale . Here, is the scaling factor for , and denotes the DAG guidance scale. These two parameters play a crucial role in our image generation process, and their impact on synthetic images is discussed in the following subsection. DAG is incorporated with the Classifier-Free Guidance (CFG) with a scale of 15.
To define the Class Alignment Token (CAT), we newly define a single token with no inherent meaning and add it to the vocabulary, using its embedding as the initial value. We utilize the token “newcls” as the CAT, but any arbitrary token without meaning can serve as the initial token. To optimize the embedding vector of the CAT, we employ the Adam optimizer with a learning rate of 0.005. We train for up to 30 epochs, each involving 20 gradient accumulation steps and generating images from latent noise initialized with different seeds. It is equivalent to a batch size of 20 and can be adjusted based on resource availability.
A.2 Detail of in DAG
The main hyperparameter of our proposed Domain Alignment Guidance (DAG) in the diffusion sampling process is , which is divided into the gradient flow scale and the DAG guidance scale and we can rewrite the Equation 10 as below.
(14) |
To determine the optimal values for the gradient flow scale() and guidance scale(), we conduct experiments across a wide range of values. Specifically, we test between 0.0001 and 1, while the is explored within the range of 0.1 to 100. As a result, we find the optimal settings as a of 0.01 and a of 20. The experimental results are shown in the Figure 11. Users can adjust these two parameters as needed, with the selection of hyperparameters guided by the generated images’ confidence score.
A.3 Training Strategy
Prompt Design.
Due to the nature of Data-Free Image Synthesis (DFIS), where we have no information about the training set beyond class label information, we design prompts to be very simple and ambiguous (e.g., “A/An {class label}.”) To demonstrate that we can effectively approximate the desired distribution with minimal information, we refrain from using any prefixes like “A photo of .”
Gradient Skipping for CAT embedding optimization.
Gradient backpropagation through all diffusion steps demands substantial memory. In our experiments, limiting gradient propagation to just the final denoising step (i.e., step (=final )) reduces the resource usage. Additionally, since the given model was trained on natural images with a similar distribution of images from the final denoising step (i.e., ), the gradient skipping technique produces an appropriate loss for representing the intended class. Although deeper backpropagation could lead to further improvements, we do not explore this approach due to memory constraints.
Early Stopping Strategy.
We use a threshold of 0.7 for the proportion of correctly predicted samples in a batch to determine early stopping. During each epoch, the generated batch is evaluated by the classifier. If over 70% of the samples are accurately predicted as the target class, we apply early stopping for the CAT embedding optimization.
A.4 Computation Overhead
In our experiments using a single RTX 4090 GPU, the optimal CAT embedding is found in approximately 7.5 minutes. Once determined, this embedding remains fixed, enabling rapid image generation via a diffusion sampling process. Compared to existing Data-Free Image Synthesis (DFIS) methods, our DDIS enhances efficiency and cost-effectiveness for large-scale ImageNet-1k dataset generation while preserving high image quality. To illustrate this, we compare the total optimization iterations needed to synthesize 100,000 ImageNet-1k images on a single RTX 4090 GPU, shown in 4.
DeepInversion (DI) (Yin et al., 2020) requires iterative optimization for each mini-batch. With an optimal batch size of 250, around 20,000 iterations per batch are needed. Generating 100,000 images necessitates repeating this 400 times, totaling 8,000,000 iterations (20,000 iterations per a mini-batch 400 batches). PlugInInversion (PII) (Ghiasi et al., 2022) optimizes a random input across seven progressive upsampling stages (77 to 224224), with 400 iterations per stage. Generating 100,000 images with a batch size of 250 results in approximately 1,120,000 total iterations (400 iterations per stage 7 stages 400 batches).
Our DDIS optimizes only class-wise CAT embedding vectors in a low-dimensional space (1784). For ImageNet-1k (1000 classes), we perform just 30 iterations per class, totaling 30,000 iterations (30 iterations/class1000 classes). After finding the CAT embedding, generating 100,000 images involves simply sampling latent vectors without further training or optimization. While CAT optimization has a slightly longer per-iteration time than prior methods, the drastically reduced total iterations make the overall process more efficient.
Cost | DeepInversion | PlugInInversion | DDIS (Ours) |
---|---|---|---|
Total Iteration for 100k samples synthesis | 8,000K | 1,120K | 30K |
Times per 1 iteration (sec) | 0.83 | 0.79 | 15.17 |
Total training cost (hours) | 18444 | 245 | 126 |
Appendix B Sample Visualization
B.1 ImageNet-1k Sample Synthesis
Since ImageNet-1k (Russakovsky et al., 2015) consists of 1,000 classes, to evaluate whether our DDIS can avoid lexical overlapping issues and accurately align with the correct class, we generate images by inverting the ResNet-34 (Sohl-Dickstein et al., 2015) pretrained on ImageNet-1k provided by TorchVision (Paszke et al., 2019), which has a top-1 accuracy of 73.31%. We generate 10,000 images with 512512 using the Stable Diffusion 2.1 and resize them to 224224 for evaluation, matching the typical resolution used during ImageNet-1k training. For comparison, we generate 10,000 images using DeepInversion (DI) (Yin et al., 2020) and PlugInInversion (PII) (Ghiasi et al., 2022), the only Data-Free Image Synthesis (DFIS) studies addressing large-resolution image synthesis, following their official GitHub implementations and proposed their ImageNet-1k parameters. Figure 12 illustrates samples for 30 classes, including those with lexical overlapping issues. The proposed method effectively generates images that precisely match the classes learned by the model through CAT embedding optimization. For example, the class “Yellow Lady’s Slipper” (class 986), which refers to a type of flower, demonstrates how DDIS alleviates lexical overlapping problems to produce accurate class images.
B.2 PACS Sample Synthesis
We are the first to address the synthesis of non-photo-specific domain datasets in DFIS. We focus on the PACS dataset synthesis (Li et al., 2017), a benchmark commonly used in domain generalization, which consists of four domains: Photo, Art Painting, Cartoon, and Sketch.We specifically handle the Art Painting and Cartoon domains, as the Photo domain’s excellence is well-described with the ImageNet-1k experiment, and the Sketch domain represents a failure case in our study. Each domain contains seven classes: dog, elephant, giraffe, guitar, horse, house, and person. We fine-tune the ResNet-34 pre-trained on ImageNet-1k with two domains, excluding the training domain data for inference on the remaining three domains to measure top-1 accuracy. The top-1 accuracies for Art Painting and Cartoon pre-trained ResNet-34 models are 54.73% and 61.78%, respectively. We generate 512512 images using Stable Diffusion 2.1 and resize them to 224224 for evaluation, aligning with the typical PACS training resolution. For comparison, we generate 3,000 images using DeepInversion (DI), PlugInInversion (PII), and our DDIS method. We reproduce the DI and PII utilizing the official GitHub implementations and ImageNet-1k parameters matching the PACS dataset resolution.
Figure 13 shows the results of synthesizing all classes in the PACS Art Painting domain, while Figure 14 displays the results for the PACS Cartoon domain. We demonstrate that our proposed DAG-based diffusion sampling process and CAT embedding optimization align the target dataset’s domain and class information more effectively than the baseline methods.
B.3 Style-Aligned Sample Synthesis
To demonstrate the effectiveness of our methodology across a broader range of domains, we utilize Style-Aligned (Hertz et al., 2024) to create high-quality, domain-specific datasets. We select the domains of manga, caricature, and sketch and synthesized 400 images for each of the seven classes within these domains using Stable Diffusion 2.1. The synthesized datasets are illustrated in the Figure 15. Subsequently, we finetune a classifier using a ResNet-34 pre-trained on ImageNet-1k, following the same approach as with ImageNet-1k and PACS experiments. The top-1 accuracies for manga, caricature, and sketch are 85.70%, 76.80%, and 98.90%, respectively.
Appendix C Additional Experiments
C.1 Efficiency of Domain Alignment Guidance
We demonstrate that in DAG, the gradient of , which aligns synthetic image statistics with the BN layer running statistics, provides stable guidance unaffected by the time step. To avoid confusion, we compare our method with Classifier-Guidance (CG) (Dhariwal & Nichol, 2021), which only provides class guidance at the last time step . First, running statistics within all BN layers are derived from the entire training set, making them more robust and stable than providing guidance based on conditional probability gradients for individual samples. Second, we demonstrate that the BN statistics of generated images remain stable across different time steps. Figure 18 illustrates the layer-wise mean and variance of images generated at every time step. We sample 30 images from the “hero” class in the Style Aligned Manga domain over time steps, passing them through a ResNet-34 pre-trained on the Manga dataset to observe the variation in layer-wise mean over time steps. The layer-wise statistics of noise samples at each time step surprisingly exhibit similar trends from time step 0 to , remaining consistent across different time steps. This result indicates that the model’s internal statistics remain stable regardless of the time step, demonstrating that guidance aligning synthetic image statistics with BN layer running statistics at each time step is highly reliable.
C.2 Comparative Analysis of Class/Domain-Wise Token
To demonstrate the effectiveness of our Class Alignment Token, we compare our CAT with the domain-wise token. The domain-wise token is trained to observe the overall classes, while class-wise tokens are trained to observe each class independently. Figure 10 illustrates the results for the PACS-cartoon domain when using a single token embedding for all classes. We observe that the domain-wise token embedding captures the domain knowledge of the training set but generates images with ignored class-specific features, averaging across all classes. In contrast, our CAT embedding encodes class-specific information, capturing precise features corresponding to each class.
C.3 Zero-shot Image Synthesis
We conduct a zero-shot image synthesis experiment to evaluate whether Stable Diffusion can effectively generate images for unseen classes. First, using Stable Diffusion V3, we select 10 classes that the Stable Diffusion 2.1 model used in our study has never encountered and generate images for those classes with SD V3. Then, we train a ResNet-50 model using the dataset of these 10 classes (see Figure 19). Finally, we utilize Stable Diffusion 2.1 to synthesize the images used for training ResNet-50 with the our proposed DDIS. Surprisingly, as shown in Figure 20, although Stable Diffusion 2.1 struggles with these classes, the CAT embedding can capture the class attributes from the training set and generate images with a distribution similar to that of the original dataset.
C.4 Further Analysis of CAT Embedding Optimization with Frozen SD Networks
To evaluate the effectiveness of our CAT embedding optimization, we conduct ablation studies on three design choices, synthesizing four lexically ambiguous ImageNet-1K classes: Kite (21), Tiger Cat (282), Beach Wagon (436), and Mail Bag (636) like 5 in the main paper. We compare confidence scores and visual quality under these settings. Firstly, we test three fine-tuning configurations for Stable Diffusion (SD): (b) UNet only, (c) text encoder only, and (d) Full fine-tuning, using the Cross-Entropy (CE) loss used for CAT optimization. As shown in Figures 21 and 5, fine-tuned SD produces distorted, low-confidence images, failing to generate class-aligned images. It suggests that the SD fine-tuning in Data-Free Image Synthesis (DFIS) disrupts the prior knowledge within the SD, degrading image quality. Generally, SD is fine-tuned using real data to adapt to specific domains or styles, but in DFIS, the lack of real images leads to unstable training (right side of Figure 21). Moreover, per-class SD fine-tuning leads to individual SD networks, increasing computational cost. In contrast, our method freezes SD and optimizes a single token, preserving SD’s image prior while efficiently generating high-quality, class-aligned images.
Secondly, we explore whether multiple CAT embeddings improve class expressivity by optimizing embeddings for one to five tokens. As shown in Figures 22 and 6, performance drops with more tokens. Lexical ambiguity persists, and confidence scores drop. In data-free settings with simple prompts (e.g., "A {class}"), more tokens amplify the effect of randomly initialized embeddings, hindering desired class-aligned image generation. Therefore, a single token is sufficient and more effective for encoding class information in DFIS.
Lastly, we test optimizing CAT embedding with BatchNorm (BN) loss alongside vanilla CE loss. As shown in Figures 23 and 7, BN loss negatively affects capturing class semantics by inducing synthetic images toward the averaged statistics of the entire dataset, which causes class mixing and reduces separability. Therefore, optimizing CAT embedding with CE loss alone effectively captures class-specific attributes.
Method | C-21 | C-282 | C-436 | C-636 | Avg. Confidence |
---|---|---|---|---|---|
(a) Ours | 94.12 | 65.94 | 85.73 | 67.66 | 78.36 |
(b) SD UNet | 0.77 | 0.67 | 2.47 | 1.53 | 1.36 |
(c) SD Text-Encoder | 0.03 | 21.25 | 17.35 | 2.57 | 10.30 |
(d) SD Full | 0.26 | 0.40 | 0.01 | 0.01 | 0.17 |
Method | C-21 | C-282 | C-436 | C-636 | Avg. Confidence |
---|---|---|---|---|---|
1 Token (Ours) | 94.12 | 65.94 | 85.73 | 67.66 | 78.36 |
2 Tokens | 0.01 | 30.19 | 5.04 | 4.10 | 9.83 |
3 Tokens | 0.01 | 30.15 | 5.08 | 3.19 | 9.60 |
4 Tokens | 0.01 | 30.17 | 5.07 | 3.59 | 9.71 |
5 Tokens | 0.01 | 29.14 | 1.32 | 3.54 | 8.50 |
Method | C-21 | C-282 | C-436 | C-636 | Avg. Confidence |
---|---|---|---|---|---|
Ours | 94.12 | 65.94 | 85.73 | 67.66 | 78.36 |
CAT optim. w/BN Loss | 0.01 | 26.15 | 8.26 | 0.03 | 8.61 |
In conclusion, the above design studies validate that desired concepts or complex information can be encoded by optimizing only token embeddings on frozen SD. Building on this, we adopt the idea to DFIS and demonstrate that a single CAT token effectively captures class semantics while preserving SD’s priors. Across all design choices, our approach consistently outperforms alternatives, highlighting its effectiveness for DFIS.












