-
Notifications
You must be signed in to change notification settings - Fork 15
Description
Hi, @weijiawu
It is really a great work to leverage the generative model to construct synthetic data for downstream tasks. At the same time, my task is to construct synthetic data in x-ray security image domain that is pseudo-colored and different from natural image domain, and I have some questions to ask for your help.
My first question is that is it feasible to fine-tune DatasetDM on x-ray dataset? Since x-ray images belong to a diiferent domain, I cannot directly apply DatasetDM. Besides, I have also made some experiments like training conditional diffusion models such as GLIGEN to generate x-ray images, but the generated results are not so good. Problems like object missing, repeating and occlusion issues have emerged, and I am not sure whether it has to do with my dataset size( training set consists of 27,708 images).
My second question is that when we say 'Synthetic Data for Perception Tasks', what is the role of synthetic data? To put it in another way, how can we judge the quality of synthetic data is perfectly suitable for our downstream tasks? For example, my downstream task is to detect all prohibited items in an x-ray security image. Sometimes, the generated images may be visually acceptable for humans but not for models, and different visual cues may be of different importance to different perception models. Is there a way for us to generate the most important synthetic data for downstream models to maximally learn from it and perform well on test sets?