研究目的
To address the problem of heavy training burdens and large model parameters in unsupervised image-to-image translation for multiple domains (n>2) by proposing a unified framework that reduces parameters and training time while maintaining or improving translation quality.
研究成果
The Domain-Bank framework effectively reduces training time and model parameters for multi-domain image translation, achieving comparable or better results than state-of-the-art methods in tasks like face attribute and painting style translation, and superior performance in domain adaptation on digit datasets, demonstrating its efficiency and effectiveness.
研究不足
The framework assumes a universal shared-latent space, which may not hold for all domains; cycle-consistency constraint is not always necessary (e.g., removed for digit tasks), indicating potential optimization needs. The method's scalability to very large n or complex domains is not fully explored.
1:Experimental Design and Method Selection:
The study uses a novel framework called Domain-Bank, based on variational autoencoders (VAEs) and generative adversarial networks (GANs), with a shared-latent space assumption to enable translation among multiple domains in a single training process.
2:Sample Selection and Data Sources:
Datasets include CelebA for face attribute translation (e.g., hair color domains), landscape photographs from Flickr and WikiArt for painting style translation (domains for Cezanne, Monet, Van Gogh, Ukiyo-e), and digit datasets (SVHN, MNIST, USPS) for domain adaptation.
3:List of Experimental Equipment and Materials:
No specific hardware or software mentioned; the method is implemented using neural networks.
4:Experimental Procedures and Operational Workflow:
The framework involves training domain-specific encoders, decoders, and discriminators with shared layers, using adversarial and VAE losses, and cycle-consistency constraints. Training is done with an alternative strategy optimizing a mini-max problem.
5:Data Analysis Methods:
Qualitative analysis through visualization of translated images and quantitative analysis via classification accuracies for domain adaptation tasks, comparing with state-of-the-art methods like UNIT and CycleGAN.
独家科研数据包,助您复现前沿成果,加速创新突破
获取完整内容