Multi-Label Conditional Generation From Pre-Trained Models

被引:0
|
作者
Proszewska, Magdalena [1 ]
Wolczyk, Maciej [1 ]
Zieba, Maciej [2 ,3 ]
Wielopolski, Patryk [4 ]
Maziarka, Lukasz [1 ]
Smieja, Marek [1 ]
机构
[1] Jagiellonian Univ, Fac Math & Comp Sci, PL-31007 Krakow, Poland
[2] Tooploox, PL-53601 Wroclaw, Poland
[3] Wroclaw Univ Sci & Technol, PL-53601 Wroclaw, Poland
[4] Wroclaw Univ Sci & Technol, PL-50370 Wroclaw, Poland
关键词
Training; Computational modeling; Adaptation models; Vectors; Data models; Aerospace electronics; Three-dimensional displays; Conditional generation; deep generative models; GANs; invertible normalizing flows; pre-trained models; VAEs;
D O I
10.1109/TPAMI.2024.3382008
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Although modern generative models achieve excellent quality in a variety of tasks, they often lack the essential ability to generate examples with requested properties, such as the age of the person in the photo or the weight of the generated molecule. To overcome these limitations we propose PluGeN (Plugin Generative Network), a simple yet effective generative technique that can be used as a plugin for pre-trained generative models. The idea behind our approach is to transform the entangled latent representation using a flow-based module into a multi-dimensional space where the values of each attribute are modeled as an independent one-dimensional distribution. In consequence, PluGeN can generate new samples with desired attributes as well as manipulate labeled attributes of existing examples. Due to the disentangling of the latent representation, we are even able to generate samples with rare or unseen combinations of attributes in the dataset, such as a young person with gray hair, men with make-up, or women with beards. In contrast to competitive approaches, PluGeN can be trained on partially labeled data. We combined PluGeN with GAN and VAE models and applied it to conditional generation and manipulation of images, chemical molecule modeling and 3D point clouds generation.
引用
收藏
页码:6185 / 6198
页数:14
相关论文
共 50 条
  • [31] Automatic Detection of Liver Cancer Using Hybrid Pre-Trained Models
    Othman, Esam
    Mahmoud, Muhammad
    Dhahri, Habib
    Abdulkader, Hatem
    Mahmood, Awais
    Ibrahim, Mina
    SENSORS, 2022, 22 (14)
  • [32] Efficient Integrated Features Based on Pre-trained Models for Speaker Verification
    Li, Yishuang
    Guan, Wenhao
    Huang, Hukai
    Miao, Shiyu
    Su, Qi
    Li, Lin
    Hong, Qingyang
    INTERSPEECH 2024, 2024, : 2140 - 2144
  • [33] Diet Code Is Healthy: Simplifying Programs for Pre-trained Models of Code
    Zhang, Zhaowei
    Zhang, Hongyu
    Shen, Beijun
    Gu, Xiaodong
    PROCEEDINGS OF THE 30TH ACM JOINT MEETING EUROPEAN SOFTWARE ENGINEERING CONFERENCE AND SYMPOSIUM ON THE FOUNDATIONS OF SOFTWARE ENGINEERING, ESEC/FSE 2022, 2022, : 1073 - 1084
  • [34] SecretGen: Privacy Recovery on Pre-trained Models via Distribution Discrimination
    Yuan, Zhuowen
    Wu, Fan
    Long, Yunhui
    Xiao, Chaowei
    Li, Bo
    COMPUTER VISION - ECCV 2022, PT V, 2022, 13665 : 139 - 155
  • [35] An analysis of pre-trained stable diffusion models through a semantic lens
    Bonechi, Simone
    Andreini, Paolo
    Corradini, Barbara Toniella
    Scarselli, Franco
    NEUROCOMPUTING, 2025, 614
  • [36] Aspect Based Sentiment Analysis using French Pre-Trained Models
    Essebbar, Abderrahman
    Kane, Bamba
    Guinaudeau, Ophelie
    Chiesa, Valeria
    Quenel, Ilhem
    Chau, Stephane
    ICAART: PROCEEDINGS OF THE 13TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE - VOL 1, 2021, : 519 - 525
  • [37] Multi-Source Log Parsing With Pre-Trained Domain Classifier
    Liu, Yilun
    Tao, Shimin
    Meng, Weibin
    Wang, Jingyu
    Yang, Hao
    Jiang, Yanfei
    IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT, 2024, 21 (03): : 2651 - 2663
  • [38] A Joint Label-Enhanced Representation Based on Pre-trained Model for Charge Prediction
    Dan, Jingpei
    Liao, Xiaoshuang
    Xu, Lanlin
    Hu, Weixuan
    Zhang, Tianyuan
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2022, PT I, 2022, 13551 : 694 - 705
  • [39] Electric Power Audit Text Classification With Multi-Grained Pre-Trained Language Model
    Meng, Qinglin
    Song, Yan
    Mu, Jian
    Lv, Yuanxu
    Yang, Jiachen
    Xu, Liang
    Zhao, Jin
    Ma, Junwei
    Yao, Wei
    Wang, Rui
    Xiao, Maoxiang
    Meng, Qingyu
    IEEE ACCESS, 2023, 11 : 13510 - 13518
  • [40] Multi-Label Classification With Label-Specific Feature Generation: A Wrapped Approach
    Yu, Ze-Bang
    Zhang, Min-Ling
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (09) : 5199 - 5210