UM  > Faculty of Science and Technology
Residential Collegefalse
Status已發表Published
T2I-Adapter: Learning Adapters to Dig Out More Controllable Ability for Text-to-Image Diffusion Models
Mou, Chong1,2; Wang, Xintao2; Xie, Liangbin2,3,4; Wu, Yanze2; Zhang, Jian1; Qi, Zhongang2; Shan, Ying2
2024-03-25
Conference Name38th AAAI Conference on Artificial Intelligence, AAAI 2024
Source PublicationProceedings of the AAAI Conference on Artificial Intelligence
Volume38
Issue5
Pages4296-4304
Conference Date20 February 2024through 27 February 2024
Conference PlaceVancouver
Abstract

The incredible generative ability of large-scale text-to-image (T2I) models has demonstrated strong power of learning complex structures and meaningful semantics. However, relying solely on text prompts cannot fully take advantage of the knowledge learned by the model, especially when flexible and accurate controlling (e.g., structure and color) is needed. In this paper, we aim to "dig out"the capabilities that T2I models have implicitly learned, and then explicitly use them to control the generation more granularly. Specifically, we propose to learn low-cost T2I-Adapters to align internal knowledge in T2I models with external control signals, while freezing the original large T2I models. In this way, we can train various adapters according to different conditions, achieving rich control and editing effects in the color and structure of the generation results. Further, the proposed T2IAdapters have attractive properties of practical value, such as composability and generalization ability. Extensive experiments demonstrate that our T2I-Adapter has promising generation quality and a wide range of applications.

KeywordCv: Computational Photography Image & Video Synthesis Cv: Multi-modal Vision
DOI10.1609/aaai.v38i5.28226
URLView the original
Language英語English
Scopus ID2-s2.0-85189556876
Fulltext Access
Citation statistics
Document TypeConference paper
CollectionFaculty of Science and Technology
THE STATE KEY LABORATORY OF INTERNET OF THINGS FOR SMART CITY (UNIVERSITY OF MACAU)
DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE
Corresponding AuthorXie, Liangbin
Affiliation1.Peking University Shenzhen Graduate School, China
2.ARC Lab, Tencent PCG, China
3.University of Macau, Macao
4.Shenzhen Institute of Advanced Technology, China
Corresponding Author AffilicationUniversity of Macau
Recommended Citation
GB/T 7714
Mou, Chong,Wang, Xintao,Xie, Liangbin,et al. T2I-Adapter: Learning Adapters to Dig Out More Controllable Ability for Text-to-Image Diffusion Models[C], 2024, 4296-4304.
APA Mou, Chong., Wang, Xintao., Xie, Liangbin., Wu, Yanze., Zhang, Jian., Qi, Zhongang., & Shan, Ying (2024). T2I-Adapter: Learning Adapters to Dig Out More Controllable Ability for Text-to-Image Diffusion Models. Proceedings of the AAAI Conference on Artificial Intelligence, 38(5), 4296-4304.
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Mou, Chong]'s Articles
[Wang, Xintao]'s Articles
[Xie, Liangbin]'s Articles
Baidu academic
Similar articles in Baidu academic
[Mou, Chong]'s Articles
[Wang, Xintao]'s Articles
[Xie, Liangbin]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Mou, Chong]'s Articles
[Wang, Xintao]'s Articles
[Xie, Liangbin]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.