作者
Jifei Song, Kaiyue Pang, Yi-Zhe Song, Tao Xiang, Timothy M Hospedales
发表日期
2018
研讨会论文
Proceedings of the IEEE conference on computer vision and pattern recognition
页码范围
801-810
简介
To see is to sketch--free-hand sketching naturally builds ties between human and machine vision. In this paper, we present a novel approach for translating an object photo to a sketch, mimicking the human sketching process. This is an extremely challenging task because the photo and sketch domains differ significantly. Furthermore, human sketches exhibit various levels of sophistication and abstraction even when depicting the same object instance in a reference photo. This means that even if photo-sketch pairs are available, they only provide weak supervision signal to learn a translation model. Compared with existing supervised approaches that solve the problem of D (E (photo))-> sketch, where E (·) and D (·) denote encoder and decoder respectively, we take advantage of the inverse problem (eg, D (E (sketch))-> photo), and combine with the unsupervised learning tasks of within-domain reconstruction, all within a multi-task learning framework. Compared with existing unsupervised approaches based on cycle consistency (ie, D (E (D (E (photo))))-> photo), we introduce a shortcut consistency enforced at the encoder bottleneck (eg, D (E (photo))-> photo) to exploit the additional self-supervision. Both qualitative and quantitative results show that the proposed model is superior to a number of state-of-the-art alternatives. We also show that the synthetic sketches can be used to train a better fine-grained sketch-based image retrieval (FG-SBIR) model, effectively alleviating the problem of sketch data scarcity.
引用总数
20182019202020212022202320244131829282512
学术搜索中的文章
J Song, K Pang, YZ Song, T Xiang, TM Hospedales - Proceedings of the IEEE conference on computer …, 2018