TransMatting: Enhancing Transparent Objects Matting with Transformers
Huanqia Cai
|
Fanglei Xue
|
Lele Xu
|
Lili Guo
|
Zhifeng Li
|
Wei Liu
|
Abstract
Image matting refers to predicting the alpha values of unknown foreground areas from natural images. Prior methods have focused on propagating alpha values from known to unknown regions. However, not all natural images have a specifically known foreground. Images of transparent objects, like glass, smoke, web, etc., have less or no known foreground. In this paper, we propose a Transformer-based network, TransMatting, to model transparent objects with a big receptive field. Specifically, we redesign the trimap as three learnable tri-tokens for introducing advanced semantic features into the self-attention mechanism. A small convolutional network is proposed to utilize the global
feature and non-background mask to guide the multi-scale feature propagation from encoder to decoder for maintaining the contexture of transparent objects. In addition, we create a high-resolution matting dataset of transparent objects with small known foreground areas. Experiments on several matting benchmarks demonstrate the superiority of our proposed method over the current state-of-the-art methods.
|
|
The structure of our TransMatting
Code of tri-token
Proposed Dataset: Transparent-460
Paper
Acknowledgements
The work was supported by the National Natural Science Foundation of Chinese under Grant 61901454.
|