Table of Links
Abstract and 1. Introduction
-
Related Works
-
MaGGIe
3.1. Efficient Masked Guided Instance Matting
3.2. Feature-Matte Temporal Consistency
-
Instance Matting Datasets
4.1. Image Instance Matting and 4.2. Video Instance Matting
-
Experiments
5.1. Pre-training on image data
5.2. Training on video data
-
Discussion and References
Supplementary Material
-
Architecture details
-
Image matting
8.1. Dataset generation and preparation
8.2. Training details
8.3. Quantitative details
8.4. More qualitative results on natural images
-
Video matting
9.1. Dataset generation
9.2. Training details
9.3. Quantitative details
9.4. More qualitative results
4. Instance Matting Datasets
This section outlines the datasets used in our experiments. With the lack of public datasets for the instance matting task, we synthesized training data from existing public instance-agnostic sources. Our evaluation combines synthetic and natural sets to assess the model’s robustness and generalization.
:::info
Authors:
(1) Chuong Huynh, University of Maryland, College Park ([email protected]);
(2) Seoung Wug Oh, Adobe Research (seoh,[email protected]);
(3) Abhinav Shrivastava, University of Maryland, College Park ([email protected]);
(4) Joon-Young Lee, Adobe Research ([email protected]).
:::
:::info
This paper is available on arxiv under CC by 4.0 Deed (Attribution 4.0 International) license.
:::
