Jiachen Li, Jitesh Jain, Humphrey Shi
[Project page
]
[ArXiv
]
[Pdf
]
[Video
]
[Demo
]
2023/07/17
: Added MAM checkpoints based on SAM ViT-L and SAM ViT-H.2023/06/28
: Getting Started is updated with training and evaluation instructions.2023/06/09
: HuggingFace Demo is released.2023/06/08
: Arxiv Preprint is released.2023/06/06
: Project Page and Demo Video are released.In this paper, we propose the Matting Anything Model (MAM), an efficient and versatile framework for estimating the alpha matte of any instance in an image with flexible and interactive visual or linguistic user prompt guidance. MAM offers several significant advantages over previous specialized image matting networks: (i) MAM is capable of dealing with various types of image matting, including semantic, instance, and referring image matting with only a single model; (ii) MAM leverages the feature maps from the Segment Anything Model (SAM) and adopts a lightweight Mask-to-Matte (M2M) module to predict the alpha matte through iterative refinement, which has only 2.7 million trainable parameters. (iii) By incorporating SAM, MAM simplifies the user intervention required for the interactive use of image matting from the trimap to the box, point, or text prompt. We evaluate the performance of MAM on various image matting benchmarks, and the experimental results demonstrate that MAM achieves comparable performance to the state-of-the-art specialized image matting models under different metrics on each benchmark. Overall, MAM shows superior generalization ability and can effectively handle various image matting tasks with fewer parameters, making it a practical solution for unified image matting.
We provide visualizations of the alpha matte predictions from SAM and MAM. Notably, we emphasize the differences in the red boxes. The visualizations demonstrate that MAM achieves improved predictions in the transition areas even without the trimap guidance. Additionally, MAM effectively addresses some of the holes present in the mask predictions generated by SAM. These visual comparisons highlight the superior performance of MAM in refining and enhancing the quality of alpha matte predictions.
Please refer to Installation Instructions for complete installation instructions for MAM.
Please refer to Getting Started for dataset preparation, training, and inference details of MAM.
@article{li2023matting,
title={Matting Anything},
author={Jiachen Li and Jitesh Jain and Humphrey Shi},
journal={arXiv: 2306.05399},
year={2023}
}
We thank the authors of SAM, Grounded-SAM, MGMatting, and InstMatt for releasing the codebases.
此处可能存在不合适展示的内容,页面不予展示。您可通过相关编辑功能自查并修改。
如您确认内容无涉及 不当用语 / 纯广告导流 / 暴力 / 低俗色情 / 侵权 / 盗版 / 虚假 / 无价值内容或违法国家有关法律法规的内容,可点击提交进行申诉,我们将尽快为您处理。