<?xml version="1.0" encoding="UTF-8"?><rdf:RDF xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:dcq="http://purl.org/dc/terms/"><records count="1" morepages="false" start="1" end="1"><record rownumber="1"><dc:product_type>Journal Article</dc:product_type><dc:title>MaskDiff: Modeling Mask Distribution with Diffusion Probabilistic Model for Few-Shot Instance Segmentation</dc:title><dc:creator>Le, Minh-Quan; Nguyen, Tam V; Le, Trung-Nghia; Do, Thanh-Toan; Do, Minh N; Tran, Minh-Triet</dc:creator><dc:corporate_author/><dc:editor/><dc:description>Few-shot instance segmentation extends the few-shot learning paradigm to the instance segmentation task, which tries to segment instance objects from a query image with a few annotated examples of novel categories. Conventional approaches have attempted to address the task via prototype learning, known as point estimation. However, this mechanism depends on prototypes (e.g. mean of K-shot) for prediction, leading to performance instability. To overcome the disadvantage of the point estimation mechanism, we propose a novel approach, dubbed MaskDiff, which models the underlying conditional distribution of a binary mask, which is conditioned on an object region and K-shot information. Inspired by augmentation approaches that perturb data with Gaussian noise for populating low data density regions, we model the mask distribution with a diffusion probabilistic model. We also propose to utilize classifier-free guided mask sampling to integrate category information into the binary mask generation process. Without bells and whistles, our proposed method consistently outperforms state-of-the-art methods on both base and novel classes of the COCO dataset while simultaneously being more stable than existing methods. The source code is available at: https://github.com/minhquanlecs/MaskDiff.</dc:description><dc:publisher>AAAI</dc:publisher><dc:date>2024-03-25</dc:date><dc:nsf_par_id>10521808</dc:nsf_par_id><dc:journal_name>Proceedings of the AAAI Conference on Artificial Intelligence</dc:journal_name><dc:journal_volume>38</dc:journal_volume><dc:journal_issue>3</dc:journal_issue><dc:page_range_or_elocation>2874 to 2881</dc:page_range_or_elocation><dc:issn>2159-5399</dc:issn><dc:isbn/><dc:doi>https://doi.org/10.1609/aaai.v38i3.28068</dc:doi><dcq:identifierAwardId>2025234</dcq:identifierAwardId><dc:subject>Segmentation</dc:subject><dc:subject>Deep Generative Models &amp; Autoencoders</dc:subject><dc:version_number/><dc:location/><dc:rights/><dc:institution/><dc:sponsoring_org>National Science Foundation</dc:sponsoring_org></record></records></rdf:RDF>