2022
DOI: 10.48550/arxiv.2207.07115
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

XMem: Long-Term Video Object Segmentation with an Atkinson-Shiffrin Memory Model

Abstract: We present XMem, a video object segmentation architecture for long videos with unified feature memory stores inspired by the Atkinson-Shiffrin memory model. Prior work on video object segmentation typically only uses one type of feature memory. For videos longer than a minute, a single feature memory model tightly links memory consumption and accuracy. In contrast, following the Atkinson-Shiffrin model, we develop an architecture that incorporates multiple independent yet deeply-connected feature memory stores… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 54 publications
(131 reference statements)
0
1
0
Order By: Relevance
“…It explores and verifies the way of external scale connection to further improve the transmission quality and performance of display images. Reference [4] proposes a structural framework called multi-scale progressive fusion neural network connection to eliminate the noise formed in dual-display images. For the noise formed by the similarity caused by different nodes, the global texture effect is captured by means of recursive calculation.…”
Section: Introductionmentioning
confidence: 99%
“…It explores and verifies the way of external scale connection to further improve the transmission quality and performance of display images. Reference [4] proposes a structural framework called multi-scale progressive fusion neural network connection to eliminate the noise formed in dual-display images. For the noise formed by the similarity caused by different nodes, the global texture effect is captured by means of recursive calculation.…”
Section: Introductionmentioning
confidence: 99%