TY - GEN
T1 - Adaptive Online Learning for Video Object Segmentation
AU - Wei, Li
AU - Xu, Chunyan
AU - Zhang, Tong
N1 - Publisher Copyright:
© 2019, Springer Nature Switzerland AG.
PY - 2019
Y1 - 2019
N2 - In this work, we address the problem of video object segmentation (VOS), namely segmenting specific objects throughout a video sequence when given only an annotated first frame. Previous VOS methods based on deep neural networks often solves this problem by fine-tuning the segmentation model in the first frame of the test video sequence, which is time-consuming and can not be well adapted to the current target video. In this paper, we proposed the adaptive online learning for video object segmentation (AOL-VOS), which adaptively optimizes the network parameters and hyperparameters of segmentation model for better predicting the segmentation results. Specifically, we first pre-train the segmentation model with the static video frames and then learn the effective adaptation strategy on the training set by optimizing both network parameters and hyperparameters. In the testing process, we learn how to online adapt the learned segmentation model to the specific testing video sequence and the corresponding future video frames, where the confidence patterns is employed to constrain/guide the implementation of adaptive learning process by fusing both object appearance and motion cue information. Comprehensive evaluations on Davis 16 and SegTrack V2 datasets well demonstrate the significant superiority of our proposed AOL-VOS over other state-of-the-arts for video object segmentation task.
AB - In this work, we address the problem of video object segmentation (VOS), namely segmenting specific objects throughout a video sequence when given only an annotated first frame. Previous VOS methods based on deep neural networks often solves this problem by fine-tuning the segmentation model in the first frame of the test video sequence, which is time-consuming and can not be well adapted to the current target video. In this paper, we proposed the adaptive online learning for video object segmentation (AOL-VOS), which adaptively optimizes the network parameters and hyperparameters of segmentation model for better predicting the segmentation results. Specifically, we first pre-train the segmentation model with the static video frames and then learn the effective adaptation strategy on the training set by optimizing both network parameters and hyperparameters. In the testing process, we learn how to online adapt the learned segmentation model to the specific testing video sequence and the corresponding future video frames, where the confidence patterns is employed to constrain/guide the implementation of adaptive learning process by fusing both object appearance and motion cue information. Comprehensive evaluations on Davis 16 and SegTrack V2 datasets well demonstrate the significant superiority of our proposed AOL-VOS over other state-of-the-arts for video object segmentation task.
KW - Adaptation
KW - Online-learning
KW - Video object segmentation
UR - http://www.scopus.com/inward/record.url?scp=85077116951&partnerID=8YFLogxK
U2 - 10.1007/978-3-030-36189-1_2
DO - 10.1007/978-3-030-36189-1_2
M3 - Conference contribution
AN - SCOPUS:85077116951
SN - 9783030361884
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 22
EP - 34
BT - Intelligence Science and Big Data Engineering. Visual Data Engineering - 9th International Conference, IScIDE 2019, Proceedings, Part 1
A2 - Cui, Zhen
A2 - Pan, Jinshan
A2 - Zhang, Shanshan
A2 - Xiao, Liang
A2 - Yang, Jian
PB - Springer
T2 - 9th International Conference on Intelligence Science and Big Data Engineering, IScIDE 2019
Y2 - 17 October 2019 through 20 October 2019
ER -