Supplementary Material for Video Propagation Networks - Varun … · 2020-06-12 · Supplementary...

7
Supplementary Material for Video Propagation Networks Varun Jampani 1 , Raghudeep Gadde 1,2 and Peter V. Gehler 1,2 1 Max Planck Institute for Intelligent Systems, T ¨ ubingen, Germany 2 Bernstein Center for Computational Neuroscience, T ¨ ubingen, Germany {varun.jampani,raghudeep.gadde,peter.gehler}@tuebingen.mpg.de 1. Parameters and Additional Results In this supplementary, we present experiment protocols and additional qualitative results for experiments on video object segmentation, semantic video segmentation and video color propagation. Table 1 shows the feature scales and other parame- ters used in different experiments. Figures 1, 2 show some qualitative results on video object segmentation with some failure cases in Fig. 3. Figure 4 shows some qualitative results on semantic video segmentation and Fig. 5 shows results on video color propagation. Experiment Feature Type Feature Scale-1, Λa Feature Scale-2, Λ b α Input Frames Loss Type Video Object Segmentation (x, y, Y, Cb, Cr, t) (0.02,0.02,0.07,0.4,0.4,0.01) (0.03,0.03,0.09,0.5,0.5,0.2) 0.5 9 Logistic Semantic Video Segmentation with CNN1 [5]-NoFlow (x, y, R, G, B, t) (0.08,0.08,0.2,0.2,0.2,0.04) (0.11,0.11,0.2,0.2,0.2,0.04) 0.5 3 Logistic with CNN1 [5]-Flow (x+ux,y+uy, R, G, B, t) (0.11,0.11,0.14,0.14,0.14,0.03) (0.08,0.08,0.12,0.12,0.12,0.01) 0.65 3 Logistic with CNN2 [3]-Flow (x+ux,y+uy, R, G, B, t) (0.08,0.08,0.2,0.2,0.2,0.04) (0.09,0.09,0.25,0.25,0.25,0.03) 0.5 4 Logistic Video Color Propagation (x,y,I,t) (0.04,0.04,0.2,0.04) No second kernel 1 4 MSE Table 1. Experiment Protocols. Experiment protocols for the different experiments presented in this work. Feature Types: Feature spaces used for the bilateral convolutions, with position (x, y) and color (R, G, B or Y,Cb,Cr) features [0, 255]. ux, uy denotes optical flow with respect to the present frame and I denotes grayscale intensity. Feature Scales (Λa, Λ b ): Validated scales for the features used. α: Exponential time decay for the input frames. Input Frames: Number of input frames for VPN. Loss Type: Type of loss used for back-propagation. “MSE” corresponds to Euclidean mean squared error loss and “Logistic” corresponds to multinomial logistic loss. 1

Transcript of Supplementary Material for Video Propagation Networks - Varun … · 2020-06-12 · Supplementary...

Page 1: Supplementary Material for Video Propagation Networks - Varun … · 2020-06-12 · Supplementary Material for Video Propagation Networks Varun Jampani 1, Raghudeep Gadde;2and Peter

Supplementary Material for Video Propagation Networks

Varun Jampani1, Raghudeep Gadde1,2 and Peter V. Gehler1,21Max Planck Institute for Intelligent Systems, Tubingen, Germany

2Bernstein Center for Computational Neuroscience, Tubingen, Germany{varun.jampani,raghudeep.gadde,peter.gehler}@tuebingen.mpg.de

1. Parameters and Additional ResultsIn this supplementary, we present experiment protocols and additional qualitative results for experiments on video object

segmentation, semantic video segmentation and video color propagation. Table 1 shows the feature scales and other parame-ters used in different experiments. Figures 1, 2 show some qualitative results on video object segmentation with some failurecases in Fig. 3. Figure 4 shows some qualitative results on semantic video segmentation and Fig. 5 shows results on videocolor propagation.

Experiment Feature Type Feature Scale-1, Λa Feature Scale-2, Λb αInput

Frames Loss Type

Video Object Segmentation (x, y, Y, Cb, Cr, t) (0.02,0.02,0.07,0.4,0.4,0.01) (0.03,0.03,0.09,0.5,0.5,0.2) 0.5 9 Logistic

Semantic Video Segmentationwith CNN1 [5]-NoFlow (x, y, R,G,B, t) (0.08,0.08,0.2,0.2,0.2,0.04) (0.11,0.11,0.2,0.2,0.2,0.04) 0.5 3 Logistic

with CNN1 [5]-Flow (x+ux, y+uy, R,G,B, t) (0.11,0.11,0.14,0.14,0.14,0.03) (0.08,0.08,0.12,0.12,0.12,0.01) 0.65 3 Logistic

with CNN2 [3]-Flow (x+ux, y+uy, R,G,B, t) (0.08,0.08,0.2,0.2,0.2,0.04) (0.09,0.09,0.25,0.25,0.25,0.03) 0.5 4 Logistic

Video Color Propagation (x, y, I, t) (0.04,0.04,0.2,0.04) No second kernel 1 4 MSE

Table 1. Experiment Protocols. Experiment protocols for the different experiments presented in this work. Feature Types: Feature spacesused for the bilateral convolutions, with position (x, y) and color (R,G,B or Y,Cb, Cr) features ∈ [0, 255]. ux, uy denotes opticalflow with respect to the present frame and I denotes grayscale intensity. Feature Scales (Λa,Λb): Validated scales for the features used.α: Exponential time decay for the input frames. Input Frames: Number of input frames for VPN. Loss Type: Type of loss used forback-propagation. “MSE” corresponds to Euclidean mean squared error loss and “Logistic” corresponds to multinomial logistic loss.

1

Page 2: Supplementary Material for Video Propagation Networks - Varun … · 2020-06-12 · Supplementary Material for Video Propagation Networks Varun Jampani 1, Raghudeep Gadde;2and Peter

Frame 5 Frame 10 Frame 15 Frame 20 Frame 25 Frame 30

Input

Vid

eo

GT

BV

SO

FL

VP

NV

PN

-DLab

Figure 1. Video Object Segmentation. Shown are the different frames in example videos with the corresponding ground truth (GT) masks,predictions from BVS [2], OFL [4], VPN (VPN-Stage2) and VPN-DLab (VPN-DeepLab) models.

Page 3: Supplementary Material for Video Propagation Networks - Varun … · 2020-06-12 · Supplementary Material for Video Propagation Networks Varun Jampani 1, Raghudeep Gadde;2and Peter

Frame 5 Frame 15 Frame 30 Frame 50

Input

Vid

eo

GT

BV

SO

FL

VP

NV

PN

-DLab

Frame 5 Frame 10 Frame 20 Frame 30

Input

Vid

eo

GT

BV

SO

FL

VP

NV

PN

-DLab

Figure 2. Video Object Segmentation. Shown are the different frames in example videos with the corresponding ground truth (GT) masks,predictions from BVS [2], OFL [4], VPN (VPN-Stage2) and VPN-DLab (VPN-DeepLab) models.

Page 4: Supplementary Material for Video Propagation Networks - Varun … · 2020-06-12 · Supplementary Material for Video Propagation Networks Varun Jampani 1, Raghudeep Gadde;2and Peter

Frame 4 Frame 14 Frame 24 Frame 36

Input

Vid

eo

GT

BV

SO

FL

VP

NV

PN

-DLab

Frame 5 Frame 15 Frame 30 Frame 50

Input

Vid

eo

GT

BV

SO

FL

VP

NV

PN

-DLab

Figure 3. Failure Cases for Video Object Segmentation. Shown are the different frames in example videos with the corresponding groundtruth (GT) masks, predictions from BVS [2], OFL [4], VPN (VPN-Stage2) and VPN-DLab (VPN-DeepLab) models.

Page 5: Supplementary Material for Video Propagation Networks - Varun … · 2020-06-12 · Supplementary Material for Video Propagation Networks Varun Jampani 1, Raghudeep Gadde;2and Peter

Input GT CNN +VPN(Ours)

Figure 4. Semantic Video Segmentation. Input video frames and the corresponding ground truth (GT) segmentation together with thepredictions of CNN [5] and with VPN-Flow.

Page 6: Supplementary Material for Video Propagation Networks - Varun … · 2020-06-12 · Supplementary Material for Video Propagation Networks Varun Jampani 1, Raghudeep Gadde;2and Peter

Frame 2 Frame 7 Frame 13 Frame 19

Input

Vid

eo

GT

-Colo

rLevin

et

al.

VP

N(O

urs)

Frame 2 Frame 7 Frame 13 Frame 19

Input

Vid

eo

GT

-Colo

rLevin

et

al.

VP

N(O

urs)

Figure 5. Video Color Propagation. Input grayscale video frames and corresponding ground-truth (GT) color images together with colorpredictions of Levin et al. [1] and VPN-Stage1 models.

Page 7: Supplementary Material for Video Propagation Networks - Varun … · 2020-06-12 · Supplementary Material for Video Propagation Networks Varun Jampani 1, Raghudeep Gadde;2and Peter

References[1] A. Levin, D. Lischinski, and Y. Weiss. Colorization using optimization. ACM Transactions on Graphics (ToG), 23(3):689–694, 2004.

6[2] N. Marki, F. Perazzi, O. Wang, and A. Sorkine-Hornung. Bilateral space video segmentation. In Computer Vision and Pattern

Recognition, IEEE Conference on, pages 743–751, 2016. 2, 3, 4[3] S. R. Richter, V. Vineet, S. Roth, and V. Koltun. Playing for data: Ground truth from computer games. In European Conference on

Computer Vision, pages 102–118. Springer, 2016. 1[4] Y.-H. Tsai, M.-H. Yang, and M. J. Black. Video segmentation via object flow. In Computer Vision and Pattern Recognition, IEEE

Conference on, 2016. 2, 3, 4[5] F. Yu and V. Koltun. Multi-scale context aggregation by dilated convolutions. International Conference on Learning Representations,

2016. 1, 5