Implicit texture mapping for multi-view video synthesis

Mohamed I Lakhal (Huawei),* Oswald Lanz (Free University of Bozen-Bolzano), ANDREA CAVALLARO (Queen Mary University of London, UK)
The 33rd British Machine Vision Conference


Multi-view video synthesis generates the scene dynamics from a viewpoint given a source view and one or more modalities of a targeted view. In this paper, we frame video synthesis as a feature learning problem and solve it as target-view motion synthesis with spatial refinement. Specifically, we propose a motion synthesis network with a novel recurrent neural layer that learns the spatio-temporal representation of the target-view. Next, a refinement network corrects the generated coarse texture by learning the residual (\textit{i.e.}~high-frequency textures) through a UNet generator. Experimental results show visual quality enhancement of the proposed pipeline over state-of-the-art methods.



author    = {Mohamed I Lakhal and Oswald Lanz and ANDREA CAVALLARO},
title     = {Implicit texture mapping for multi-view video synthesis},
booktitle = {33rd British Machine Vision Conference 2022, {BMVC} 2022, London, UK, November 21-24, 2022},
publisher = {{BMVA} Press},
year      = {2022},
url       = {}

Copyright © 2022 The British Machine Vision Association and Society for Pattern Recognition
The British Machine Vision Conference is organised by The British Machine Vision Association and Society for Pattern Recognition. The Association is a Company limited by guarantee, No.2543446, and a non-profit-making body, registered in England and Wales as Charity No.1002307 (Registered Office: Dept. of Computer Science, Durham University, South Road, Durham, DH1 3LE, UK).

Imprint | Data Protection