TY - GEN
T1 - Realtime novel view synthesis with eigen-texture regression
AU - Nakashima, Yuta
AU - Okura, Fumio
AU - Kawai, Norihiko
AU - Kawasaki, Hiroshi
AU - Blanco, Ambrosio
AU - Ikeuchi, Katsushi
N1 - Publisher Copyright:
© 2017. The copyright of this document resides with its authors.
PY - 2017
Y1 - 2017
N2 - Realtime novel view synthesis, which generates a novel view of a real object or scene in realtime, enjoys a wide range of applications including augmented reality, telepresence, and immersive telecommunication. Image-based rendering (IBR) with rough geometry can be done using only an off-the-shelf camera and thus can be used by many users. However, IBR from images in the wild (e.g., lighting condition changes or the scene contains objects with specular surfaces) has been a tough problem due to color discontinuity; IBR with rough geometry picks up appropriate images for a given viewpoint, but the image used for a rendering unit (a face or pixel) switches when the viewpoint moves, which may cause noticeable changes in color. We use the eigen-texture technique, which represents images for a certain face using a point in the eigenspace. We propose to regress a new point in this space, which moves smoothly, given a viewpoint so that we can generate an image whose color smoothly changes according to the point. Our regressor is based on a neural network with a single hidden layer and hyperbolic tangent nonlinearity. We demonstrate the advantages of our IBR approach using our own datasets as well as publicly available datasets for comparison.
AB - Realtime novel view synthesis, which generates a novel view of a real object or scene in realtime, enjoys a wide range of applications including augmented reality, telepresence, and immersive telecommunication. Image-based rendering (IBR) with rough geometry can be done using only an off-the-shelf camera and thus can be used by many users. However, IBR from images in the wild (e.g., lighting condition changes or the scene contains objects with specular surfaces) has been a tough problem due to color discontinuity; IBR with rough geometry picks up appropriate images for a given viewpoint, but the image used for a rendering unit (a face or pixel) switches when the viewpoint moves, which may cause noticeable changes in color. We use the eigen-texture technique, which represents images for a certain face using a point in the eigenspace. We propose to regress a new point in this space, which moves smoothly, given a viewpoint so that we can generate an image whose color smoothly changes according to the point. Our regressor is based on a neural network with a single hidden layer and hyperbolic tangent nonlinearity. We demonstrate the advantages of our IBR approach using our own datasets as well as publicly available datasets for comparison.
UR - http://www.scopus.com/inward/record.url?scp=85088201917&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85088201917&partnerID=8YFLogxK
U2 - 10.5244/c.31.83
DO - 10.5244/c.31.83
M3 - Conference contribution
AN - SCOPUS:85088201917
T3 - British Machine Vision Conference 2017, BMVC 2017
BT - British Machine Vision Conference 2017, BMVC 2017
PB - BMVA Press
T2 - 28th British Machine Vision Conference, BMVC 2017
Y2 - 4 September 2017 through 7 September 2017
ER -