Light field (LF) imaging captures multiple intensities and directions of light per pixel during acquisition in a 3D scene, so that novel images of different viewpoints or focal points can be synthesized. However, transmitting all LF data before viewer observation incurs a large startup delay. To avoid such delay, we propose a new interactive LF streaming framework, where a client periodically requests viewpoint images, and in response a server synthesizes and transmits each requested image as a carefully chosen sparse linear combination of sub-aperture images. For each received synthesized image, the client "decodes" and recovers a new sub-aperture image using a cache of known sub-aperture images. As the cache of decoded sub-aperture images grows over time, the client becomes capable of synthesizing new view/focal point images, reducing overall transmission cost. Experimental results show that our proposed scheme can deliver synthesized images at high quality, even though only sparse sub-aperture images are used for synthesis. Moreover, compared with a scenario where the requested synthesized images are always transmitted, our proposed scheme achieves a significant reduction in accumulated rate when a sufficient number of images are transmitted.
Eduardo Peixoto received the Engineering and M.Sc. degrees from Universidade de Brasília, Brazil, in 2005 and 2008, respectively, and the Ph.D. degree from Queen Mary, University of London, in 2012, all in Electrical Engineering. In 2005-2006 he worked at BRISA developing Java applications for Digital Television and Mobile Phones. In 2006, he joined the Digital Signal Processing Group (GPDS) to work with Distributed Video Coding and Transcoding. In 2008, he joined Queen Mary, University of London to pursue a Ph.D., where he worked with Video Coding based on Wavelet Transforms, and with different video transcoders. In 2012, he joined Universidade de Brasília, first as a Substitute Lecturer and then, in 2014, as an Adjunct Professor. Currently, he is with the Department of Electrical Engineering at the University of Brasilia (UnB), where he is a member of the Graduate Program on Electronic Systems and Automation Engineering (PGEA). He is also a member of IEEE and the IEEE Signal Processing Society. His research interests include image and video processing and coding, as well as the development of applications for embedded systems. For more details, please visit: http://www.image.unb.br/eduardo WEB
The proliferation of new capabilities in affordable smart devices capable of capturing, processing and rendering audio-visual media content triggers a need for coordination and orchestration between these devices and their capabilities, and of the content flowing from and to such devices. The upcoming MPEG Media Orchestration standard (“MORE”, ISO/IEC 23001-13) enables the temporal and spatial orchestration of multiple media and metadata streams. Temporal orchestration is about time synchronisation of media and sensor captures, processing and renderings, for which the MORE standard uses and extends a DVB standard. Spatial orchestration is about the alignment of (global) position, altitude and orientation, for which the MORE standard provides dedicated timed metadata. Other types of orchestration involve timed metadata for region of interest, perceptual quality of media, audio-feature extraction and media timeline correlation.