Center for Digital Innovations (CDI)
Associate Director Steven Peterson
3871 Slichter Hall
156704
(Ping, Please eliminate all bullets/dots at left in the item below.)
Capture and Compression of Multi-Viewpoint Video
Allen Klinger
Background
Panoramic and multi-viewpoint images enable three-dimensional reconstruction of solid models of scenes. Digital computer programs process the constituent images. In some situations these programs are made efficient by special purpose hardware. That kind of hardware implements video capture and processing functions. The value of resultant three-dimensional information and the low cost of imaging devices makes research in this area likely to have application benefits.
Most panoramic techniques are either fisheye or use adjoined images stitching based. There are significant technical challenges in capturing and processing the huge amount of video data in the existing systems for acquiring multi-viewpoint data. The result is that current technology can only produce still images: it cannot handle time-varying or video data.
This proposal proceeds in cooperation with a commercial company that owns proprietary technology. That technology includes techniques to capture both panoramic and multi-viewpoint imagery. The company, Reality Commerce Corporation (RCC), will supply technical knowledge, information and practical experience. That information will be based on their systems - see description below (System Architecture).
We will be conducting joint research regarding enabling new methods for handling video imagery. That activity will be based on specific RCC technology, including proprietary real-time data compression.
Improved System Architecture
Two RCC patents pending are the foundation of the research:
1. Parallel Multi-Viewpoint Video Capturing and Compression (PMVCC): Method and Apparatus, U.S. Ser. No. 60/191,721.
2. Subject Video Streaming: Methods and Systems, U.S. Ser. No. 60/191,754.
Accelerated Algorithmic Processing
Graphics-based approaches to creating multi-viewpoint images involve first establishing a geometric model. This is followed by a texture rendering process. In contrast, the proposed system is based on individual images. This approach uses the actual images. They are taken simultaneously at multiple cameras. Each camera captures a different viewpoint.
The processing of individual images improves the overall system performance. Nevertheless the key is the careful registration and that is based on data structures and algorithms [1]. Where graphics-based methods are usually not real-time, our methods have high processing speed. The image-based system we propose can process both still image and moving pictures.
Several media processors are capable of capturing and compressing one or two channels of real-time video images. We propose in depth study of technical and economic aspects of designing and implementing a multiple input multiple data (MIMD) video parallel processing system, in order to extend that capability to the massive information handling needed in practical immersion or three-dimensional imaging situations. We further will investigate related compression algorithms for such hardware.
Applications and Markets
Online entertainment interactive multi-viewpoint 3D video images of live performances, sports, concerts and theatre presentations.
E-commerce Industry e-commerce platforms for businesses and consumers through web host designers, service providers and large retailers.
Online and Training and e-learning Industry interactive 3D training techniques tailored to distance learning, training and customer support facilities.
Surveillance and Remote Sensing agricultural crops and natural disaster monitoring from air and space.
As soon as I have an agreement with RCC I will supply information about this added application, possibly leading to a future source of business.
Competition
We have identified several competing technologies with products tailored to the 3D image and video marketplace. Currently, there are no competing 3D multi-viewpoint video imaging solutions available. Competing technologies include the following methods:
Digital Image Capture several companies have developed hardware capable of overlapping a series of still images and stitching them together to form a flat panoramic image. The process is time consuming as the process is not automated and requires tedious human intervention. In contrast, RCCs 3D multi-viewpoint image capturing device is a fully automated high performance capture device that can be used for both still image and video production without stitching artifacts.
Model Based Rendering Techniques software modeling technology has advanced and several companies are aggressively developing graphic renderings of animated objects for e-commerce marketing. The process is extremely time consuming and expensive to complete as a model must be developed for each image. Perception is also a problem as the final product is an animated view of the source object.
Panoramic 3D Images several groups have concentrated on the development of an optical system capable of capturing a 360 degree image around the camera in a single frame. This avoids the need of stitching, however, the image suffers from nonlinear distortion or "fish eye" viewing. Multi-viewpoint capture and content production is more challenging than panoramic content. Panoramic content captures only one viewing position. Multi-viewpoint content demands the capture and realization of picture taking at multiple spatial positions.
Research Objectives:
Design and analyze MIMD video parallel processing system architecture.
Design and analyze compression algorithm.
Compare the solution we proposed with other completing technologies
Research Team
Dr. Allen Klinger, Professor of computer science department, UCLA
Mr. Ping Liu, M.Sc, Post Engineer in UCLA, Director of Hardware Development, Reality Commerce Corp,.
(if resume is needed please let me know)
Research Schedule
Month 1: more investigation on related research
Month 2~4: Media Parallel Processing System Hardware Architecture Design and analysis
Month 5~7: Parallel video compression algorithm design and analysis
Month 8~9: Research project reporting
Research Project Cost
Desk-top Computer: $2,500.00
Travel : $8,000.00
Literature and industrial standard purchasing: $1,000.00
Video capturing card: $1,000.00
Two Video Camera: $1,000.00($500 each)
Video utility software: $3,000.00
Total: $16,500.00
Immersion Systems
1. National Tele-immersion Initiative Web site:
http://http://www. advanced.org/teleimmersion.html
2. Tele-immersion at Brown University:
http://www. cs.brown.edu/~lsh/telei.html
Andries van Dam, Loring Holden, Robert C. Zeleznik
http://www. cs.brown.edu/~lsh/telei.html
3. Tele-immersion at the University of North Carolina at Chapel Hill:
http://www.cs.unc.edu/Research/stc/teleimmersion/
Team Members: Henry Fuchs, Herman Towles, Greg Welch, Wei-Chao Chen, Ruigang Yang, Sang-Uok Kum, Andrew Nashel, Srihari Sukumaran
http://www. cs.unc.edu/Research/stc/ teleimmersion/
4. Tele-immersion at the University of Pennsylvania:
http://www.cis.upenn.edu/~sequence/teleim1.html
Ruzena Bajcsy, Kostas Daniilidis, Jane Mulligan, Ibrahim Volkan Isler
http://www. cis.upenn.edu/~sequence/teleim2.html
5. Tele-immersion site at Internet2:
http://www.internet2.edu/html/tele-immersion.html
6. Advanced Networks and Services:
http://www.advanced.org/teleimmersion.html
Jaron Lanier, Amela Sadagic