Award Abstract # 1910085
CNS Core: Small: Collaborative: Content-Based Viewport Prediction Framework for Live Virtual Reality Streaming

NSF Org: CNS
Division Of Computer and Network Systems
Recipient: RUTGERS, THE STATE UNIVERSITY
Initial Amendment Date: August 15, 2019
Latest Amendment Date: August 15, 2019
Award Number: 1910085
Award Instrument: Standard Grant
Program Manager: Deepankar Medhi
dmedhi@nsf.gov
 (703)292-2935
CNS
 Division Of Computer and Network Systems
CSE
 Direct For Computer & Info Scie & Enginr
Start Date: August 15, 2019
End Date: July 31, 2022 (Estimated)
Total Intended Award Amount: $214,874.00
Total Awarded Amount to Date: $214,874.00
Funds Obligated to Date: FY 2019 = $214,874.00
History of Investigator:
  • Sheng Wei (Principal Investigator)
    sheng.wei@rutgers.edu
Recipient Sponsored Research Office: Rutgers University New Brunswick
3 RUTGERS PLZ
NEW BRUNSWICK
NJ  US  08901-8559
(848)932-0150
Sponsor Congressional District: 12
Primary Place of Performance: Rutgers University New Brunswick
NJ  US  08854-3925
Primary Place of Performance
Congressional District:
06
Unique Entity Identifier (UEI): M1LVPE5GLSD9
Parent UEI:
NSF Program(s): CSR-Computer Systems Research,
Networking Technology and Syst
Primary Program Source: 01001920DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 7923
Program Element Code(s): 735400, 736300
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

Virtual reality (VR) video streaming has been gaining popularity recently with the rapid adoption of mobile head mounted display (HMD) devices in the consumer video market. As the cost for the immersive experience drops, VR video streaming introduces new bandwidth and performance challenges, especially in live streaming, due to the delivery of 360-degree views. This project develops a new content-based viewport prediction framework to improve the bandwidth and performance in live VR streaming, which predicts the user's viewport through a fusion of tracking the moving objects in the video, extracting the video semantics, and modeling the user's viewport of interest.

This project consists of three research thrusts. First, it develops a content-based viewport prediction framework for live VR streaming by tracking the motions and semantics of the objects. Second, it employs hardware and software techniques to facilitate real-time execution and scale the viewport prediction mechanism to a large number of users. Third, it develops evaluation frameworks to verify the functionality, performance, and scalability of the approach. The project uniquely considers the correlation between video content and user behavior, which leverages the deterministic nature of the former to conquer the randomness of the latter.

With the rapidly increasing popularity of VR systems in domain-specific immersive environments, the project will benefit several VR-related fields of studies with significant bandwidth savings and performance improvements, such as VR-based live broadcast, healthcare, and scientific visualization. Moreover, the interdisciplinary nature of the project will enhance the education and recruitment of underrepresented minorities in several science, technology, engineering, and mathematics (STEM) fields.

The project repository will be stored on a publicly accessible server (https://github.com/hwsel). All the project data will be maintained for at least five years following the end of the grant period.

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

Zhu, Zichen and Feng, Xianglong and Tang, Zhongze and Jiang, Nan and Guo, Tian and Xu, Lisong and Wei, Sheng "Power-Efficient Live Virtual Reality Streaming Using Edge Offloading" Proceedings of the 32nd Workshop on Network and Operating Systems Support for Digital Audio and Video , 2022 https://doi.org/10.1145/3534088.3534351 Citation Details
Feng, Xianglong and Li, Weitian and Wei, Sheng "LiveROI: Region of Interest Analysis for Viewport Prediction in Live Mobile Virtual Reality Streaming" ACM Multimedia Systems Conference , 2021 https://doi.org/10.1145/3458305.3463378 Citation Details
Feng, Xianglong and Bao, Zeyang and Wei, Sheng "LiveObj: Object Semantics-based Viewport Prediction for Live Mobile Virtual Reality Streaming" IEEE Transactions on Visualization and Computer Graphics , v.27 , 2021 https://doi.org/10.1109/TVCG.2021.3067686 Citation Details
Feng, Xianglong and Liu, Yao and Wei, Sheng "LiveDeep: Online Viewport Prediction for Live Virtual Reality Streaming Using Lifelong Deep Learning" IEEE Conference on Virtual Reality and 3D User Interfaces (VR) , 2020 10.1109/VR46266.2020.00104 Citation Details

PROJECT OUTCOMES REPORT

Disclaimer

This Project Outcomes Report for the General Public is displayed verbatim as submitted by the Principal Investigator (PI) for this award. Any opinions, findings, and conclusions or recommendations expressed in this Report are those of the PI and do not necessarily reflect the views of the National Science Foundation; NSF has not approved or endorsed its content.

This project developed a new content-based viewport prediction framework to reduce the bandwidth consumption when streaming live virtual reality (VR) videos from the content server to the end-user head mounted displays. During the course of the project, the PIs have accomplished three major research outcomes:

(1) a content-based viewport prediction framework for live VR video streaming considering the motion and semantics of the objects in the video;

(2) hardware/software techniques to enhance the speed and scalability of the proposed framework; and

(3) an evaluation framework to verify the functionality, performance, and scalability of the proposed viewport prediction framework.

Along with the research outcomes, the project has generated the following major products that address the intellectual merit and broader impacts of the conducted work.


Products that address intellectual merit: 

(1) Publications: The project has generated research publications in top conferences/journals of multimedia, virtual reality, mobile computing, computer vision, machine learning signal processing, and computer architecture/hardware.

(2) Software releases: The project has generated several open source software releases for the viewport prediction framework, available online at https://github.com/hwsel/

(3) Dataset: The project has generated a 360-degree human activity video dataset, available online at https://egok360.github.io/


Products that address broader impacts: 

(1) Education materials: The content and outcomes of the project have been incorporated as education materials in several undergraduate/graduate courses offered by the PIs in their specific institutions, such as Introduction to Machine Learning, Advanced Image Processing and Computer Vision, Deep Learning Embedded Systems, Advanced Topics on Computer Vision and Multimedia, and Advances in Deep Learning. The courses and education materials benefitted a large group of undergraduate and graduate students in their programs of studies and future career development.

(2) Student mentoring: The project has supported the PIs mentoring multiple graduate and undergraduate students in their specific institutions. The students have worked closely with the PIs in the planned research activities, with regularly scheduled research meetings and mentoring sessions. The students have learned the required knowledge and expertise in multimedia systems/signal processing, deep learning, and computer vision.

(3) Presentations and media coverages: The project has resulted in invited presentations at conferences, research institutions, and industry, as well as coverages at various media channels. The presentations and media coverages serve as important means of research dissemination to the society for potential technology transfer.


Last Modified: 11/29/2022
Modified by: Sheng Wei

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page