Imagine streaming immersive 3D content without the hefty bandwidth demands. NYU Tandon’s latest research makes this a reality, predicting what users will see next to optimize data transmission.
Key Points at a Glance
- New AI-driven method predicts users’ field of view in 3D environments.
- Reduces bandwidth requirements for VR/AR streaming by up to 85%.
- Utilizes transformer-based graph neural networks and recurrent neural networks.
- Enhances accessibility of immersive technologies on standard devices.
- Potential applications in education, entertainment, and beyond.
In a significant advancement for virtual and augmented reality (VR/AR) technologies, researchers at NYU Tandon School of Engineering have developed an innovative streaming method that dramatically reduces the bandwidth needed for immersive 3D content. By accurately predicting the viewer’s field of view (FoV), the system transmits only the data necessary for the user’s immediate perspective, cutting bandwidth usage by up to 85%.
Traditional 3D streaming methods often involve transmitting the entire scene, regardless of where the user is looking, leading to substantial data consumption. The new approach, however, leverages advanced machine learning techniques to anticipate the user’s gaze direction and stream only the relevant portions of the 3D environment.
“Our system is akin to having your eyes guide the streaming process,” explained Professor Yong Liu, who led the research team. “It focuses on delivering just what you need to see, enhancing efficiency without compromising the immersive experience.”
The technology employs a combination of transformer-based graph neural networks to understand spatial relationships within the 3D space and recurrent neural networks to model how visibility patterns change over time. This dual approach allows the system to predict what a user will see 2 to 5 seconds ahead, a significant improvement over previous methods that could only forecast a fraction of a second into the future.

Beyond reducing bandwidth, the system maintains real-time performance, achieving over 30 frames per second even with complex point cloud videos containing more than a million points per frame. This ensures a seamless and responsive experience for users, even on devices with standard internet connections.
The implications of this research are far-reaching. By lowering the data requirements for high-quality 3D streaming, the technology opens doors for broader adoption of VR/AR applications in fields such as education, where it can facilitate remote learning experiences, and entertainment, where it can enhance interactive media consumption.
Furthermore, the research team has made their code publicly available, encouraging further development and integration of the technology into various applications. This move aligns with the broader goal of democratizing access to immersive technologies, making them more accessible to a wider audience.
As VR and AR continue to evolve, innovations like this predictive streaming method are crucial for overcoming existing limitations and expanding the potential of immersive experiences. By intelligently managing data transmission, NYU Tandon’s research represents a significant step toward more efficient and accessible 3D streaming solutions.
Source: NYU Tandon School of Engineering