PANews reported on October 16 that WorldLabs, a spatial intelligence company founded by Stanford University professor Fei-Fei Li, officially released a research preview version of its new generation real-time generation model RTFM (Real-Time Frame Model), which can generate videos and 3D scenes in real time during the interaction process, realizing a persistent virtual world. RTFM is designed around three core principles: efficiency, scalability, and persistence. The model runs at interactive frame rates on a single NVIDIA H100 GPU, generating scenes with complex effects like reflections, shadows, and glossiness from a single image without explicit 3D modeling. Its "spatial memory" mechanism maintains world persistence through frame pose modeling and context juggling, supporting long-term interactive exploration.PANews reported on October 16 that WorldLabs, a spatial intelligence company founded by Stanford University professor Fei-Fei Li, officially released a research preview version of its new generation real-time generation model RTFM (Real-Time Frame Model), which can generate videos and 3D scenes in real time during the interaction process, realizing a persistent virtual world. RTFM is designed around three core principles: efficiency, scalability, and persistence. The model runs at interactive frame rates on a single NVIDIA H100 GPU, generating scenes with complex effects like reflections, shadows, and glossiness from a single image without explicit 3D modeling. Its "spatial memory" mechanism maintains world persistence through frame pose modeling and context juggling, supporting long-term interactive exploration.

World Labs releases real-time generative world model (RTFM): a single H100 GPU can generate 3D worlds in real time

2025/10/16 23:20

PANews reported on October 16 that WorldLabs, a spatial intelligence company founded by Stanford University professor Fei-Fei Li, officially released a research preview version of its new generation real-time generation model RTFM (Real-Time Frame Model), which can generate videos and 3D scenes in real time during the interaction process, realizing a persistent virtual world.

RTFM is designed around three core principles: efficiency, scalability, and persistence. The model runs at interactive frame rates on a single NVIDIA H100 GPU, generating scenes with complex effects like reflections, shadows, and glossiness from a single image without explicit 3D modeling. Its "spatial memory" mechanism maintains world persistence through frame pose modeling and context juggling, supporting long-term interactive exploration.

Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact service@support.mexc.com for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.