PANews reported on October 16 that WorldLabs, a spatial intelligence company founded by Stanford University professor Fei-Fei Li, officially released a research preview version of its new generation real-time generation model RTFM (Real-Time Frame Model), which can generate videos and 3D scenes in real time during the interaction process, realizing a persistent virtual world. RTFM is designed around three core principles: efficiency, scalability, and persistence. The model runs at interactive frame rates on a single NVIDIA H100 GPU, generating scenes with complex effects like reflections, shadows, and glossiness from a single image without explicit 3D modeling. Its "spatial memory" mechanism maintains world persistence through frame pose modeling and context juggling, supporting long-term interactive exploration.PANews reported on October 16 that WorldLabs, a spatial intelligence company founded by Stanford University professor Fei-Fei Li, officially released a research preview version of its new generation real-time generation model RTFM (Real-Time Frame Model), which can generate videos and 3D scenes in real time during the interaction process, realizing a persistent virtual world. RTFM is designed around three core principles: efficiency, scalability, and persistence. The model runs at interactive frame rates on a single NVIDIA H100 GPU, generating scenes with complex effects like reflections, shadows, and glossiness from a single image without explicit 3D modeling. Its "spatial memory" mechanism maintains world persistence through frame pose modeling and context juggling, supporting long-term interactive exploration.

World Labs releases real-time generative world model (RTFM): a single H100 GPU can generate 3D worlds in real time

2025/10/16 23:20

PANews reported on October 16 that WorldLabs, a spatial intelligence company founded by Stanford University professor Fei-Fei Li, officially released a research preview version of its new generation real-time generation model RTFM (Real-Time Frame Model), which can generate videos and 3D scenes in real time during the interaction process, realizing a persistent virtual world.

RTFM is designed around three core principles: efficiency, scalability, and persistence. The model runs at interactive frame rates on a single NVIDIA H100 GPU, generating scenes with complex effects like reflections, shadows, and glossiness from a single image without explicit 3D modeling. Its "spatial memory" mechanism maintains world persistence through frame pose modeling and context juggling, supporting long-term interactive exploration.

Sorumluluk Reddi: Bu sitede yeniden yayınlanan makaleler, halka açık platformlardan alınmıştır ve yalnızca bilgilendirme amaçlıdır. MEXC'nin görüşlerini yansıtmayabilir. Tüm hakları telif sahiplerine aittir. Herhangi bir içeriğin üçüncü taraf haklarını ihlal ettiğini düşünüyorsanız, kaldırılması için lütfen service@support.mexc.com ile iletişime geçin. MEXC, içeriğin doğruluğu, eksiksizliği veya güncelliği konusunda hiçbir garanti vermez ve sağlanan bilgilere dayalı olarak alınan herhangi bir eylemden sorumlu değildir. İçerik, finansal, yasal veya diğer profesyonel tavsiye niteliğinde değildir ve MEXC tarafından bir tavsiye veya onay olarak değerlendirilmemelidir.

Ayrıca Şunları da Beğenebilirsiniz