MiniMax Hailuo Video Team Drops VTP: The First Open-Source Scalable Visual Tokenizer Pre-Training Framework — Revolutionizing Generative Video Pipelines
On December 16, 2025, the MiniMax Hailuo Video team officially open-sourced VTP (Visual Tokenizer Pre-training) — a groundbreaking unified framework for pre-training visual tokenizers optimized for downstream generation tasks. By jointly optimizing contrastive, self-supervised, and reconstruction losses, VTP creates semantic-rich latent spaces that scale dramatically better than traditional autoencoders, delivering 65.8% FID gains in DiT-based video/image generation with just more pre-training FLOPs. Models (0.2B-0.3B) and code are now live on GitHub and Hugging Face, empowering the community to build next-gen Hailuo-level video models without starting from scratch.
Luma AI Scores $900M Mega-Round: Fueling 2GW Saudi Supercluster to Dominate Multimodal Video AGI and Crush the Compute Crunch
Luma AI, the breakout star behind Ray3 video generation, just inked a $900 million Series C on November 19, 2025 — led by Saudi powerhouse HUMAIN (PIF-backed), with heavy hits from AMD Ventures and a16z. The war chest unlocks exclusive access to Project Halo, HUMAIN's beastly 2-gigawatt AI supercluster in Saudi Arabia, set to go live Q1 2026 and rival xAI's Colossus in raw scale. Valued at $4B+, this isn't just cash — it's a geopolitical compute coup positioning Luma to train world-modeling AGI from video floods, leaving OpenAI and Google scrambling in the multimodal dust.


