Apple released STARFlow-V LAST WEEK, a 7-billion parameter video generation model with open weights—a surprising move from a company known for locking down AI research. The model uses normalizing flows instead of the diffusion models dominating video AI, representing the first time this architecture has worked at scale for video generation. It’s trending on Hacker News with 225 points, though developers quickly pointed out the license isn’t truly open.
This marks a significant shift for Apple, which rarely shares AI model weights. But before celebrating Apple’s newfound openness, read the fine print.
The “Open” Weights Come With Strings Attached
STARFlow-V is available for research, but commercial use requires written approval from Apple. Redistributing modified weights? Also needs Apple’s explicit consent. The Hacker News community was blunt: “It’s more appropriate to call it ‘weights available’ rather than truly open.”
This restrictive licensing reflects Apple’s ongoing internal tension around AI openness. Earlier this year, software chief Craig Federighi vetoed plans to fully open-source Apple Intelligence models, reportedly worried it would reveal how much their performance degraded when compressed for iPhones. The decision triggered a talent exodus, with Ruoming Pang—head of Apple’s foundation models team—leaving for Meta.
STARFlow-V represents Apple’s compromise: gain research credibility through a NeurIPS 2025 spotlight while maintaining control through restrictive licensing. Federighi now instructs teams to “do whatever it takes” for AI features, even using competitors’ open models. This release signals Apple testing a middle path between its secrecy culture and the AI industry’s collaborative norms.
Normalizing Flows vs. Diffusion: A Different Architecture
While Runway Gen 4.5, OpenAI Sora, and Stable Video Diffusion all use diffusion models, STARFlow-V takes a different approach with normalizing flows. The advantage? Single-pass generation instead of iterative refinement, exact likelihood evaluation, and better handling of causal video sequences.
Apple’s architecture combines a deep Transformer block for representational capacity with shallow blocks for efficiency. The 7-billion parameter model generates 480p video at 16 fps in one forward pass—trained on 70 million text-video pairs and 400 million text-image pairs. According to Apple’s GitHub repository, this is “the first successful demonstration of normalizing flows operating effectively at this scale and resolution.”
The technical innovation is real. Normalizing flows have historically struggled with high-resolution content. But architectural novelty doesn’t guarantee better results. Community feedback on the generated examples was lukewarm, with one commenter noting they had “the feeling of the early Will Smith noodles videos”—a reference to notorious AI generation artifacts.
Where STARFlow-V Fits in the Video AI Landscape
STARFlow-V occupies an awkward middle ground. It’s more accessible than Runway Gen 4.5 or OpenAI Sora—both closed-source commercial products requiring paid API access. But it’s less open than Stable Video Diffusion, which has a permissive license allowing unrestricted commercial use and redistribution.
Runway Gen 4.5 currently dominates the Video Arena leaderboard, beating both OpenAI Sora 2 Pro and Google Veo 3 in blind user tests. It excels at realistic physics, human motion, and camera control—generating HD video that’s a clear step above STARFlow-V’s 480p output. For developers needing the highest quality NOW, closed commercial models remain the only option.
STARFlow-V makes more sense for research projects and specific use cases where local deployment matters. The 7-billion parameter size means it might run on high-end consumer GPUs with layer swapping techniques. But there’s a catch: the model weights aren’t actually available yet. Apple’s GitHub repository says they’re “coming soon” with no specific timeline.
Accessibility Applications Show the Most Promise
The strongest enthusiasm from developers centers on accessibility. Video description generation could enable blind users to understand video content independently—a use case where 480p resolution and imperfect quality matter less than having any automated understanding at all. One blind developer on Hacker News described how AI video tools have “changed my life,” opening up content previously inaccessible.
Research applications also make sense despite the licensing restrictions. Academic projects exploring normalizing flows for video generation now have a reference implementation to study. The model’s architecture and training approach provide valuable insights even if commercial deployment requires navigating Apple’s approval process.
Apple’s AI Strategy Remains Contradictory
STARFlow-V reveals Apple caught between competing pressures. The company wants research credibility—hence the NeurIPS spotlight and public release. But it can’t fully embrace the open-source ethos that Meta and Google have adopted. The result is a half-measure that satisfies neither camp completely.
Compare this to Meta’s approach: when Meta releases AI research, the weights come with permissive licenses enabling broad commercial use. Google similarly publishes research with accessible implementations. Apple’s restrictive licensing suggests it’s more interested in appearing open than actually being open.
The timing is notable. After Federighi’s veto on open-sourcing earlier this year caused talent to leave for competitors, Apple appears to be recalibrating. STARFlow-V tests whether controlled releases can rebuild research credibility without sacrificing the company’s obsession with control.
Key Takeaways
- Apple released open weights for STARFlow-V, a 7B parameter video model using normalizing flows—but the license restricts commercial use and requires approval for redistribution
- Technical innovation is real: first normalizing flow model working at scale for video, with single-pass generation and exact likelihood evaluation
- Quality trails closed leaders like Runway Gen 4.5 based on community feedback, generating 480p @ 16fps vs HD from commercial alternatives
- Best for accessibility and research, where restricted licensing matters less and local deployment offers advantages over API-only access
- Apple’s AI strategy remains contradictory, caught between research credibility needs and control obsession, testing a middle path that satisfies neither open-source advocates nor secrecy traditionalists









