Original Reddit post

So I’ve been experimenting with a more structured workflow where the system starts with a single prompt then plans the sequence scene-by-scene before generation instead of treating the whole film as one giant prompt. Made this 40s cinematic train sequence using that approach. Prompt: “Create a cinematic travel film for a remote mountain railway in winter. Show snow, steam, steel, cold morning light, and small human moments inside the train. Let the film feel poetic and grounded, with connected scene transitions that make the journey feel continuous and real.” Workflow was roughly: storyboard planning scene-level visual mapping different continuity strategies per shot chaining from previous scene endings when needed automatic clip generation + sequencing Some scenes start fresh. Others inherit visual continuity from previous shots. The interesting part for me is that the workflow stays editable at the scene level instead of locking everything into one generation pass. Attached: final output visual planning workflow before generation Still seeing limitations with: object permanence dynamic motion consistency maintaining identity through complex camera movement But orchestration/control feels like the bigger unlock now, not just raw generation quality. Curious where people think this goes long term. If future models eventually generate perfectly coherent long-form films on their own, does that actually reduce creative control for filmmakers? Feels like the more interesting direction might be systems where the AI handles execution, but humans still shape pacing, continuity, scene structure, and intent at a granular level. submitted by /u/meet_og

Originally posted by u/meet_og on r/ArtificialInteligence