Original Reddit post

I was wondering if this either exists, or would be a useful addition to current video generation llm’s (and any other tbh). I think it would be useful for a video generation LLM (like seedance), to confer with a storyteller LLM, a director LLM, a cinematography LLM, an acting LLM, etc, before it comes up with the next token. It could take the results of those suggestions, and add them into the context of the next token generation. Perhaps not for every token, but only ones symbolizing the end of a “scene”, for example. This could lead to better camera angles, better acting, but story elements, plot points, etc. Do you think this could work? Or is it already being done? submitted by /u/BECOMING_A_TURTLE

Originally posted by u/BECOMING_A_TURTLE on r/ArtificialInteligence