Been thinking about this a lot lately and I want to see if anyone else has gone down this rabbit hole. When Cursor came out, it wasn’t just “AI autocomplete in VS Code.” The key shift was that the AI had structural understanding of your entire codebase, how one file affects another, what a function is actually doing in context. It stopped being a plugin and became the environment itself. Music production hasn’t had that moment yet. Every AI tool for DAWs right now is essentially a smarter plugin. It sits inside Ableton or Logic. The host doesn’t change. The metaphor doesn’t change. We’re still working on a 40-year-old “tape machine” abstraction. So what would it actually look like to build a DAW from scratch around AI context? A few ideas I keep coming back to: A “Sonic AST” (semantic understanding): instead of the AI seeing isolated plugin parameters, it indexes the whole project. Key, scale, the timbre of your kick, the frequency buildup across your synths, every patch state. So instead of asking for “a preset,” you say: “make the bassline sit under the kick without losing its warmth”, and the AI applies sidechain or dynamic EQ because it actually sees the conflict. Tab-to-complete for arrangement — you play a 2-bar melody, ghost notes suggest the next 2 bars based on your patterns and genre theory. Hit Tab to accept. Or you highlight the gap between your verse and chorus, hit Cmd+K: “4-bar build with a white noise riser and a drum fill that slows into the drop”… and it generates the automation! Natural language mixing: instead of knowing “boxy = cut at 300Hz,” you just say “vocals feel boxy, push them further back in the room.” The AI analyzes the specific frequencies and adjusts EQ and reverb wet/dry in context, not from a preset. Semantic sample search, vector database of your local library. You type: “find a snare like the one in Midnight City by M83 but grittier” and it maps the closest match to your sampler, pre-tuned to your key. The biggest shift though is conceptual: you stop being the person turning the knobs and become the director. “Rearrange this 8-bar loop into a rough 3-minute pop structure” — it slices, duplicates, drafts the skeleton, and you edit from there. Is anyone actually building something like this? I know tools like Udio/Suno exist but they generate audio as an output, that’s not the same as an AI-native production environment where you’re still in creative control. Feels like the Cursor equivalent for music is still wide open. What am I missing? submitted by /u/KNTRL108
Originally posted by u/KNTRL108 on r/ClaudeCode
