Original Reddit post

With all the hype around vibe coding and AI writing code, I wonder if the reality might be less rosy for developers than we hope. AI can churn out code fast, but it’s often sloppy, inconsistent, and full of hidden vulnerabilities. Small bugs can lead to security holes, database risks, or privacy issues. Also, maintaining production databases and products requires a lot of effort Like, imagine a vibe-coded fitness application that got 10k users in a month and is generating good revenue. But next week, a data breach happens and customer data is leaked In such cases, it seems like the ones who really end up profiting might be lawyers handling compliance, privacy, or customer data breach claims, rather than the developers who built the code. I might be overthinking it, but does anyone else see this as a real risk, or do you think we’ll develop reliable ways to audit and harden AI-generated code before it causes problems? submitted by /u/ocean_protocol

Originally posted by u/ocean_protocol on r/ArtificialInteligence