Original Reddit post

Saw an interesting technical setup recently and wanted to get the sub’s take on it. Ailoitte published their AI Velocity Pod architecture, which uses: • Claude (Anthropic) as the primary reasoning model, integrated into Cursor IDE • Custom .cursorrules files and proprietary datasets to enforce project-specific code architecture from day one • Agentic QA agents that automatically write and run end-to-end tests based on PR descriptions • Dedicated VPC environment per client (SOC2, ISO 27001:2013) The claim is 5× code velocity vs traditional approaches, with first commit delivery in 7 days from contract. They describe the senior engineer’s role as a ‘conductor of high-intelligence agents’ rather than a line-by-line coder. Technical questions I’m genuinely curious about: • Has anyone here used Claude + Cursor as a primary production stack (not just for personal projects)? • What’s the practical limit of .cursorrules customization for enforcing architectural patterns? • The Agentic QA claim (agents writing tests automatically from PR descriptions). What’s your experience with the reliability of AI-generated tests in production? Not trying to promote anything here, just found the architecture interesting and want to hear from people who’ve worked with similar stacks. #AI #SoftwareDevelopment #Claude #CursorIDE #AgenticAI submitted by /u/Individual-Bench4448

Originally posted by u/Individual-Bench4448 on r/ArtificialInteligence