Original Reddit post

Disclosure: I’m the builder. I’m not an engineer. I’m a PM by background. This is the first thing I’ve ever built end to end. All code was generated via AI-assisted development (ChatGPT 5.5 primarily, but Claude contributed as well). I made the architectural decisions and iterated on prompts. I did not write or debug code by hand. Why I built it I was tracking 100+ sources to stay current on AI (news sites, blogs, newsletters, podcasts, Substacks, Discord). Signal-to-noise was awful and general news apps treat AI as one tag among twenty. I wanted the inverse: AI as the entire feed. Stack Ingestion: RSS pipeline pulling 100+ feeds. Polling cadence varies by source priority. Dedup on title similarity + canonical URL. Categorization: LLM classification into a taxonomy (product, startups, finance, health, etc.) plus a “significance” score. Ranking: Hybrid of recency decay, source authority weight, and LLM significance. Pure-recency and pure-LLM both lost to the hybrid in informal A/B against my own reading patterns. Audio: TTS per article. Aggressive caching because regenerating audio is the biggest cost line. Ask tab: Conversational news experience. Speech-to-text into structured DB queries against the article store, then a generated answer with sources. Wiring the STT output into reliable query intent (filters, time windows, entities) was the hardest part to get right. Client: Native iOS. SwiftUI with UIKit where it made sense. No React Native, no web wrapper. What broke / lessons Source quality is everything. A wide net pulled in low-quality SEO content the classifier kept boosting because it was technically “AI news.” Hand-curation fixed more than any model change. LLM categorization drifts. Same article, different runs, different categories. Added a majority-vote stability layer for borderline items. Push notifications are a UX minefield. Permission timing, APNs setup, and “deny” handling took longer than the ranking algorithm. STT to DB query is harder than it looks. People speak in fragments (“what’d Anthropic ship this week”). Mapping that to filters and time windows reliably took more iteration than any other part of the app. It’s still not perfect, but works. TTS cost discipline matters early. Without caching and batching, audio alone would’ve killed unit economics. “No code” is the right frame for me. I didn’t write or debug Swift. Coding isn’t one of my skills. What I brought was product thinking: defining the system, the taxonomy, the ranking signals, the conversational UX. The AI did the engineering. The honest takeaway is that this kind of build is now possible for non-engineers who can specify clearly and stay opinionated about the product. Codex (ChatGPT 5.5) vs Claude 4.7 : Based on my experience, I started leaning more toward Codex vs Claude. Codex one-shotted the scenario often when Claude ran into a cycle of bugs. Limitations US/English-biased source coverage. Significance score calibrated to my reading taste. Personalization is the obvious next step. Coarse categorization. “Models/products” lumps frontier releases with fine-tune announcements. No Android yet. Links: https://news.thisweek.ai/ https://apps.apple.com/us/app/thisweek-ai/id6759996996 Happy to answer any questions to take feedback on the app. submitted by /u/techila

Originally posted by u/techila on r/ArtificialInteligence