Original Reddit post

Been building Izwi, a fully local audio inference stack for speech workflows. No cloud APIs, no data leaving your machine. What’s inside: Text-to-speech & speech recognition (ASR) Voice cloning & voice design Chat/audio-chat models OpenAI-compatible API ( /v1 routes) Apple Silicon acceleration (Metal) Stack: Rust backend (Candle/MLX), React/Vite UI, CLI-first workflow. Everything runs locally. Pull models from Hugging Face, benchmark throughput, or just izwi tts “Hello world” and go. Apache 2.0, actively developed. Would love feedback from anyone working on local ML in Rust! GitHub: github.com/agentem/izwi submitted by /u/zinyando

Originally posted by u/zinyando on r/ArtificialInteligence