Hey r/ArtificialInteligence , I’ve been working on a Flutter-based Android AI assistant called JARVIS that goes beyond just chatting — it actually controls your device. AI & Chat
- Conversational AI powered by Groq (fast inference)
- Choose from multiple models: Llama 3.3 70B, Llama 4 Scout/Maverick, Qwen3 32B, Kimi K2, and more
- Full conversation history with multi-session support
- Text-to-Speech responses
- Wake word detection for hands-free activation Voice Commands
- “Open YouTube” → launches the app instantly
- “What’s the weather in Tokyo?” → real-time weather via OpenWeather API
- “Open WiFi settings” → jumps straight to the setting
- “Go to github.com” → opens in browser
- “What time is it?” / “What’s today’s date?” → pulls from device
- “Calculate 25 * 4 + 10” → instant math
- “Set a reminder for 5 minutes”
- Task creation and management by voice System Control (Accessibility Service)
- Navigate back, home, recent apps
- Take screenshots
- Lock screen
- Read what’s on screen
- Click buttons by description
- Auto-fill text fields
- Read notifications
- Swipe, pinch, scroll, double tap, long press gestures
- Split screen toggle UI & Other
- Floating overlay window — accessible from any app
- Blueprint/Canvas screen for visual planning
- Dark minimalist design with Geist font
- Google Sign-In + Firebase backend
- Fully configurable (API keys, voice settings, custom commands) The AI side: The app connects to Groq’s API for fast inference and lets you switch between models on the fly:
- Llama 3.3 70B
- Llama 4 Scout & Maverick
- Qwen3 32B
- Kimi K2
- Llama 3.1 8B Tool use / function calling: The AI has access to real tools it can invoke:
get_current_time/get_current_date— pulls live from deviceget_weather_info— OpenWeather API with GPS coordinatescalculate— math expression evaluatorsearch_web— web searchopen_app— launches any installed appopen_setting— opens any system settingopen_link— opens URLs in browserperform_system_action— back, home, screenshot, lock, gestures, etc.get_screen_content— reads what’s on screenclick_by_description— clicks UI elementsfill_text_field— auto-fills inputsget_recent_notifications— reads notification panel- Task management tools (create, update, complete tasks) It’s essentially an agentic assistant that can reason about what you need and take action on your device. Wake word detection keeps it always ready. Currently sideloadable via ADB. Full accessibility features require Android 12 or lower, or the upcoming Play Store release. Curious what the community thinks about the tool-use design and model selection approach. If you want to try it or follow development, join the Discord: https://discord.com/invite/JGBYCGk5WC
submitted by /u/Mikeeeyy04
Originally posted by u/Mikeeeyy04 on r/ArtificialInteligence
You must log in or # to comment.
