Original Reddit post

Hey r/ArtificialInteligence , I’ve been working on a Flutter-based Android AI assistant called JARVIS that goes beyond just chatting — it actually controls your device. AI & Chat

  • Conversational AI powered by Groq (fast inference)
  • Choose from multiple models: Llama 3.3 70B, Llama 4 Scout/Maverick, Qwen3 32B, Kimi K2, and more
  • Full conversation history with multi-session support
  • Text-to-Speech responses
  • Wake word detection for hands-free activation Voice Commands
  • “Open YouTube” → launches the app instantly
  • “What’s the weather in Tokyo?” → real-time weather via OpenWeather API
  • “Open WiFi settings” → jumps straight to the setting
  • “Go to github.com” → opens in browser
  • “What time is it?” / “What’s today’s date?” → pulls from device
  • “Calculate 25 * 4 + 10” → instant math
  • “Set a reminder for 5 minutes”
  • Task creation and management by voice System Control (Accessibility Service)
  • Navigate back, home, recent apps
  • Take screenshots
  • Lock screen
  • Read what’s on screen
  • Click buttons by description
  • Auto-fill text fields
  • Read notifications
  • Swipe, pinch, scroll, double tap, long press gestures
  • Split screen toggle UI & Other
  • Floating overlay window — accessible from any app
  • Blueprint/Canvas screen for visual planning
  • Dark minimalist design with Geist font
  • Google Sign-In + Firebase backend
  • Fully configurable (API keys, voice settings, custom commands) The AI side: The app connects to Groq’s API for fast inference and lets you switch between models on the fly:
  • Llama 3.3 70B
  • Llama 4 Scout & Maverick
  • Qwen3 32B
  • Kimi K2
  • Llama 3.1 8B Tool use / function calling: The AI has access to real tools it can invoke:
  • get_current_time / get_current_date — pulls live from device
  • get_weather_info — OpenWeather API with GPS coordinates
  • calculate — math expression evaluator
  • search_web — web search
  • open_app — launches any installed app
  • open_setting — opens any system setting
  • open_link — opens URLs in browser
  • perform_system_action — back, home, screenshot, lock, gestures, etc.
  • get_screen_content — reads what’s on screen
  • click_by_description — clicks UI elements
  • fill_text_field — auto-fills inputs
  • get_recent_notifications — reads notification panel
  • Task management tools (create, update, complete tasks) It’s essentially an agentic assistant that can reason about what you need and take action on your device. Wake word detection keeps it always ready. Currently sideloadable via ADB. Full accessibility features require Android 12 or lower, or the upcoming Play Store release. Curious what the community thinks about the tool-use design and model selection approach. If you want to try it or follow development, join the Discord: https://discord.com/invite/JGBYCGk5WC

submitted by /u/Mikeeeyy04

Originally posted by u/Mikeeeyy04 on r/ArtificialInteligence