Skip to main content
This guide will help you get Mobilerun installed and running quickly, controlling your Android device through natural language in minutes.

Prerequisites

Before installing Mobilerun, ensure you have:
  1. Python 3.11-3.13 installed on your system
  2. Android Debug Bridge (adb) installed and configured
  3. Android device with:

Installation

Mobilerun is installed using uv, a fast Python package installer and resolver. Install uv (if not already installed):
# macOS/Linux
curl -LsSf https://astral.sh/uv/install.sh | sh

# Windows (PowerShell)
powershell -c "irm https://astral.sh/uv/install.ps1 | iex"
Choose your installation method: For CLI usage only:
uv tool install mobilerun
For CLI + Python code integration:
uv pip install mobilerun
Most LLM providers (Google Gemini, OpenAI, Ollama, OpenRouter) are included by default. For additional providers, install extras: uv tool install 'mobilerun[anthropic,deepseek]'.

Set Up the Portal APK

Mobilerun requires the Portal app to be installed on your Android device for device control. The Portal app provides accessibility services that expose the UI accessibility tree, enabling the agent to see and interact with UI elements.
mobilerun setup
This command automatically:
  1. Downloads the latest Portal APK
  2. Installs it on your connected device
  3. Enables the accessibility service

Test Connection

Verify that Mobilerun can communicate with your device:
mobilerun ping
If successful, you’ll see:
Portal is installed and accessible. You're good to go!

Configure Your LLM

Run the configure wizard to choose your provider, auth method (API key or OAuth), and model:
mobilerun configure
Alternatively, you can set an API key as an environment variable:
# For Google Gemini (default)
export GOOGLE_API_KEY=your-api-key-here

# For OpenAI
export OPENAI_API_KEY=your-api-key-here

# For Anthropic Claude
export ANTHROPIC_API_KEY=your-api-key-here

Run Your First Command via CLI

Now you’re ready to control your device with natural language:
# Using default configuration (Google Gemini)
mobilerun run "Open the settings app and tell me the Android version"

# Override provider and model
mobilerun run "Check the battery level" --provider OpenAI --model gpt-4o

# Enable vision mode (sends screenshots to LLM)
mobilerun run "What app is currently open?" --vision

# Enable reasoning mode (uses Manager-Executor workflow for complex tasks)
mobilerun run "Find a contact named John and send him an email" --reasoning
Common CLI flags:
  • --provider - LLM provider (GoogleGenAI, OpenAI, Anthropic, etc.)
  • --model - Model name (gemini-3.1-flash-lite-preview, gpt-4o, etc.)
  • --vision - Enable screenshot processing
  • --reasoning - Enable multi-agent planning mode
  • --steps N - Maximum execution steps (default: 15)
  • --debug - Enable detailed logging

Create a Simple Agent via Script

For complex automation or integration into your Python projects, create a script:
import asyncio
from mobilerun import MobileAgent, MobileConfig

async def main():
    # Use default configuration with built-in LLM profiles
    config = MobileConfig()

    # Create agent
    # LLMs are automatically loaded from config.llm_profiles
    agent = MobileAgent(
        goal="Open Settings and check battery level",
        config=config,
    )

    # Run agent
    result = await agent.run()

    # Check results (result is a ResultEvent object)
    print(f"Success: {result.success}")
    print(f"Reason: {result.reason}")
    print(f"Steps: {result.steps}")

if __name__ == "__main__":
    asyncio.run(main())

Next Steps

Now that you’ve got Mobilerun running, explore these topics: