- Platform (Recommended)
- Local Development
With the platform, profiles are configured on platform.minitap.ai and referenced by name in your code.
- No config files needed
- Update models without code changes
- Access to all OpenRouter models
- Share profiles across your team
What are Agent Profiles?
An agent profile defines the LLM configuration for the various sub-agents that power mobile-use:- Planner - Creates high-level plans from goals
- Orchestrator - Coordinates execution steps
- Cortex - Visual understanding and decision-making (most important)
- Executor - Performs specific actions
- Utils - Helper agents (outputter, hopper)
Platform Profiles
Using the platform? Create and manage profiles on platform.minitap.ai/llm-profiles, then reference them by name in your tasks.
Creating Platform Profiles
- Go to LLM Profiles on the platform
- Click Create Profile
- Configure each agent component with your preferred models
- All OpenRouter models are available (no API key management needed)
- Reference the profile by name in your code:
Platform profiles can be updated anytime without changing code - perfect for A/B testing different models!
Local Profiles
For local development, profiles are defined in config files or code:From Configuration File
The recommended approach for local production:Programmatic Configuration
For dynamic configuration:You cannot specify both
llm_config and from_file - they are mutually exclusive.Using Profiles
Setting a Default Profile
Configure an agent with a default profile:Multiple Profiles
Register multiple profiles and switch between them:Profile Use Cases
- Speed vs Accuracy
- Cost Optimization
- Provider Diversity
Fast Profile - Quick tasks, simple UI navigationAccurate Profile - Complex analysis, detailed extraction
Component Roles
Cortex (Most Important)
Cortex (Most Important)
The cortex is the visual understanding component - the โeyes and brainโ of the agent.
- Analyzes screenshots
- Understands UI elements
- Makes decisions about actions
- Recommendation: Use the best vision model available (e.g.,
o4-mini,gpt-5) - Context window: Needs at least 128k tokens
Planner
Planner
Creates high-level plans from natural language goals.
- Breaks down goals into subgoals
- Estimates complexity
- Recommendation: Use fast, capable text models (e.g.,
gpt-5-nano)
Orchestrator
Orchestrator
Coordinates execution and decides when to replicate.
- Manages task flow
- Handles errors and retries
- Recommendation: Use fast models (e.g.,
gpt-5-nano)
Executor
Executor
Translates decisions into device actions.
- Generates Maestro commands
- Handles action formatting
- Recommendation: Use fast, instruction-following models
Utils - Hopper
Utils - Hopper
Digs through large batches of data to extract the most relevant information for reaching the goal.
- Processes extensive historical context and screen data
- Extracts relevant information without modifying it
- Context window: Needs at least 256k tokens (handles huge data batches)
- Recommendation: Use models with large context (e.g.,
gpt-4.1)
Utils - Outputter
Utils - Outputter
Extracts structured output from final results.
- Formats data into Pydantic models
- Ensures type safety
- Recommendation: Use capable text models
Supported Providers
Configure API keys in your.env file:
.env
Fallback Configuration
The cortex supports fallback models for reliability:Best Practices
Optimize the Cortex
Invest in the best vision model for cortex - it has the biggest impact
Use Fast Models for Planner
Planner and orchestrator donโt need the most powerful models
Large Context for Hopper
Ensure hopper has at least 256k context window
Test Profile Performance
Benchmark different profile configurations for your use cases