- Platform (Recommended)
- Local Development
With the platform, profiles are configured on platform.minitap.ai and referenced by name in your code.
- No config files needed
- Update models without code changes
- Access to all OpenRouter models
- Share profiles across your team
What are Agent Profiles?
An agent profile defines the LLM configuration for the various sub-agents that power mobile-use:- Planner - Creates high-level plans from goals
- Orchestrator - Coordinates execution steps
- Contextor - Gathers device context and enforces app lock constraints (optional, conditionally invokes LLM)
- Cortex - High-level reasoning and decision-making (requires vision capabilities)
- Executor - Performs specific actions
- Utils - Helper agents (outputter, hopper)
Platform Profiles
Using the platform? Create and manage profiles on platform.minitap.ai/llm-profiles, then reference them by name in your tasks.
Creating Platform Profiles
- Go to LLM Profiles on the platform
- Click Create Profile
- Configure each agent component with your preferred models
- All OpenRouter models are available (no API key management needed)
- Reference the profile by name in your code:
Local Profiles
For local development, profiles are defined in config files or code:From Configuration File
The recommended approach for local production:Programmatic Configuration
For dynamic configuration:Using Profiles
Setting a Default Profile
Configure an agent with a default profile:Multiple Profiles
Register multiple profiles and switch between them:Profile Use Cases
- Speed vs Accuracy
- Cost Optimization
- Provider Diversity
Fast Profile - Quick tasks, simple UI navigationAccurate Profile - Complex analysis, detailed extraction
Component Roles
Contextor (App Lock Enforcement)
Contextor (App Lock Enforcement)
Gathers device context and enforces app lock constraints.
- Always collects current device state (UI hierarchy, focused app, screen dimensions, date/time)
- When app lock is enabled: validates user remains in specified app
- Decides if app deviations are intentional (e.g., OAuth flows)
- Relaunches the app when necessary
- Model Type: Non-vision (text-only)
- Invokes LLM: Only when app lock is active AND app mismatch detected
- Recommendation: Use fast, decision-focused models (e.g.,
gpt-5-nano,meta-llama/llama-3.1-8b-instruct) - Context window: Standard (does not need extended context)
The Contextor is optional and only needed if you plan to use the app lock feature. When disabled or when thereโs no app lock, it functions as a simple data-gathering node without LLM inference.
Cortex (High-Level Reasoning)
Cortex (High-Level Reasoning)
The cortex handles high-level reasoning and decision-making.
- Makes strategic decisions about task execution
- Coordinates between planner and executor
- Handles complex reasoning tasks
- Analyzes both UI hierarchy and screenshots for complete device understanding
- IMPORTANT: Requires a vision-capable model (e.g.,
gpt-4o,gpt-5) - Context window: Needs at least 128k tokens and image input support
Planner
Planner
Creates high-level plans from natural language goals.
- Breaks down goals into subgoals
- Estimates complexity
- Recommendation: Use fast, capable text models (e.g.,
gpt-5-nano)
Orchestrator
Orchestrator
Coordinates execution and decides when to replicate.
- Manages task flow
- Handles errors and retries
- Recommendation: Use fast models (e.g.,
gpt-5-nano)
Executor
Executor
Translates decisions into device actions.
- Generates Maestro commands
- Handles action formatting
- Recommendation: Use fast, instruction-following models
Utils - Hopper
Utils - Hopper
Digs through large batches of data to extract the most relevant information for reaching the goal.
- Processes extensive historical context and screen data
- Extracts relevant information without modifying it
- Context window: Needs at least 256k tokens (handles huge data batches)
- Recommendation: Use models with large context (e.g.,
gpt-4.1)
Utils - Outputter
Utils - Outputter
Extracts structured output from final results.
- Formats data into Pydantic models
- Ensures type safety
- Recommendation: Use capable text models
Supported Providers
Configure API keys in your.env file:
.env
Fallback Configuration
The cortex supports fallback models for reliability:Best Practices
Optimize the Cortex
Invest in the best vision-capable model for cortex - it has the biggest impact
Use Fast Models for Planner
Planner and orchestrator donโt need the most powerful models
Large Context for Hopper
Ensure hopper has at least 256k context window
Test Profile Performance
Benchmark different profile configurations for your use cases