- Platform (Recommended)
- Local Development
With the platform, tasks are configured on platform.minitap.ai and executed using
PlatformTaskRequest.- Define tasks once on the platform
- Update prompts without code changes
- Built-in observability and cost tracking
Task Characteristics
Goal-based
Define what you want using natural language
Traceable
Record execution for debugging and visualization
Structured Output
Return typed Pydantic models
Platform Tasks
Using the platform? Create tasks on platform.minitap.ai/tasks and execute them with
PlatformTaskRequest.Creating Platform Tasks
- Go to Tasks on the platform
- Click Create Task
- Configure task details:
- Name: Unique identifier (e.g.,
check-notifications) - Agent Prompt: Detailed instructions
- Output Description: Optional structured output format
- Name: Unique identifier (e.g.,
- Use in your code:
Platform Task with Structured Output
Platform Task Benefits
Centralized Management
Update task prompts on the platform without redeploying code
Built-in Observability
View execution details, costs, and agent thoughts on the platform
Team Collaboration
Share tasks across your organization
Version Control
Track changes to task configurations over time
Local Tasks
For local development, define tasks directly in code:Simple String Output
The most basic way to run a local task:Structured Output with Pydantic
Get type-safe, validated output:With Output Description
Provide guidance for unstructured output:Task Options
Naming Tasks
Give your tasks descriptive names for logging:Using Different Profiles
Switch agent profiles for specific tasks:Maximum Steps
Control how many actions the agent can take:Task Builder Pattern
For advanced configuration, use the builder pattern:Tracing and Debugging
Enable trace recording to capture screenshots and execution steps:Traces include screenshots at each step, making it easy to debug failed tasks.
Saving Output
Save LLM Output
Save the final LLM response to a file:Save Agent Thoughts
Capture the agent’s reasoning process:Complex Output Structures
Define complex, nested output structures:Task Execution Flow
1
Goal Analysis
LLM analyzes the goal and creates a plan
2
Screen Observation
Agent captures current screen state
3
Action Decision
LLM decides next action based on goal and screen
4
Action Execution
Hardware bridge performs the action
5
Verification
Agent checks if goal is achieved
6
Output Extraction
If specified, extract structured output
Best Practices
Be specific in your goals
Be specific in your goals
Use Pydantic for structured output
Use Pydantic for structured output
Define clear field descriptions to help the LLM understand what to extract
Break complex tasks into simpler ones
Break complex tasks into simpler ones
Instead of one complex task, run multiple simpler tasks in sequence
Enable tracing for debugging
Enable tracing for debugging
Always enable tracing when developing or debugging tasks
Example: Multi-Step Workflow
- Platform
- Local