* Added a **pyproject.toml** file to define project metadata and dependencies. * Added **run\_maestro.py** and **osworld\_run\_maestro.py** to provide the main execution logic. * Introduced multiple new modules, including **Evaluator**, **Controller**, **Manager**, and **Sub-Worker**, supporting task planning, state management, and data analysis. * Added a **tools module** containing utility functions and tool configurations to improve code reusability. * Updated the **README** and documentation with usage examples and module descriptions. These changes lay the foundation for expanding the Maestro project’s functionality and improving the user experience. Co-authored-by: Hiroid <guoliangxuan@deepmatrix.com>
8.5 KiB
Supported Model Providers and Model Lists
LLM Model Providers
1. OpenAI
Provider
openai
Supported Models:
gpt-5Window: 400,000 Max Output Tokens: 128,000gpt-5-miniWindow: 400,000 Max Output Tokens: 128,000gpt-4.1-nanoWindow: 400,000 Max Output Tokens: 128,000gpt-4.1Window: 1,047,576 Max Output Tokens: 32,768gpt-4.1-miniWindow: 1,047,576 Max Output Tokens: 32,768gpt-4.1-nanoWindow: 1,047,576 Max Output Tokens: 32,768gpt-4oWindow: 128,000 Max Output Tokens: 16,384gpt-4o-miniWindow: 128,000 Max Output Tokens: 16,384o1Window: 200,000 Max Output Tokens: 100,000o1-proWindow: 200,000 Max Output Tokens: 100,000o1-miniWindow: 200,000 Max Output Tokens: 100,000o3Window: 200,000 Max Output Tokens: 100,000o3-proWindow: 200,000 Max Output Tokens: 100,000o3-miniWindow: 200,000 Max Output Tokens: 100,000o4-miniWindow: 200,000 Max Output Tokens: 100,000
Embedding Models:
text-embedding-3-smalltext-embedding-3-largetext-embedding-ada-002
📚 Reference Link: https://platform.openai.com/docs/pricing
2. Anthropic Claude
Provider
anthropic
Supported Models:
claude-opus-4-1-20250805Context window: 200K Max output: 32000claude-opus-4-20250514Context window: 200K Max output: 32000claude-sonnet-4-20250514Context window: 200K Max output: 64000claude-3-7-sonnet-20250219Context window: 200K Max output: 64000-
claude-3-5-sonnet-20240620Context window: 200K Max output: 64000
claude-3-5-haiku-20241022Context window: 200K Max output: 8192
📚 Reference Link: https://www.anthropic.com/api
3. AWS Bedrock
Provider
bedrock
Supported Claude Models:
Claude-Opus-4Claude-Sonnet-4Claude-Sonnet-3.7Claude-Sonnet-3.5
📚 Reference Link: https://aws.amazon.com/bedrock/
4. Google Gemini
Provider
gemini
Supported Models:
gemini-2.5-proin: 1,048,576 out: 65536gemini-2.5-flashin: 1,048,576 out: 65536gemini-2.0-flashin: 1,048,576 out: 8192gemini-1.5-proin: 2,097,152 out: 8192gemini-1.5-flashin: 1,048,576 out: 8192
Embedding Models:
gemini-embedding-001
📚 Reference Link: https://ai.google.dev/gemini-api/docs/pricing
5. Groq
Provider
groq
Supported Models:
Kimi-K2-InstructLlama-4-Scout-17B-16E-InstructLlama-4-Maverick-17B-128E-InstructLlama-Guard-4-12BDeepSeek-R1-Distill-Llama-70BQwen3-32BLlama-3.3-70B-Instruct
📚 Reference Link: https://groq.com/pricing
6. Monica (Proxy Platform)
Provider
monica
OpenAI Models:
gpt-4.1gpt-4.1-minigpt-4.1-nanogpt-4o-2024-11-20gpt-4o-mini-2024-07-18o4-minio3
Anthropic Claude Models:
claude-opus-4-20250514claude-sonnet-4-20250514claude-3-7-sonnet-latestclaude-3-5-sonnet-20241022claude-3-5-sonnet-20240620claude-3-5-haiku-20241022
Google Gemini Models:
gemini-2.5-pro-preview-03-25gemini-2.5-flash-litegemini-2.5-flash-preview-05-20gemini-2.0-flash-001gemini-1.5-pro-002gemini-1.5-flash-002
DeepSeek Models:
deepseek-reasonerdeepseek-chat
Meta Llama Models:
Llama-4-Scout-17B-16E-InstructContext length: 10M tokensLlama-4-Maverick-17B-128E-InstructContext length: 1M tokensllama-3.3-70b-instructllama-3-70b-instructllama-3.1-405b-instruct
xAI Grok Models:
grok-3-betagrok-beta
📚 Reference Link: https://platform.monica.im/docs/en/models-and-pricing
7. OpenRouter (Proxy Platform)
Provider
openrouter
OpenAI Models:
gpt-4.1gpt-4.1-minio1o1-proo1-minio3o3-proo3-minio4-mini
xAI Grok Models:
grok-4Total Context: 256K Max Output: 256Kgrok-3grok-3-mini
Anthropic Claude Models:
claude-opus-4claude-sonnet-4
Google Gemini Models:
gemini-2.5-flashgemini-2.5-pro
📚 Reference Link: https://openrouter.ai/models
8. Azure OpenAI
Provider
azure
Supported Models:
gpt-4.1gpt-4.1-minigpt-4.1-nanoo1o3o4-mini
📚 Reference Link: https://azure.microsoft.com/en-us/pricing/details/cognitive-services/openai-service/
9. Lybic AI
Provider:
lybic
Supported Models:
gpt-5gpt-4.1gpt-4.1-minigpt-4.1-nanogpt-4.5-previewgpt-4ogpt-4o-realtime-previewgpt-4o-minio1o1-proo1-minio3o3-proo3-minio4-mini
Note: Lybic AI provides OpenAI-compatible API endpoints with the same model names and pricing structure.
📚 Reference Link: https://aigw.lybicai.com/
10. DeepSeek
Provider
deepseek
Supported Models:
deepseek-chatContext length: 128K, Output length: Default 4K, Max 8Kdeepseek-reasonerContext length: 128K, Output length: Default 32K, Max 64K
📚 Reference Link: https://platform.deepseek.com/
11. Alibaba Cloud Qwen
Supported Models:
qwen-max-latestContext window: 32,768 Max input token length: 30,720 Max generation token length: 8,192qwen-plus-latestContext window: 131,072 Max input token length: 98,304 (thinking) Max generation token length: 129,024 Max output: 16,384qwen-turbo-latestContext window: 1,000,000 Max input token length: 1,000,000 Max generation token length: 16,384qwen-vl-max-latest(Grounding) Context window: 131,072 Max input token length: 129,024 Max generation token length: 8,192qwen-vl-plus-latest(Grounding) Context window: 131,072 Max input token length: 129,024 Max generation token length: 8,192
Embedding Models:
text-embedding-v4text-embedding-v3
📚 Reference Link: https://bailian.console.aliyun.com/?tab=doc#/doc/?type=model&url=https%3A%2F%2Fhelp.aliyun.com%2Fdocument_detail%2F2840914.html&renderType=iframe
12. ByteDance Doubao
Supported Models:
doubao-seed-1-6-flash-250615Context window: 256k Max input token length: 224k Max generation token length: 32k Max thinking content token length: 32kdoubao-seed-1-6-thinking-250715Context window: 256k Max input token length: 224k Max generation token length: 32k Max thinking content token length: 32kdoubao-seed-1-6-250615Context window: 256k Max input token length: 224k Max generation token length: 32k Max thinking content token length: 32kdoubao-1.5-vision-pro-250328(Grounding) Context window: 128k Max input token length: 96k Max generation token length: 16k Max thinking content token length: 32kdoubao-1-5-thinking-vision-pro-250428(Grounding) Context window: 128k Max input token length: 96k Max generation token length: 16k Max thinking content token length: 32kdoubao-1-5-ui-tars-250428(Grounding) Context window: 128k Max input token length: 96k Max generation token length: 16k Max thinking content token length: 32k
Embedding Models:
doubao-embedding-large-text-250515doubao-embedding-text-240715
📚 Reference Link: https://console.volcengine.com/ark/region:ark+cn-beijing/model?vendor=Bytedance&view=LIST_VIEW
13. Zhipu GLM
Supported Models:
GLM-4.5Max in: 128k Max output: 0.2KGLM-4.5-XMax in: 128k Max output: 0.2KGLM-4.5-AirMax in: 128k Max output: 0.2KGLM-4-PlusGLM-4-Air-250414GLM-4-AirX(Grounding)GLM-4V-Plus-0111(Grounding)
Embedding Models:
Embedding-3Embedding-2
📚 Reference Link: https://open.bigmodel.cn/pricing
14. SiliconFlow
Supported Models:
Kimi-K2-InstructContext Length: 128KDeepSeek-V3DeepSeek-R1Qwen3-32B
📚 Reference Link: https://cloud.siliconflow.cn/sft-d1pi8rbk20jc73c62gm0/models
🔤 Dedicated Embedding Providers
15. Jina AI
Embedding Models:
jina-embeddings-v4jina-embeddings-v3
📚 Reference Link: https://jina.ai/embeddings
🔍 AI Search Engines
16. Bocha AI
Service Type: AI Research & Search
📚 Reference Link: https://open.bochaai.com/overview
17. Exa
Service Type: AI Research & Search
Pricing Model:
- $5.00 / 1k agent searches
- $5.00 / 1k exa-research agent page reads
- $10.00 / 1k exa-research-pro agent page reads
- $5.00 / 1M reasoning tokens
📚 Reference Link: https://dashboard.exa.ai/home