* Added a **pyproject.toml** file to define project metadata and dependencies. * Added **run\_maestro.py** and **osworld\_run\_maestro.py** to provide the main execution logic. * Introduced multiple new modules, including **Evaluator**, **Controller**, **Manager**, and **Sub-Worker**, supporting task planning, state management, and data analysis. * Added a **tools module** containing utility functions and tool configurations to improve code reusability. * Updated the **README** and documentation with usage examples and module descriptions. These changes lay the foundation for expanding the Maestro project’s functionality and improving the user experience. Co-authored-by: Hiroid <guoliangxuan@deepmatrix.com>
386 lines
8.5 KiB
Markdown
386 lines
8.5 KiB
Markdown
# Supported Model Providers and Model Lists
|
|
|
|
## LLM Model Providers
|
|
|
|
### 1. OpenAI
|
|
|
|
**Provider**
|
|
|
|
- `openai`
|
|
|
|
**Supported Models:**
|
|
|
|
- `gpt-5` Window: 400,000 Max Output Tokens: 128,000
|
|
- `gpt-5-mini` Window: 400,000 Max Output Tokens: 128,000
|
|
- `gpt-4.1-nano` Window: 400,000 Max Output Tokens: 128,000
|
|
- `gpt-4.1` Window: 1,047,576 Max Output Tokens: 32,768
|
|
- `gpt-4.1-mini` Window: 1,047,576 Max Output Tokens: 32,768
|
|
- `gpt-4.1-nano` Window: 1,047,576 Max Output Tokens: 32,768
|
|
- `gpt-4o` Window: 128,000 Max Output Tokens: 16,384
|
|
- `gpt-4o-mini` Window: 128,000 Max Output Tokens: 16,384
|
|
- `o1` Window: 200,000 Max Output Tokens: 100,000
|
|
- `o1-pro` Window: 200,000 Max Output Tokens: 100,000
|
|
- `o1-mini` Window: 200,000 Max Output Tokens: 100,000
|
|
- `o3` Window: 200,000 Max Output Tokens: 100,000
|
|
- `o3-pro` Window: 200,000 Max Output Tokens: 100,000
|
|
- `o3-mini` Window: 200,000 Max Output Tokens: 100,000
|
|
- `o4-mini` Window: 200,000 Max Output Tokens: 100,000
|
|
|
|
**Embedding Models:**
|
|
|
|
- `text-embedding-3-small`
|
|
- `text-embedding-3-large`
|
|
- `text-embedding-ada-002`
|
|
|
|
📚 **Reference Link:** <https://platform.openai.com/docs/pricing>
|
|
|
|
---
|
|
|
|
### 2. Anthropic Claude
|
|
|
|
**Provider**
|
|
|
|
- `anthropic`
|
|
|
|
**Supported Models:**
|
|
|
|
- `claude-opus-4-1-20250805` Context window: 200K Max output: 32000
|
|
- `claude-opus-4-20250514` Context window: 200K Max output: 32000
|
|
- `claude-sonnet-4-20250514` Context window: 200K Max output: 64000
|
|
- `claude-3-7-sonnet-20250219` Context window: 200K Max output: 64000
|
|
- - `claude-3-5-sonnet-20240620` Context window: 200K Max output: 64000
|
|
- `claude-3-5-haiku-20241022` Context window: 200K Max output: 8192
|
|
|
|
📚 **Reference Link:** <https://www.anthropic.com/api>
|
|
|
|
---
|
|
|
|
### 3. AWS Bedrock
|
|
|
|
**Provider**
|
|
|
|
- `bedrock`
|
|
|
|
|
|
**Supported Claude Models:**
|
|
|
|
- `Claude-Opus-4`
|
|
- `Claude-Sonnet-4`
|
|
- `Claude-Sonnet-3.7`
|
|
- `Claude-Sonnet-3.5`
|
|
|
|
📚 **Reference Link:** <https://aws.amazon.com/bedrock/>
|
|
|
|
---
|
|
|
|
### 4. Google Gemini
|
|
|
|
**Provider**
|
|
|
|
- `gemini`
|
|
|
|
**Supported Models:**
|
|
|
|
- `gemini-2.5-pro` in: 1,048,576 out: 65536
|
|
- `gemini-2.5-flash` in: 1,048,576 out: 65536
|
|
- `gemini-2.0-flash` in: 1,048,576 out: 8192
|
|
- `gemini-1.5-pro` in: 2,097,152 out: 8192
|
|
- `gemini-1.5-flash` in: 1,048,576 out: 8192
|
|
|
|
**Embedding Models:**
|
|
|
|
- `gemini-embedding-001`
|
|
|
|
📚 **Reference Link:** <https://ai.google.dev/gemini-api/docs/pricing>
|
|
|
|
---
|
|
|
|
### 5. Groq
|
|
|
|
**Provider**
|
|
|
|
- `groq`
|
|
|
|
**Supported Models:**
|
|
|
|
- `Kimi-K2-Instruct`
|
|
- `Llama-4-Scout-17B-16E-Instruct`
|
|
- `Llama-4-Maverick-17B-128E-Instruct`
|
|
- `Llama-Guard-4-12B`
|
|
- `DeepSeek-R1-Distill-Llama-70B`
|
|
- `Qwen3-32B`
|
|
- `Llama-3.3-70B-Instruct`
|
|
|
|
📚 **Reference Link:** <https://groq.com/pricing>
|
|
|
|
---
|
|
|
|
### 6. Monica (Proxy Platform)
|
|
|
|
**Provider**
|
|
|
|
- `monica`
|
|
|
|
**OpenAI Models:**
|
|
|
|
- `gpt-4.1`
|
|
- `gpt-4.1-mini`
|
|
- `gpt-4.1-nano`
|
|
- `gpt-4o-2024-11-20`
|
|
- `gpt-4o-mini-2024-07-18`
|
|
- `o4-mini`
|
|
- `o3`
|
|
|
|
**Anthropic Claude Models:**
|
|
|
|
- `claude-opus-4-20250514`
|
|
- `claude-sonnet-4-20250514`
|
|
- `claude-3-7-sonnet-latest`
|
|
- `claude-3-5-sonnet-20241022`
|
|
- `claude-3-5-sonnet-20240620`
|
|
- `claude-3-5-haiku-20241022`
|
|
|
|
|
|
**Google Gemini Models:**
|
|
|
|
- `gemini-2.5-pro-preview-03-25`
|
|
- `gemini-2.5-flash-lite`
|
|
- `gemini-2.5-flash-preview-05-20`
|
|
- `gemini-2.0-flash-001`
|
|
- `gemini-1.5-pro-002`
|
|
- `gemini-1.5-flash-002`
|
|
|
|
**DeepSeek Models:**
|
|
|
|
- `deepseek-reasoner`
|
|
- `deepseek-chat`
|
|
|
|
**Meta Llama Models:**
|
|
|
|
- `Llama-4-Scout-17B-16E-Instruct` Context length: 10M tokens
|
|
- `Llama-4-Maverick-17B-128E-Instruct ` Context length: 1M tokens
|
|
- `llama-3.3-70b-instruct`
|
|
- `llama-3-70b-instruct`
|
|
- `llama-3.1-405b-instruct`
|
|
|
|
**xAI Grok Models:**
|
|
|
|
- `grok-3-beta`
|
|
- `grok-beta`
|
|
|
|
📚 **Reference Link:** <https://platform.monica.im/docs/en/models-and-pricing>
|
|
|
|
---
|
|
|
|
### 7. OpenRouter (Proxy Platform)
|
|
|
|
**Provider**
|
|
|
|
- `openrouter`
|
|
|
|
**OpenAI Models:**
|
|
|
|
- `gpt-4.1`
|
|
- `gpt-4.1-mini`
|
|
- `o1`
|
|
- `o1-pro`
|
|
- `o1-mini`
|
|
- `o3`
|
|
- `o3-pro`
|
|
- `o3-mini`
|
|
- `o4-mini`
|
|
|
|
**xAI Grok Models:**
|
|
|
|
- `grok-4` Total Context: 256K Max Output: 256K
|
|
- `grok-3`
|
|
- `grok-3-mini`
|
|
|
|
**Anthropic Claude Models:**
|
|
|
|
- `claude-opus-4`
|
|
- `claude-sonnet-4`
|
|
|
|
**Google Gemini Models:**
|
|
|
|
- `gemini-2.5-flash`
|
|
- `gemini-2.5-pro`
|
|
|
|
📚 **Reference Link:** <https://openrouter.ai/models>
|
|
|
|
---
|
|
|
|
### 8. Azure OpenAI
|
|
|
|
**Provider**
|
|
|
|
- `azure`
|
|
|
|
|
|
**Supported Models:**
|
|
|
|
- `gpt-4.1`
|
|
- `gpt-4.1-mini`
|
|
- `gpt-4.1-nano`
|
|
- `o1`
|
|
- `o3`
|
|
- `o4-mini`
|
|
|
|
📚 **Reference Link:** <https://azure.microsoft.com/en-us/pricing/details/cognitive-services/openai-service/>
|
|
|
|
---
|
|
|
|
### 9. Lybic AI
|
|
|
|
**Provider:**
|
|
|
|
- `lybic`
|
|
|
|
**Supported Models:**
|
|
|
|
- `gpt-5`
|
|
- `gpt-4.1`
|
|
- `gpt-4.1-mini`
|
|
- `gpt-4.1-nano`
|
|
- `gpt-4.5-preview`
|
|
- `gpt-4o`
|
|
- `gpt-4o-realtime-preview`
|
|
- `gpt-4o-mini`
|
|
- `o1`
|
|
- `o1-pro`
|
|
- `o1-mini`
|
|
- `o3`
|
|
- `o3-pro`
|
|
- `o3-mini`
|
|
- `o4-mini`
|
|
|
|
**Note:** Lybic AI provides OpenAI-compatible API endpoints with the same model names and pricing structure.
|
|
|
|
📚 **Reference Link:** <https://aigw.lybicai.com/>
|
|
|
|
---
|
|
|
|
### 10. DeepSeek
|
|
|
|
**Provider**
|
|
|
|
- `deepseek`
|
|
|
|
**Supported Models:**
|
|
|
|
- `deepseek-chat` Context length: 128K, Output length: Default 4K, Max 8K
|
|
- `deepseek-reasoner` Context length: 128K, Output length: Default 32K, Max 64K
|
|
|
|
📚 **Reference Link:** <https://platform.deepseek.com/>
|
|
|
|
---
|
|
|
|
### 11. Alibaba Cloud Qwen
|
|
|
|
**Supported Models:**
|
|
|
|
- `qwen-max-latest` Context window: 32,768 Max input token length: 30,720 Max generation token length: 8,192
|
|
- `qwen-plus-latest` Context window: 131,072 Max input token length: 98,304 (thinking) Max generation token length: 129,024 Max output: 16,384
|
|
- `qwen-turbo-latest` Context window: 1,000,000 Max input token length: 1,000,000 Max generation token length: 16,384
|
|
- `qwen-vl-max-latest` (Grounding) Context window: 131,072 Max input token length: 129,024 Max generation token length: 8,192
|
|
- `qwen-vl-plus-latest` (Grounding) Context window: 131,072 Max input token length: 129,024 Max generation token length: 8,192
|
|
|
|
**Embedding Models:**
|
|
|
|
- `text-embedding-v4`
|
|
- `text-embedding-v3`
|
|
|
|
📚 **Reference Link:** <https://bailian.console.aliyun.com/?tab=doc#/doc/?type=model&url=https%3A%2F%2Fhelp.aliyun.com%2Fdocument_detail%2F2840914.html&renderType=iframe>
|
|
|
|
---
|
|
|
|
### 12. ByteDance Doubao
|
|
|
|
**Supported Models:**
|
|
|
|
- `doubao-seed-1-6-flash-250615` Context window: 256k Max input token length: 224k Max generation token length: 32k Max thinking content token length: 32k
|
|
- `doubao-seed-1-6-thinking-250715` Context window: 256k Max input token length: 224k Max generation token length: 32k Max thinking content token length: 32k
|
|
- `doubao-seed-1-6-250615` Context window: 256k Max input token length: 224k Max generation token length: 32k Max thinking content token length: 32k
|
|
- `doubao-1.5-vision-pro-250328` (Grounding) Context window: 128k Max input token length: 96k Max generation token length: 16k Max thinking content token length: 32k
|
|
- `doubao-1-5-thinking-vision-pro-250428` (Grounding) Context window: 128k Max input token length: 96k Max generation token length: 16k Max thinking content token length: 32k
|
|
- `doubao-1-5-ui-tars-250428` (Grounding) Context window: 128k Max input token length: 96k Max generation token length: 16k Max thinking content token length: 32k
|
|
|
|
**Embedding Models:**
|
|
|
|
- `doubao-embedding-large-text-250515`
|
|
- `doubao-embedding-text-240715`
|
|
|
|
📚 **Reference Link:** <https://console.volcengine.com/ark/region:ark+cn-beijing/model?vendor=Bytedance&view=LIST_VIEW>
|
|
|
|
---
|
|
|
|
### 13. Zhipu GLM
|
|
|
|
**Supported Models:**
|
|
|
|
- `GLM-4.5` Max in: 128k Max output: 0.2K
|
|
- `GLM-4.5-X` Max in: 128k Max output: 0.2K
|
|
- `GLM-4.5-Air` Max in: 128k Max output: 0.2K
|
|
- `GLM-4-Plus`
|
|
- `GLM-4-Air-250414`
|
|
- `GLM-4-AirX` (Grounding)
|
|
- `GLM-4V-Plus-0111` (Grounding)
|
|
|
|
**Embedding Models:**
|
|
|
|
- `Embedding-3`
|
|
- `Embedding-2`
|
|
|
|
📚 **Reference Link:** <https://open.bigmodel.cn/pricing>
|
|
|
|
---
|
|
|
|
### 14. SiliconFlow
|
|
|
|
**Supported Models:**
|
|
|
|
- `Kimi-K2-Instruct` Context Length: 128K
|
|
- `DeepSeek-V3`
|
|
- `DeepSeek-R1`
|
|
- `Qwen3-32B`
|
|
|
|
📚 **Reference Link:** <https://cloud.siliconflow.cn/sft-d1pi8rbk20jc73c62gm0/models>
|
|
|
|
---
|
|
|
|
## 🔤 Dedicated Embedding Providers
|
|
|
|
### 15. Jina AI
|
|
|
|
**Embedding Models:**
|
|
|
|
- `jina-embeddings-v4`
|
|
- `jina-embeddings-v3`
|
|
|
|
📚 **Reference Link:** <https://jina.ai/embeddings>
|
|
|
|
---
|
|
|
|
## 🔍 AI Search Engines
|
|
|
|
### 16. Bocha AI
|
|
|
|
**Service Type:** AI Research & Search
|
|
|
|
📚 **Reference Link:** <https://open.bochaai.com/overview>
|
|
|
|
---
|
|
|
|
### 17. Exa
|
|
|
|
**Service Type:** AI Research & Search
|
|
|
|
**Pricing Model:**
|
|
|
|
- $5.00 / 1k agent searches
|
|
- $5.00 / 1k exa-research agent page reads
|
|
- $10.00 / 1k exa-research-pro agent page reads
|
|
- $5.00 / 1M reasoning tokens
|
|
|
|
📚 **Reference Link:** <https://dashboard.exa.ai/home>
|