Update claude endpoint
This commit is contained in:
@@ -15,7 +15,6 @@ import dashscope
|
|||||||
import google.generativeai as genai
|
import google.generativeai as genai
|
||||||
import openai
|
import openai
|
||||||
import requests
|
import requests
|
||||||
import wandb
|
|
||||||
from PIL import Image
|
from PIL import Image
|
||||||
from google.api_core.exceptions import InvalidArgument
|
from google.api_core.exceptions import InvalidArgument
|
||||||
|
|
||||||
@@ -523,20 +522,30 @@ class PromptAgent:
|
|||||||
claude_messages[1]['content'].insert(0, claude_system_message_item)
|
claude_messages[1]['content'].insert(0, claude_system_message_item)
|
||||||
claude_messages.pop(0)
|
claude_messages.pop(0)
|
||||||
|
|
||||||
|
# headers = {
|
||||||
|
# "x-api-key": os.environ["ANTHROPIC_API_KEY"],
|
||||||
|
# "anthropic-version": "2023-06-01",
|
||||||
|
# "content-type": "application/json"
|
||||||
|
# }
|
||||||
|
|
||||||
headers = {
|
headers = {
|
||||||
"x-api-key": os.environ["ANTHROPIC_API_KEY"],
|
"Accept": "application / json",
|
||||||
"anthropic-version": "2023-06-01",
|
"Authorization": "Bearer " + os.environ["ANTHROPIC_API_KEY"],
|
||||||
"content-type": "application/json"
|
"User-Agent": "Apifox/1.0.0 (https://apifox.com)",
|
||||||
|
"Content-Type": "application/json"
|
||||||
}
|
}
|
||||||
|
|
||||||
payload = {
|
payload = {
|
||||||
"model": self.model,
|
"model": self.model,
|
||||||
"max_tokens": max_tokens,
|
"max_tokens": max_tokens,
|
||||||
"messages": claude_messages
|
"messages": claude_messages,
|
||||||
|
"temperature": temperature,
|
||||||
|
"top_p": top_p
|
||||||
}
|
}
|
||||||
|
|
||||||
response = requests.post(
|
response = requests.post(
|
||||||
"https://api.anthropic.com/v1/messages",
|
# "https://chat.claude.com/v1/chat/completions",
|
||||||
|
"https://api.aigcbest.top/v1/chat/completions",
|
||||||
headers=headers,
|
headers=headers,
|
||||||
json=payload
|
json=payload
|
||||||
)
|
)
|
||||||
@@ -546,8 +555,10 @@ class PromptAgent:
|
|||||||
logger.error("Failed to call LLM: " + response.text)
|
logger.error("Failed to call LLM: " + response.text)
|
||||||
time.sleep(5)
|
time.sleep(5)
|
||||||
return ""
|
return ""
|
||||||
|
# else:
|
||||||
|
# return response.json()['content'][0]['text']
|
||||||
else:
|
else:
|
||||||
return response.json()['content'][0]['text']
|
return response.json()['choices'][0]['message']['content']
|
||||||
|
|
||||||
|
|
||||||
elif self.model.startswith("mistral"):
|
elif self.model.startswith("mistral"):
|
||||||
@@ -568,10 +579,8 @@ class PromptAgent:
|
|||||||
for part in message["content"]:
|
for part in message["content"]:
|
||||||
mistral_message['content'] = part['text'] if part['type'] == "text" else ""
|
mistral_message['content'] = part['text'] if part['type'] == "text" else ""
|
||||||
|
|
||||||
|
|
||||||
misrtal_messages.append(mistral_message)
|
misrtal_messages.append(mistral_message)
|
||||||
|
|
||||||
|
|
||||||
# openai.api_base = "http://localhost:8000/v1"
|
# openai.api_base = "http://localhost:8000/v1"
|
||||||
# response = openai.ChatCompletion.create(
|
# response = openai.ChatCompletion.create(
|
||||||
# messages=misrtal_messages,
|
# messages=misrtal_messages,
|
||||||
@@ -615,7 +624,8 @@ class PromptAgent:
|
|||||||
|
|
||||||
for part in message["content"]:
|
for part in message["content"]:
|
||||||
if part['type'] == "image_url":
|
if part['type'] == "image_url":
|
||||||
cog_message['content'].append({"type": "image_url", "image_url": {"url": part['image_url']['url'] } })
|
cog_message['content'].append(
|
||||||
|
{"type": "image_url", "image_url": {"url": part['image_url']['url']}})
|
||||||
|
|
||||||
if part['type'] == "text":
|
if part['type'] == "text":
|
||||||
cog_message['content'].append({"type": "text", "text": part['text']})
|
cog_message['content'].append({"type": "text", "text": part['text']})
|
||||||
|
|||||||
Reference in New Issue
Block a user