格式转换
This commit is contained in:
@@ -2,22 +2,25 @@
|
||||
api:
|
||||
key: "sk-oYh3Xrhg8oDY2gW02c966f31C84449Ad86F9Cd9dF6E64a8d"
|
||||
base_url: "https://vip.apiyi.com/v1"
|
||||
temperature: 0
|
||||
temperature: -1 # 默认使用模型的温度设置
|
||||
max_retries: 10
|
||||
# 支持多个模型
|
||||
models:
|
||||
- "qwen-max-2025-01-25"
|
||||
- "gpt-4o"
|
||||
- "deepseek-chat"
|
||||
- "claude-sonnet-4-20250514"
|
||||
- "deepseek-r1"
|
||||
# 或者使用单个模型(向后兼容)
|
||||
# model: "qwen-max-2025-01-25"
|
||||
|
||||
# 系统提示词
|
||||
system_prompt: "You are an expert in the field of materials science, adept at answering questions related to fundamental aspects of materials science, including material structure, properties, processing, and applications."
|
||||
system_prompt: None
|
||||
|
||||
# 评估配置
|
||||
evaluation:
|
||||
max_workers: 8
|
||||
input_file: "/home/ubuntu/50T/LYT/MatBench/layer1/ALL-merge/merged.json"
|
||||
max_workers: 20
|
||||
# input_file: "/home/ubuntu/50T/LYT/MatBench/layer1/ALL-merge/merged.json"
|
||||
input_file: "/home/ubuntu/50T/LYT/MatBench/layer2/PGEE/code/stepz_final_choice_questions.json"
|
||||
# 输出配置
|
||||
output:
|
||||
base_dir: "results"
|
||||
|
||||
@@ -144,7 +144,7 @@ def main():
|
||||
logger.info(f"Evaluating model {i}/{len(models)}: {model_name}")
|
||||
|
||||
try:
|
||||
model_result = evaluate_single_model(model_name, data[:10], config, output_dir)
|
||||
model_result = evaluate_single_model(model_name, data, config, output_dir)
|
||||
all_results[model_name] = model_result
|
||||
|
||||
# 打印当前模型的结果
|
||||
|
||||
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
@@ -51,7 +51,7 @@ class Evaluator:
|
||||
|
||||
# 格式化选择项
|
||||
formatted_choices = " ".join([f"({lbl}) {txt}" for lbl, txt in zip(label, text)])
|
||||
user_input = f"{question} {formatted_choices}. {prompt}"
|
||||
user_input = f"{prompt} \n {question} {formatted_choices}"
|
||||
|
||||
# 获取LLM响应
|
||||
llm_answer = self.llm_client.get_response(user_input, self.system_prompt)
|
||||
|
||||
@@ -48,14 +48,27 @@ class LLMClient:
|
||||
retries = 0
|
||||
while retries < self.max_retries:
|
||||
try:
|
||||
response = self.client.chat.completions.create(
|
||||
model=self.model,
|
||||
messages=[
|
||||
if system_prompt == 'None':
|
||||
messages = [
|
||||
{"role": "user", "content": user_input}
|
||||
]
|
||||
else:
|
||||
messages = [
|
||||
{"role": "system", "content": system_prompt},
|
||||
{"role": "user", "content": user_input}
|
||||
],
|
||||
temperature=self.temperature
|
||||
)
|
||||
]
|
||||
|
||||
if self.temperature == -1:
|
||||
response = self.client.chat.completions.create(
|
||||
model=self.model,
|
||||
messages=messages,
|
||||
)
|
||||
else:
|
||||
response = self.client.chat.completions.create(
|
||||
model=self.model,
|
||||
messages=messages,
|
||||
temperature=self.temperature
|
||||
)
|
||||
answer = response.choices[0].message.content
|
||||
return answer
|
||||
|
||||
|
||||
Reference in New Issue
Block a user