后端添加cache;前端改动agents
This commit is contained in:
@@ -2,9 +2,9 @@ from autogen.coding import LocalCommandLineCodeExecutor
|
||||
from autogen.agentchat.contrib.multimodal_conversable_agent import MultimodalConversableAgent
|
||||
import autogen
|
||||
from .utils import get_max_uv_wavelength_from_txt, get_max_pl_peak_from_txt
|
||||
from .constant import config_list, STREAM, SILENT, WORK_DIR
|
||||
from .constant import config_list, STREAM, SILENT, WORK_DIR, CACHE
|
||||
|
||||
llm_config = {"config_list": config_list, "stream": STREAM}
|
||||
llm_config = {"config_list": config_list, "stream": STREAM, "cache_seed": CACHE}
|
||||
|
||||
def init_optimize_group(work_dir, venv_context):
|
||||
optimize_env = LocalCommandLineCodeExecutor(
|
||||
|
||||
Reference in New Issue
Block a user