|
@@ -85,7 +85,7 @@ class LLMClient(ABC):
|
|
|
match = re.search(self.final_answer_pattern, result)
|
|
|
if not match:
|
|
|
if self.session.answer_interpreter:
|
|
|
- interpreter_client = get_client(self.session.answer_interpreter.llm_model)
|
|
|
+ interpreter_client = get_client(self.session.answer_interpreter.llm_model, self.session)
|
|
|
question = self.session.answer_interpreter.prompt.replace("$QUESTION", self.stats['answer'])
|
|
|
result = interpreter_client.send_question(question)
|
|
|
self.stats['answer'] += f"\nInterpreter: {result}"
|
|
@@ -213,7 +213,6 @@ class OllamaClient(LLMClient):
|
|
|
self.stats['instruction'] = self.messages_2_instruction(messages)
|
|
|
|
|
|
model_parameters = self.get_chat_params()
|
|
|
-
|
|
|
response = self.client.chat(model=self.model, messages=messages, **model_parameters)
|
|
|
|
|
|
self.stats['answer'] = response['message']['content']
|