diff --git a/tests/test_agents_template.py b/tests/test_agents_template.py index 8ad5dc891..0497283b9 100644 --- a/tests/test_agents_template.py +++ b/tests/test_agents_template.py @@ -137,6 +137,7 @@ def test_initialize_agent(client: TestClient): "self-ask-with-search", "conversational-react-description", "openai-functions", + "openai-multi-functions", ], "name": "agent", "type": "str", diff --git a/tests/test_graph.py b/tests/test_graph.py index 16dd514b8..02bb180f2 100644 --- a/tests/test_graph.py +++ b/tests/test_graph.py @@ -336,7 +336,7 @@ def test_get_result_and_thought(basic_graph): responses = [ "Final Answer: I am a response", ] - message = "Hello" + message = {"input": "Hello"} # Find the node that is an LLMNode and change the # _built_object to a FakeListLLM llm_node = get_node_by_type(basic_graph, LLMVertex) @@ -349,8 +349,5 @@ def test_get_result_and_thought(basic_graph): # now build again and check if FakeListLLM was used # Get the result and thought - result, thought = get_result_and_thought(langchain_object, message) - # The result should be a str - assert isinstance(result, str) - # The thought should be a Thought - assert isinstance(thought, str) + result = get_result_and_thought(langchain_object, message) + assert isinstance(result, dict)