From 9f8b7e24ccee6a3b5501454f48d08cc8776b4d53 Mon Sep 17 00:00:00 2001 From: Dillon Laird Date: Mon, 15 Apr 2024 15:35:14 -0700 Subject: [PATCH] fixed test case --- tests/test_llm.py | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/tests/test_llm.py b/tests/test_llm.py index 0a671ca5..bbcc203e 100644 --- a/tests/test_llm.py +++ b/tests/test_llm.py @@ -18,7 +18,7 @@ def test_generate_with_mock(openai_llm_mock): # noqa: F811 response = llm.generate("test prompt") assert response == "mocked response" openai_llm_mock.chat.completions.create.assert_called_once_with( - model="gpt-4-turbo-preview", + model="gpt-4-turbo", messages=[{"role": "user", "content": "test prompt"}], ) @@ -31,7 +31,7 @@ def test_chat_with_mock(openai_llm_mock): # noqa: F811 response = llm.chat([{"role": "user", "content": "test prompt"}]) assert response == "mocked response" openai_llm_mock.chat.completions.create.assert_called_once_with( - model="gpt-4-turbo-preview", + model="gpt-4-turbo", messages=[{"role": "user", "content": "test prompt"}], ) @@ -44,14 +44,14 @@ def test_call_with_mock(openai_llm_mock): # noqa: F811 response = llm("test prompt") assert response == "mocked response" openai_llm_mock.chat.completions.create.assert_called_once_with( - model="gpt-4-turbo-preview", + model="gpt-4-turbo", messages=[{"role": "user", "content": "test prompt"}], ) response = llm([{"role": "user", "content": "test prompt"}]) assert response == "mocked response" openai_llm_mock.chat.completions.create.assert_called_with( - model="gpt-4-turbo-preview", + model="gpt-4-turbo", messages=[{"role": "user", "content": "test prompt"}], )