From 4463b8fd7e0bcf2f5bf0f895bafec0b11a72e695 Mon Sep 17 00:00:00 2001 From: Yuanwen Tian <140994104+yuanwen-tian@users.noreply.github.com> Date: Tue, 24 Sep 2024 22:28:42 +0800 Subject: [PATCH] fix: add func name to florencev2 fine tune (#248) * fix: add func name to florencev2 fine tune * fix * fix: format * remove extra code --------- Co-authored-by: wuyiqunLu --- vision_agent/agent/vision_agent.py | 19 +++++-------------- vision_agent/tools/tools.py | 7 ++++++- 2 files changed, 11 insertions(+), 15 deletions(-) diff --git a/vision_agent/agent/vision_agent.py b/vision_agent/agent/vision_agent.py index b54d08b8..bf35e5e9 100644 --- a/vision_agent/agent/vision_agent.py +++ b/vision_agent/agent/vision_agent.py @@ -30,12 +30,6 @@ if str(WORKSPACE) != "": os.environ["PYTHONPATH"] = f"{WORKSPACE}:{os.getenv('PYTHONPATH', '')}" -STUCK_IN_LOOP_ERROR_MESSAGE = { - "name": "Error when running conversation agent", - "value": "Agent is stuck in conversation loop, exited", - "traceback_raw": [], -} - class BoilerplateCode: pre_code = [ @@ -298,13 +292,6 @@ def chat_with_code( # sometimes it gets stuck in a loop, so we force it to exit if last_response == response: response["let_user_respond"] = True - self.streaming_message( - { - "role": "assistant", - "content": "{}", - "error": STUCK_IN_LOOP_ERROR_MESSAGE, - } - ) finished = response["let_user_respond"] @@ -317,7 +304,11 @@ def chat_with_code( { "role": "assistant", "content": "{}", - "error": STUCK_IN_LOOP_ERROR_MESSAGE, + "error": { + "name": "Error when running conversation agent", + "value": "Agent is stuck in conversation loop, exited", + "traceback_raw": [], + }, "finished": finished and code_action is None, } ) diff --git a/vision_agent/tools/tools.py b/vision_agent/tools/tools.py index fca3819c..34c037ab 100644 --- a/vision_agent/tools/tools.py +++ b/vision_agent/tools/tools.py @@ -1181,7 +1181,12 @@ def florence2_phrase_grounding( fine_tuning=FineTuning(job_id=UUID(fine_tune_id)), ) data = data_obj.model_dump(by_alias=True) - detections = send_inference_request(data, "tools", v2=False) + detections = send_inference_request( + data, + "tools", + v2=False, + metadata_payload={"function_name": "florence2_phrase_grounding"}, + ) else: data = { "image": image_b64,