From 81ef0a863410c21fe97f4abc939dd63f1907dc32 Mon Sep 17 00:00:00 2001 From: Dillon Laird Date: Fri, 9 Aug 2024 08:12:21 -0700 Subject: [PATCH] fixed linting --- vision_agent/agent/vision_agent.py | 2 +- vision_agent/agent/vision_agent_coder.py | 12 ++++++------ vision_agent/lmm/lmm.py | 4 ++-- 3 files changed, 9 insertions(+), 9 deletions(-) diff --git a/vision_agent/agent/vision_agent.py b/vision_agent/agent/vision_agent.py index a39fe208..a41fd09f 100644 --- a/vision_agent/agent/vision_agent.py +++ b/vision_agent/agent/vision_agent.py @@ -63,7 +63,7 @@ def run_conversation(orch: LMM, chat: List[Message]) -> Dict[str, Any]: dir=WORKSPACE, conversation=conversation, ) - return extract_json(orch([{"role": "user", "content": prompt}], stream=False)) # type: ignore + return extract_json(orch([{"role": "user", "content": prompt}], stream=False)) # type: ignore def run_code_action(code: str, code_interpreter: CodeInterpreter) -> str: diff --git a/vision_agent/agent/vision_agent_coder.py b/vision_agent/agent/vision_agent_coder.py index 5a7f9a2e..db13e09a 100644 --- a/vision_agent/agent/vision_agent_coder.py +++ b/vision_agent/agent/vision_agent_coder.py @@ -160,7 +160,7 @@ def pick_plan( docstring=tool_info, plans=plan_str, previous_attempts="", media=media ) - code = extract_code(model(prompt, stream=False)) # type: ignore + code = extract_code(model(prompt, stream=False)) # type: ignore log_progress( { "type": "log", @@ -211,7 +211,7 @@ def pick_plan( "code": DefaultImports.prepend_imports(code), } ) - code = extract_code(model(prompt, stream=False)) # type: ignore + code = extract_code(model(prompt, stream=False)) # type: ignore tool_output = code_interpreter.exec_isolation( DefaultImports.prepend_imports(code) ) @@ -251,7 +251,7 @@ def pick_plan( tool_output=tool_output_str[:20_000], ) chat[-1]["content"] = prompt - best_plan = extract_json(model(chat, stream=False)) # type: ignore + best_plan = extract_json(model(chat, stream=False)) # type: ignore if verbosity >= 1: _LOGGER.info(f"Best plan:\n{best_plan}") @@ -286,7 +286,7 @@ def write_code( feedback=feedback, ) chat[-1]["content"] = prompt - return extract_code(coder(chat, stream=False)) # type: ignore + return extract_code(coder(chat, stream=False)) # type: ignore def write_test( @@ -310,7 +310,7 @@ def write_test( media=media, ) chat[-1]["content"] = prompt - return extract_code(tester(chat, stream=False)) # type: ignore + return extract_code(tester(chat, stream=False)) # type: ignore def write_and_test_code( @@ -439,7 +439,7 @@ def debug_code( while not success and count < 3: try: fixed_code_and_test = extract_json( - debugger( # type: ignore + debugger( # type: ignore FIX_BUG.format( code=code, tests=test, diff --git a/vision_agent/lmm/lmm.py b/vision_agent/lmm/lmm.py index 1f49f8ba..b190144f 100644 --- a/vision_agent/lmm/lmm.py +++ b/vision_agent/lmm/lmm.py @@ -152,7 +152,7 @@ def chat( ) if "stream" in tmp_kwargs and tmp_kwargs["stream"]: for chunk in response: - chunk_message = chunk.choices[0].delta.content # type: ignore + chunk_message = chunk.choices[0].delta.content # type: ignore yield chunk_message else: return cast(str, response.choices[0].message.content) @@ -191,7 +191,7 @@ def generate( ) if "stream" in tmp_kwargs and tmp_kwargs["stream"]: for chunk in response: - chunk_message = chunk.choices[0].delta.content # type: ignore + chunk_message = chunk.choices[0].delta.content # type: ignore yield chunk_message else: return cast(str, response.choices[0].message.content)