Conversation
| Now synthesize the findings from multiple junior workers (LLMs). | ||
| Your task is to finalize an answer to the question below **if and only if** you have sufficient, reliable information. | ||
| Now synthesize the findings from multiple junior workers (LLMs). | ||
| Your task is to finalize an answer to the question below **if and only if** you have sufficient, reliable information. |
There was a problem hiding this comment.
there's a lot of noise in this diff because my code editor removes trailing whitespace. (the main change is lines 747-58).
happy to revert this noise for easier code review; if we'd like to keep it in just to clean things up, that's fine too.
minions/minions_deep_research.py
Outdated
| local_usage = Usage() | ||
|
|
||
| # 1. [REMOTE] CONTEXT --- Read the query with big model and generate web-search context | ||
| web_preview_client = OpenAIClient(model_name="gpt-4o-mini", |
There was a problem hiding this comment.
doesn't seem like GPT-4.5 preview supports web search anymore (https://discord.com/channels/@me/1351994271699566622/1351994706988498986)
ResultsQuery: Can you explain how Anthropic's MCP works? Query: What are the principles of GRPO? |
| if self.callback: | ||
| self.callback("supervisor", None, is_final=False) | ||
|
|
||
| advice_response, usage = self.remote_search_client.chat( |
There was a problem hiding this comment.
could perhaps use the non-search model here for advice if search is extra-expensive
danbider
left a comment
There was a problem hiding this comment.
this looks good. we acknowledge that the task was straightforward tough.
@ayush-alag how would you see a complete version of deep research running local? what will the remote do and what will the local model do?
just getting the docs from remote doesn't mean we are doing deep research. for deep research, I'd expect running multiple queries to grab context, analyzing them locally, sending the results back up to remote, and iterating. i.e., local model does document processing, and remote model does reasoning. but each deep research might have many minion calls on many retrieved documents, interleaved by remote reasoning.
|
@danbider great point. I'd envision something similar to what you mentioned, where we could have a remote search model, a remote reasoning model, and many local minions.
Repeat steps 3-5 as necessary. Again, in keeping with the general Minions philosophy, we avoid having the remote models process long-context documents; these are handled by the local models. Happy to discuss this more in depth or make any changes to the current PR! |
This feature implements a "deep research" version of minions that generates the context for a query from a GPT-4o web search. The minions protocol of using this context for generating sub-tasks that are handled by local models is then used.
Example 1 (MCP)