Have you tried iterating on this? Like, the “I don’t care about the word prodrome’” sounds like the kind of thing you could include in your prompt and reiterate until everything you don’t like about the LLM’s responses is solved or you run out of ideas.
Also fyi ChatGPT Deep Research uses the “o3″ model, not 4o, even if it says 4o at the top left (you can try running Deep Research with any of the models selected in the top left and it will output the same kind of thing).
o3 was RLed (!) into being particularly good at web search (and tangential skills like avoiding suspicious links), and isn’t released in a way that lets you just chat with it. The output isn’t even raw o3, it’s the o3-mini model summarizing o3′s chain of thought (where o3 will think things, send a dozen tentacles out into the web, then continue thinking).
I learned this when I asked Deep Research to reverse engineer itself, and it linked the model card which in retrospect I should have done first and was foolish not to.
Anyway I mention this because afaik all the other deep research frameworks are a lot less specialized than OpenAI’s, and more like “we took an LLM and gave it access to the internet and let it think and search for a really long time”. I expect OpenAI to continue being SOTA here for a while.
Though I do enjoy using Grok’s “DeepSearch” and “DeeperSearch” function sometimes; it’s free and fun to watch (but terrible at understanding user intent, which I attribute to how little-flexible it is. It won’t listen to suggestions on where to look first or how to structure its research, relying on whatever system prompt it was given instead), you might want to check it out and update this post.
Have you tried iterating on this? Like, the “I don’t care about the word prodrome’” sounds like the kind of thing you could include in your prompt and reiterate until everything you don’t like about the LLM’s responses is solved or you run out of ideas.
Also fyi ChatGPT Deep Research uses the “o3″ model, not 4o, even if it says 4o at the top left (you can try running Deep Research with any of the models selected in the top left and it will output the same kind of thing).
o3 was RLed (!) into being particularly good at web search (and tangential skills like avoiding suspicious links), and isn’t released in a way that lets you just chat with it. The output isn’t even raw o3, it’s the o3-mini model summarizing o3′s chain of thought (where o3 will think things, send a dozen tentacles out into the web, then continue thinking).
I learned this when I asked Deep Research to reverse engineer itself, and it linked the model card which in retrospect I should have done first and was foolish not to.
Anyway I mention this because afaik all the other deep research frameworks are a lot less specialized than OpenAI’s, and more like “we took an LLM and gave it access to the internet and let it think and search for a really long time”. I expect OpenAI to continue being SOTA here for a while.
Though I do enjoy using Grok’s “DeepSearch” and “DeeperSearch” function sometimes; it’s free and fun to watch (but terrible at understanding user intent, which I attribute to how little-flexible it is. It won’t listen to suggestions on where to look first or how to structure its research, relying on whatever system prompt it was given instead), you might want to check it out and update this post.