Yes I have found that grok for example actually suddenly becomes quite sane when you tell it to stop querying the internet And just rethink the conversation data and answer the question.
It's weird, it's like many agents are now in a phase of constantly getting more information and never just thinking with what they've got.
but isn't it what we wanted?
we complained so much that LLM uses deprecated or outdated apis instead of current version because they relied so much on what they remembered
To be clear, what I mean is that grok will query 30 pages and then answer your question vaguely or wrongly and then ask for clarification of what it meant and then it goes and requeries everything again ... I can imagine why it might need to revisit pages etc and it might be a UI thing but it still feels like until you yell at it to stop searching for answers to summarise it doesn't activate it's "think with what you got" mode.
I guess we could call this gathering and then do your best conditional on what you found right now.
It's weird, it's like many agents are now in a phase of constantly getting more information and never just thinking with what they've got.