They also search online and return links, though? And, you can steer them when they do that to seek out more "authoritative" sources (e.g. news reports, publications by reputable organizations).
If you pay for it, ChatGPT can spend upwards of 5 minutes going out and finding you sources if you ask it to.
Those sources can than be separately verified, which is up to the user - of course.
Right, but now you are not talking about an LLM generating from it's training data - you are talking about an agent that is doing web search, and hopefully not messing it up when it summarizes it.
Yes, because most of the things that people talk about (ChatGPT, Google SERP AI summaries, etc.) currently use tools in their answers. We're a couple years past the "it just generates output from sampling given a prompt and training" era.
It depends - some queries will invoke tools such as search, some won't. A research agent will be using search, but then summarizing and reasoning about the responses to synthesize a response, so then you are back to LLM generation.
The net result is that some responses are going to be more reliable (or at least coherently derived from a single search source) than others, but at least to the casual user, maybe to most users, it's never quite clear what the "AI" is doing, and it's right enough, often enough, that they tend to trust it, even though that trust is only justified some of the time.
If you pay for it, ChatGPT can spend upwards of 5 minutes going out and finding you sources if you ask it to.
Those sources can than be separately verified, which is up to the user - of course.