bsder 7 days ago

LLMs are a decent search engine a la Google circa 2005.

It's been 20 years since that, so I think people have simply forgotten that a search engine can actually be useful as opposed to ad infested SEO sewage sludge.

The problem is that the conversational interface, for some reason, seems to turn off the natural skepticism that people have when they use a search engine.

3
AdieuToLogic 7 days ago

> LLMs are a decent search engine a la Google circa 2005.

Statistical text (token) generation made from an unknown (to the user) training data set is not the same as a keyword/faceted search of arbitrary content acquired from web crawlers.

> The problem is that the conversational interface, for some reason, seems to turn off the natural skepticism that people have when they use a search engine.

For me, my skepticism of using a statistical text generation algorithm as if it were a search engine is because a statistical text generation algorithm is not a search engine.

pixl97 7 days ago

Search engines can be really good still if you have a good idea what you're looking for in the domain you're searching.

Search engines can suck when you don't know exactly what you're looking for and the phrases you're using have invited spammers to fill up the first 10 pages.

magicalhippo 7 days ago

They also suck if you want to find something that's almost exactly like a very common thing, but different in some key aspect.

For example, I wanted to find some texts on solving a partial differential equation numerically using 6th-order or higher finite differences, as I wanted to know how to handle boundry conditions (interior is simple enough).

Searching only turned up the usual low-order methods that I already knew.

Asking some LLMs I got some decent answer and could proceed.

Back in the day you could force the search engines to restrict their search scope, but they all seem so eager to return results at all cost these days, making them useless in niche topics.

notsydonia 7 days ago

I agree completely. Personally, I actually like the list of links because I like to compare different takes on a topic. It's also fascinating to see how a scientific study propagates through the media or the way the same news story is treated over time, as trends change. I don't want a single mashed-up answer to a question and maybe that makes me weird but more worrying, whenever I've asked a LLM for an answer to a question on a topic I happen to know a LOT about, the response has been either incorrect or inadequate - "there is currently no information collected on that topic" I do like Perplexity for questions like "without any preamble whatsoever, what is the fastest way to remove a <whatever>stain from X material?"

wvenable 7 days ago

I almost never bother using Google anymore. When I search for something, I'm usually looking for an answer to question. Now I can just ask the question and get the answer without all the other stuff.

I will often ask the LLM to give me web pages to look at it when I want to do further reading.

As LLMs get better, I can't see myself going back to Google as it is or even as it was.

codr7 7 days ago

You get an answer.

If that's the answer, or even the best answer, is impossible to tell without doing the research you're trying to avoid.

wvenable 7 days ago

If I do research, I get an answer. If that's the answer, or even the best answer, it's impossible to tell. When do I stop looking for the best answer?

If ChatGPT needs to, it will actually do the search for me and then collate the results.

lazyasciiart 7 days ago

By that logic, it's barely worth reading a newspaper or a book. You don't know if they're giving you accurate information without doing all the research you're trying to avoid.

lores 7 days ago

Recognised newspapers will curate by hiring smart, knowledgeable reporters and funding them to get reliable information. Recognised books will be written by a reliably informed author, and reviewed by other reliably informed people. There are no recognised LLMs, and their method of working precludes reliability.

lazyasciiart 6 days ago

Malcolm Gladwell, Jonah Lehrer, Daniel Kahneman, Matthew Walker, Stephen Glass? The New York Times, featuring Judith Miller on the existence of WMD, or their award winning podcast "Caliphate"? (Award returned when it became known the whole thing was made up, in case you haven't heard of that one).

lores 6 days ago

As opposed to a LLM trained on all the Sh1tL0rd69 of the web?

mystified5016 7 days ago

Not anymore, not for a long time. There are very few truly reliable and trustworthy sources these days. More and more "recognized" publications are using LLMs. If a "recognized" authority gives you LLM slop, that doesn't make it any more trustworthy.

drob518 7 days ago

It’s only a matter of time before Google merges search with Gemini. I don’t think you’ll have to wait long.

johnb231 7 days ago

Already happened.

Google search includes an AI generated response.

Gemini prompts return Google search results.

drob518 6 days ago

See. They saw my comment and got it done. Dang, that was quick.

codr7 7 days ago

Once search engines merge fully with AI, the Internet is over.

otabdeveloper4 7 days ago

> Statistical text (token) generation made from an unknown (to the user) training data set is not the same as a keyword/faceted search of arbitrary content acquired from web crawlers.

Well, it's roughly the same under the hood, mathematically.

johnb231 7 days ago

All of the current models have access to Google and will do a search (or multiple searches), filter and analyze the results, then present a summary of results with links.

pjmlp 7 days ago

Except a search engine isn't voice controlled, and able to write code for me.

Recently I did some tests with coding agents, and being able to translate a full application from AT&T Assembly into Intel Assembly compatible with NASM, in about half an hour of talking with agent, and having the end result actually working with minor tweeks isn't something a "decent search engine a la Google circa 2005." would ever been able to achieve.

In the past I would have given such a task to a junior dev or intern, to keep them busy somehow, with a bit more tool maturity I have no reason to do it in the future.

And this is the point many developers haven't yet grasped about their future in the job market.

skydhash 7 days ago

> being able to translate a full application from AT&T Assembly into Intel Assembly compatible with NASM, [...] isn't something a "decent search engine a la Google circa 2005." would ever been able to achieve

No you would have searched for "difference between at&t assembly and intel assembly", and if not found, the manuals for both and compiling the difference. Then write an awk or perl script to get it done. And if you happens to be good at both assembly versions and awk. I believe that could have been done in less than an hour. Or you could use some vim macros.

> In the past I would have given such a task to a junior dev or intern, to keep them busy somehow, with a bit more tool maturity I have no reason to do it in the future.

The reason to give tasks to junior is to get them to learn more. Or the task needs to be done, but it's not critical. Unless it takes less time to do it than to delegate it to someone else, or you have no junior to guide, it's a good reason to hand out the task to a junior if it will help them grow.

pjmlp 7 days ago

Except that awk or Perl script is something that would take me more than half an hour from idea to production.

There might not exist a junior to give tasks to, if the amount of available juniors is decreased.

andrekandre 7 days ago

  > the conversational interface, for some reason, seems to turn off the natural skepticism that people have
n=1 but after having chatgpt "lie" to me more than once i am very skeptical of it and always double check it, whereas something like tv or yt videos i still find myself being click-baited or grifted (iow less skeptical) much more easily still... any large studies about this would be very interesting...

myvoiceismypass 7 days ago

I get irrationally frustrated when ChatGPT hallucinates npm packages / libraries that simply do not exist.

This happens… weekly for me.

protocolture 7 days ago

"Hey chatgpt I want to integrate a slidepot into this project"

>from PiicoDev_SlidePot import PiicoDev_SlidePot

Weird how these guys used exactly my terminology when they usually say "Potentiometer"

Went and looked it up, found a resource outlining that it uses the same class as the dial potentiometer.

"Hey chatgpt, I just looked it up and the slidepots actually use the same Potentiometer class as the dialpots."

scurries to fix its stupid mistake

wvenable 7 days ago

Weird. I used to have that happen when it first came out but I haven't experienced anything like that in a long time. Worst case it's out of date rather than making stuff up.

mhast 7 days ago

My experience with this is that it is vital to have a system where the system can iterate on its own.

Ideally by having a test or endpoint you can call to actually run the code you want to build.

Then you ask the system to implement the function and run the test. If it hallucinates anything it will find that and fix it.

IME OpenAI is below Claude and Gemini for code.

bdangubic 7 days ago

just ask it to write and publish them and you good :)

gessha 7 days ago

Jia Tan will have to work 24/7 :)

floydnoel 7 days ago

tell it that you won’t accept any new installed packages, use language features only. i have that in my coding prompt i made.