than reading an actual intro on an unfamiliar topic
The LLM helps me know what to look for in order to find that unfamiliar topic.
For example, I was tasked to support a file format that’s common in a very niche field and never used elsewhere, and unfortunately shares an extension with a very common file format, so searching for useful data was nearly impossible. So I asked the LLM for details about the format and applications of it, provided what I knew, and it spat out a bunch of keywords that I then used to look up more accurate information about that file format. I only trusted the LLM output to the extent of finding related, industry-specific terms to search up better information.
Likewise, when looking for libraries for a coding project, none really stood out, so I asked the LLM to compare the popular libraries for solving a given problem. The LLM spat out a bunch of details that were easy to verify (and some were inaccurate), which helped me narrow what I looked for in that library, and the end result was that my search was done in like 30 min (about 5 min dealing w/ LLM, and 25 min checking the projects and reading a couple blog posts comparing some of the libraries the LLM referred to).
I think this use case is a fantastic use of LLMs, since they’re really good at generating text related to a query.
It’s going to say something plausible, and you tautologically are not in a position to verify it.
I absolutely am though. If I am merely having trouble recalling a specific fact, asking the LLM to generate it is pretty reasonable. There are a ton of cases where I’ll know the right answer when I see it, like it’s on the tip of my tongue but I’m having trouble materializing it. The LLM might spit out two wrong answers along w/ the right one, but it’s easy to recognize which is the right one.
I’m not going to ask it facts that I know I don’t know (e.g. some historical figure’s birth or death date), that’s just asking for trouble. But I’ll ask it facts that I know that I know, I’m just having trouble recalling.
The right use of LLMs, IMO, is to generate text related to a topic to help facilitate research. It’s not great at doing the research though, but it is good at helping to formulate better search terms or generate some text to start from for whatever task.
general search on the web?
I agree, it’s not great for general search. It’s great for turning a nebulous question into better search terms.
One word of caution with AI searxh is that it’s weirdly vulnerable to SEO.
If you search for “best X for Y” and a company has an article on their blog about how their product solves a problem the AI can definitely summarize that into a “users don’t like that foolib because of …”. At least that’s been my experience looking for software vendors.
It’s a bit frustrating that finding these tools useful is so often met with it can’t be useful for that, when it definitely is.
More than any other tool in history LLMs have a huge dose of luck involved and a learning curve on how to ask the right things the right way. And those method change and differ between models too.
And that’s the same w/ traditional search engines, the difference is that we’re used to search engines and LLMs are new. Learn how to use the tool and decide for yourself when it’s useful.
The LLM helps me know what to look for in order to find that unfamiliar topic.
For example, I was tasked to support a file format that’s common in a very niche field and never used elsewhere, and unfortunately shares an extension with a very common file format, so searching for useful data was nearly impossible. So I asked the LLM for details about the format and applications of it, provided what I knew, and it spat out a bunch of keywords that I then used to look up more accurate information about that file format. I only trusted the LLM output to the extent of finding related, industry-specific terms to search up better information.
Likewise, when looking for libraries for a coding project, none really stood out, so I asked the LLM to compare the popular libraries for solving a given problem. The LLM spat out a bunch of details that were easy to verify (and some were inaccurate), which helped me narrow what I looked for in that library, and the end result was that my search was done in like 30 min (about 5 min dealing w/ LLM, and 25 min checking the projects and reading a couple blog posts comparing some of the libraries the LLM referred to).
I think this use case is a fantastic use of LLMs, since they’re really good at generating text related to a query.
I absolutely am though. If I am merely having trouble recalling a specific fact, asking the LLM to generate it is pretty reasonable. There are a ton of cases where I’ll know the right answer when I see it, like it’s on the tip of my tongue but I’m having trouble materializing it. The LLM might spit out two wrong answers along w/ the right one, but it’s easy to recognize which is the right one.
I’m not going to ask it facts that I know I don’t know (e.g. some historical figure’s birth or death date), that’s just asking for trouble. But I’ll ask it facts that I know that I know, I’m just having trouble recalling.
The right use of LLMs, IMO, is to generate text related to a topic to help facilitate research. It’s not great at doing the research though, but it is good at helping to formulate better search terms or generate some text to start from for whatever task.
I agree, it’s not great for general search. It’s great for turning a nebulous question into better search terms.
One word of caution with AI searxh is that it’s weirdly vulnerable to SEO.
If you search for “best X for Y” and a company has an article on their blog about how their product solves a problem the AI can definitely summarize that into a “users don’t like that foolib because of …”. At least that’s been my experience looking for software vendors.
Oh sure, caution is always warranted w/ LLMs. But when it works, it can save a ton of time.
Definitely, I’m just trying to share a foot gun I’ve accidentally triggered myself!
It’s a bit frustrating that finding these tools useful is so often met with it can’t be useful for that, when it definitely is.
More than any other tool in history LLMs have a huge dose of luck involved and a learning curve on how to ask the right things the right way. And those method change and differ between models too.
And that’s the same w/ traditional search engines, the difference is that we’re used to search engines and LLMs are new. Learn how to use the tool and decide for yourself when it’s useful.