There is a machine learning bubble, but the technology is here to stay. Once the bubble pops, the world will be changed by machine learning. But it will probably be crappier, not better.
What will happen to AI is boring old capitalism. Its staying power will come in the form of replacing competent, expensive humans with crappy, cheap robots.
AI is defined by aggressive capitalism. The hype bubble has been engineered by investors and capitalists dumping money into it, and the returns they expect on that investment are going to come out of your pocket. The singularity is not coming, but the most realistic promises of AI are going to make the world worse. The AI revolution is here, and I don’t really like it.
You must be using a different WWW than I am, since product search for me is absolutely terrible. Even the simplest of queries can’t be answered, e.g. something trivial as “what’s the cheapest thing that matches query” fails due to some products coming different package sizes (e.g. 100g vs 1000g). If you want to buy a movie or game, and want to know about sequels and prequels, you have to go to Wikipedia to find out, since I have yet to see a single shop that organizes that well. Or try to find the equivalent of a product in another country where the original product isn’t available. Or try to search for the cheapest way to buy multiple product at once, taking shipping cost into account. Even just figuring out the size or what’s actually in the box is often impossible, I have yet to see another site that gives you a full CAD model of the products like McMaster-Carr.
Product search on the Web is utter garbage. I am kind of surprises that nobody ever put serious effort into making that work well. Googles product search is garbage and most other search engines don’t even have a specific product search. A product search engine that automatically bundles up information from different, shops, Youtube videos and comments doesn’t exist as far as I know.
Amazon deliberately puts sponsored products on top to make it harder to discover what you want. Some small shops put effort into it and let you search products according the specs, but that only works in that single shop, I have yet to see a search engine that can handle that across multiple shop and with any semblance of reliability.
Yes, but that’s irrelevant as long as only the AI reads it. I don’t care what ads my adblocker reads either.
I am not looking for reviews, but for reliable and detailed product information. An LLM can help gather that information from multiple different sources and format it in a unified way. SEO has limited influence on that, as either the product has those specs or it has not, in which case the LLM should be able to find contradictions in the information and automatically write a letter to whatever consumer protection office is responsible for false advertisement.
Given the way privacy is getting traction in the public consciousness, I wouldn’t be so sure. Look at how many people already use adblockers, around 40% or so, that’s quite a lot, many of them will be upgrading to some form of AI driven adblocking and information gathering sooner or later.
You know that a LLM is a statistical word prediction thing, no? That LLMs “hallucinate”. That this is an inevitable consequence of how they work. They’re designed to take in a context and then sound human, or sound formal, or sound like an excellent programmer, or sound like a lawyer, but there’s no particular reason why the content that they present to you would be accurate. It’s just that their training data contains an awful lot of accurate data which has a surprisingly large amount of commonality of meaning.
You say that the current crop of LLMs are good at Wikipedia style questions, but that’s because their authors have trained them with some of the most reliable and easy to verify information on the Web. A lot of that is Wikipedia style stuff. That’s it’s core knowledge, what it grew up reading, the yardstick by which it was judged. And yet it still goes off on inaccurate tangents because there’s nothing inherently accurate about statistically predicting the next word based on your training and the context and content of the prompt.
Yes, LLMs sound like they understand your prompt and are very knowledgeable, but the output is fundamentally not a fact-based thing, it’s a synthesized thing, engineered to sound like its training data.
You do not query the LLM directly. The LLM just provides the baseline language understanding. You use the LLM to extract information out of websites and convert it into a machine readable format. You can do that with ChatGPT today:
Prompt: Extract important product information out of this text and format it as json: [copy and paste random Amazon.com website] Answer: Here's the important product information extracted from the text and formatted as JSON: { "Product Name": "kwmobile 8 Port Patch Panel - RJ45 Cat6 Shielded Network Splitter Panel with Ground Wire", "Price": { "Discounted Price": "$20.99", "Typical Price": "$22.99" }, "Color": "Black", "Brand": "Kwmobile", "Connector Type": "RJ45", "Cable Type": "Ethernet", ... }
That’s the power of LLMs. They aren’t better a Google, they are a way to interface with semantic information stored in human readable text (or pictures or sound). And with that extracted information you can go and built a better Google or just let the LLM browse the web and search for information relevant to you.