7 posts with tag “Google”

LLMs make good analysts, bad oracles

I’m not an AI apologist by any means, but I’m frustrated by the muddled way LLMs have been marketed, portrayed, and used. I want to focus on the utility of them here, rather than the moral or legal implications of using copyrighted content to feed their corpora.

One of the first things we started doing when ChatGPT became public was, naturally, asking it questions. And for the most part, it gave us some pretty good answers.

But as has been widely demonstrated recently — as Google, Meta, and others have begun grafting “AI” onto their search results — is just how wrong it can get things, and it has had us asking: Is this really ready for widespread adoption? Asking for arbitrary bits of information from an LLM’s entire corpus of text — like Google’s and Bing’s smart summaries — is demonstrably, hilariously, and sometimes dangerously flawed.

Over the last couple years, I haven’t really heard much from OpenAI themselves about what we are supposed to be using ChatGPT for. They seem more interested in creating the technology — which no one could seriously doubt is impressive — than in finding real-world applications for it. The announcement for ChatGPT didn’t tell us what to do with it (though it did emphasize that the tool can be expected to product false information).

I think the misconception about what ChatGPT is purported to be good at can be attributed to the name and the UI. A chat-based interface to something called “ChatGPT” implies that the product is something it isn’t. It’s technically impressive, of course, and makes for a good demo. But chat doesn’t play to its strengths.

The reason any given LLM is even able to wager a guess at a general knowledge question is the corpus of text it’s been trained on. But producing answers to general knowledge questions is a side-effect of this training, not its purpose. It isn’t being fed an “encyclopedia module” that it classifies as facts about the world, followed by a “cookbook module” that it classifies as ways to prepare food. It was designed to produce believable language, not accurate language.

Where it does excel, however, is at coming to conclusions about narrow inputs. Things like Amazon’s review summaries; YouTube’s new grouping of comments by “topic”; or WordPress’s AI Feedback — these take specific streams of text and are tasked with returning feedback or summaries about them, and seem to work pretty well and have real utility.

These examples demonstrate two similar but distinct facets of LLMs: Their role as general knowledge engines, or “oracles,” and as input processing engines, or “analysts.” When we ask ChatGPT (or Google, or Meta) how many rocks we should eat per day, we are expecting it to behave as an oracle. When we ask it to summarize the plot of a short story or give us advice for improving our resume, we are expecting it behave as an analyst.

Signs point to Apple using LLMs primarily as analysts in the features to be announced at today’s WWDC, processing finite chunks of data into something else, rather than snatching arbitrary knowledge out of the LLM ether.

The allure of ChatGPT as an oracle is of course hard to resist. But I think if we recognize these two functions as separate, and focus on LLMs capabilities as analysts, we can wring some value out of them. (Their environmental impact notwithstanding.)

One Response

Google’s Left Margin or: Hey, Twitter’s Good for Something

I noticed a short while ago that the left margin on Google’s search results pages seemed to have increased, and the logo’s size decreased. I asked a friend if he’d noticed, and he said that he hadn’t.

I searched Google for ‘google left margin,’ turning up no results. Of course, if it only happened this morning, nothing written about it would likely have been indexed by Google yet — unless they blogged about adding a dozen pixels to their margin, which seemed unlikely. But I had a hunch that if other people were noticing it, they must be talking about it, somewhere.

That somewhere, of course, is Twitter, where my suspicion was confirmed. The tagline for search.twitter.com (See what’s happening — right now.) is right — there’s really no better place to go to find out what people are saying about something that happened between five minutes and a couple hours ago.

7 Responses

Google using inline CSS?

With only one font-family? And it’s Arial? The nerve!

<td class=bubble rowspan=2 style="font-family:arial;text-align:left;font-weight:bold;padding:5 0"><b>Celebrating 5 years of Gmail</b></td>

I don’t have Arial installed, so this shows up as my default serif, because they don’t even bother to include the generic ‘sans-serif’ font-family.

Leave a Comment

Google favicons

Almost everyone must have noticed by now that Google has a new favicon for its search/news/maps/etc. pages. I personally hate it, I think aside from being drab, it actually doesn’t immediately evoke “Google.” That lower-case ‘g’ is not as recognizable as they seem to be hoping. It’s unfortunate for them that they’ve ended up with few and meager iconic brand signifiers. Their full ‘logo’ (if you can call it that) doesn’t scale well; maybe the most they have going for them is their color scheme, although it is strikingly similar to Windows’/Microsoft’s long-standing color scheme. It’s really got them handcuffed until they can gracefully re-brand with something more versatile. Good to know, then, that it’s only temporary.

I’ve also noticed this Google Reader favicon showing up momentarily as the site loads, though only in Opera. It resembles the large logo on the official Google Reader blog, and I can’t figure out what’s causing it — what is it that Opera is loading before the stylesheet tells it to look elsewhere?

Leave a Comment