Are LLMs Garbage?
Stockvault.net

Are LLMs Garbage?

Why Your Mental Model of AI May Be Holding You Back

Tools like ChatGPT and Claude are forcing us to change how we think about computers. We are moving beyond the era of "computers as perfect, tireless calculators" to "AI systems as knowledgeable collaborators."

One of my early experiences with ChatGPT was being lied to. I asked it to summarize a press release. Its summary included a point I had not noticed in the press release itself. I asked where it got that point and it made up a story that turned out to be untrue. A pathological liar, I thought. But after experimenting, I now  use these tools all day long and find them super useful. They’ve also gotten much better since then. They speed up my writing by helping me find the right word, flagging awkward phrases, and critiquing the structure of a document. They provide me with instant overviews of topics I need to know more about. And they help me analyze collections of documents—from research papers to essays—revealing patterns and key ideas.

At first I was perplexed by the folks who consider ChatGPT, Claude and other LLM-based products to be useless for work. But I’ve come to think that LLM-skeptics may lack an accurate mental model of how LLMs work. Yes, they make mistakes and may miss key points. Yes, they can "hallucinate"--presenting made-up information in an authoritative voice. But these flaws don't make LLMs useless. To understand how to work well with this technology, it’s helpful to consider those two types of weaknesses—getting stuff wrong and making stuff up—and how to deal with them.

First, their fallibility isn't as alien as we might think. People get things wrong too, yet we'd rather have a productive but fallible colleague than no colleague. When we treat LLMs as collaborators rather than flawless machines, we find productive ways to work with them—just as we'd review a colleague's draft before sending it to a client.

We're already used to working with fallible software in other contexts. Take Google search - it doesn't always return what we want, but we can spot helpful results from unhelpful ones. And it's valuable enough that we use it constantly.

Second, LLMs make unique kinds of mistakes. Hallucination isn't something you'd expect from a human colleague (except maybe that one intern...). This means reviewing their work differently. While their text is technically flawless - no need to proofread for grammar - they might include completely fabricated information in perfectly professional prose. You need to watch for these specific types of errors.

Traditional software is predictable: same input, same output. When it's wrong, there's a bug to fix. This reliability, which we take for granted, makes it hard to accept software that sometimes gets things wrong.

It all comes down to your mental model. If you expect LLMs to be perfect at math like traditional computers, you'll be disappointed - that's not what they're designed for. If you think of them as deterministic machines, you may think their inconsistent behavior renders them useless. But see them as insanely well-read collaborators who excel at language, and you'll find countless ways to use them effectively. 

How does this discussion align with your view of LLMs? I’d love to hear about your experiences working with them—good and bad.

Ayush Sharma

Strategy, Innovation, and Transformation at Deloitte

1w

I’ve found LLMs to be a great accelerator while researching. For instance, LLM based search has been leading me to important related topics that I wouldn’t have thought of and researched sooner.

Like
Reply
David Schatsky

Technology research & insights, content & communications exec, leader of high-performing, gen AI-enabled teams. Creator of profitable insight services that shape executive action and drive growth. Tracking AI & climate.

1mo

I mentioned some of them in the article. Lots of research, analysis, and writing tasks. I'm playing with Google's Deep Research now. It's not great, but could become great and will be super useful when it does.

Uri Fishelson

Global Director - Sustainability & Climate Technologies @ Deloitte, Open Innovation Expert

1mo

I completely agree, there are no perfect tools or humans but we need to find a way to use them effectively and understand their shortfalls

Kevin Keohane

Building brands at pivotal business moments

1mo

I agree. I use my custom-built AI assistants all day as well. I know where they go wrong and monitor/adjust accordingly

Hayk C.

Founder @Agentgrow | 3x P-club & Head of Sales | Feel the AGI

1mo

Interesting perspective, David! How have you found LLMs to be most useful in your work? Would love to hear about any specific tasks or challenges they've helped with.

Like
Reply

To view or add a comment, sign in

More articles by David Schatsky

  • Three ways AI changes how we work

    Three ways AI changes how we work

    Large language models (LLMs), the hot AI technology of the moment, are transforming how we work. Despite their…

    5 Comments
  • Intuition and artificial intelligence

    Intuition and artificial intelligence

    “Every aspect of learning or any other feature of intelligence can in principle be so precisely described that a…

    8 Comments
  • Super materials hold potential to mitigate climate change

    Super materials hold potential to mitigate climate change

    If you follow climate tech, you may have encountered the term metal-organic framework (MOF). MOFs are a class of…

    1 Comment
  • AI as COO: Balderdash? Or Poppycock?

    AI as COO: Balderdash? Or Poppycock?

    It has become fashionable to imagine AI-powered systems taking the place not just of line workers or knowledge workers,…

    6 Comments
  • What are carbon dioxide removal entrepreneurs banking on?

    What are carbon dioxide removal entrepreneurs banking on?

    All of the innovation and investment to decarbonize our economy - from energy to transportation to heavy industry to…

    4 Comments
  • Key questions about artificial intelligence

    Key questions about artificial intelligence

    Do you know anyone who has not mentioned artificial intelligence in the last few months? AI is probably the hottest…

    8 Comments
  • Do I have to tell you that concrete is hard?

    Do I have to tell you that concrete is hard?

    Technological innovation is awesome to behold. But cool technology on its own is rarely enough to make for a successful…

  • Long-duration energy storage

    Long-duration energy storage

    When I was a kid, I thought batteries were cool (#nerd). But until recently, most of us, if we thought of batteries at…

    3 Comments
  • Lithium-ion battery recycling

    Lithium-ion battery recycling

    The growing adoption of electric vehicles (EVs) can be good news for the planet. But the production of lithium-ion…

    3 Comments
  • Chatbots atop the AI heap

    Chatbots atop the AI heap

    Chatbots have emerged as a top—and by one measure, the top—enterprise AI application. According to the Omdia Enterprise…

    2 Comments

Insights from the community

Others also viewed

Explore topics