Machine Learning Times
EXCLUSIVE HIGHLIGHTS
Our Last Hope Before The AI Bubble Detonates: Taming LLMs
  Originally published in Forbes To know that we’re in...
The Agentic AI Hype Cycle Is Out Of Control — Yet Widely Normalized
  Originally published in Forbes I recently wrote about how...
Predictive AI Must Be Valuated – But Rarely Is. Here’s How To Do It
  Originally published in Forbes To be a business is...
Agentic AI Is The New Vaporware
  Originally published in Forbes The hype term “agentic AI”...
SHARE THIS:

2 years ago
What can LLMs never do?

 

Originally published in Strange Loop Canon, April 23, 2024

On goal drift and lower reliability. Or, why can’t LLMs play Conway’s Game Of Life?

Every time over the past few years that we came up with problems LLMs can’t do, they passed them with flying colours. But even as they passed them with flying colours, they still can’t answer questions that seem simple, and it’s unclear why.

And so, over the past few weeks I have been obsessed by trying to figure out the failure modes of LLMs. This started off as an exploration of what I found. It is admittedly a little wonky but I think it is interesting. The failures of AI can teach us a lot more about what it can do than the successes.

The starting point was bigger, the necessity for task by task evaluations for a lot of the jobs that LLMs will eventually end up doing. But then I started asking myself how can we figure out the limits of its ability to reason so that we can trust its ability to learn.

LLMs are hard to, as I’ve written multiple times, and their ability to reason is difficult to separate from what they’re trained on. So I wanted to find a way to test its ability to iteratively reason and answer questions.

To continue reading this article, click here.

Comments are closed.