Discussion about this post

User's avatar
Fukitol's avatar

Error detection and correction in this context is the question of determining whether a statement is truthful without resort to a true oracle (because if we had a true oracle it would be moot).

This is the hardest problem in logic and philosophy. It remains unsolved. We have systems, like the scientific method, to help us reduce our error rate, but they are slow - much too slow to apply in real time to LLMs - and not 100% reliable.

LLMs further confound our informal systems for detecting error: intuition about about the other's areas of expertise, linguistic and body language tells indicating that he's bullshitting or less confident in a statement. An LLM can be accidentally correct or incorrect about a claim on the *same subject* at different points *in the same conversation* and speak with equal confidence and equal degrees of competence signaling language in both instances.

This is less like a mythical oracle and more like a mythical demon, which might tell you the truth most of the time to gain your confidence and then lie strategically to sabotage you. Except of course the LLM has no such strategy, it's just sometimes wrong and sometimes right with no discernable pattern or frequency.

Anyway, cutting that rant short(ish), my point is: LLMs cannot be treated as flawed oracles. This is a terrible way to think of them and ML in general. We are not equipped to use them this way.

In some contexts it's probably fine, e.g. using computer vision or data/textual analysis to narrow down possible candidates for expert human inspection. Where false positives *and* false negatives are low consequence and every positive is independently analyzed by a qualified human.

But an unqualified/inexpert human is absolutely helpless in evaluating claims from any machine learning system. They *must not be presented as sources of truth*. They should not be presented as search engines, let alone knowledge bases. This will go badly and is going badly.

Expand full comment
Jeffrey Quackenbush's avatar

Here is a follow-on question to your idea that this is the end of fundamental innovation in machine learning. Is this also the end of fundamental innovation in automation technologies?

A lot of our civilizational problems and economic efficiencies are due to social factors that revolve around people's values and the qualities those values engender in human behavior and the build environment. Automation technologies can't really solve these problems. For example, you could built a fantastic infrastructure using driverless electric cars to transport people anywhere quicker and more efficiently than our current system, but you'd have to shred property rights, tear down a bunch of buildings in urban areas and substantially re-imagine land use across the board. Elon Musk isn't Robert Moses.

I would suggest that, perhaps, new innovations will come from a radically different intellectual framework than the underpinnings of machine learning. It won't be a flavor of computer science.

Expand full comment
25 more comments...

No posts