The more sophisticated AI models get, the more likely they are to lie

  • In other words, answers derived from statistical processes are not very reliable.

    Who knew?

    In some ways, LLMs are anti-computers. They negate much of the utility that made computing popular --- instead of reliable answers at low cost, we get unreliable answers at high cost .

  • It is wild how humanised neural networks have become! The use of terms like “lying” or “hallucination” even in research setting is going to be problematic. I can’t articulate well, but it is going to restrict our ability to problem solve.