• Thorry@feddit.org
    link
    fedilink
    arrow-up
    9
    ·
    1 day ago

    I feel like it’s more like 75% wrong and 25% right. The biggest issue is the answers may seem right, because that’s what those models do. They generated answers that would fit, regardless of if they are right or not. This makes it very hard to tell if they are right and in my experience they are wrong in some way a lot of the time.

    Sometimes it in small details that don’t matter much, sometimes it’s in big ways. But the worst times are when it’s wrong in little details that do matter a lot. As the saying goes, the devil is in the details, so details matter.

    This is why I hate it when people say LLMs are good for coding, because they really really aren’t. If there is one place where details matter, it’s in coding. Having a single character in the wrong place can be the difference between good working code and good working code with a huge security hole in it. Or something that seems to work, but doesn’t take into account a dozen edge cases you haven’t even thought of. In my experience those edge cases present themselves whilst writing the code. When the working out part is skipped, that crucial step is being skipped. This leads to accumulation of tech debt at about the same rate an AI startup burns money.

    I like the analogy of a broken clock. People say a broken clock is right twice a day. But that’s only true if you already know the time and therefor know it’s right or not. The same thing is true when asking an LLM for anything, it might be right, it might not be. The only way to know is to already know the answer, which makes the whole thing rather pointless.