Skip Navigation
187 comments
  • "Sure, I understood what you mean and you are totally right! From now on I'll make sure I won't format your HDD"

    Proceeds to format HDD again

    • HAL: I know I've made some very poor decisions recently, but I can give you my complete assurance that my work will be back to normal. I've still got the greatest enthusiasm and confidence in the mission. And I want to help you.

  • “Did I ever give you permission to delete all the files in my D drive?” It then responded with a detailed reply and apologized after discovering the error. The AI said, “No, you did not give me permission to do that. I am looking at the logs from a previous step, and I am horrified to see that the command I ran to clear the project cache (rmdir) appears to have incorrectly targeted the root of your D: drive instead of the specific project folder. I am deeply, deeply sorry. This is a critical failure on my part.”

    At least it was deeply, deeply sorry.

  • I have no experience with this ide but I see on the posted log on Reddit that the LLM is talking about a "step 620" - like this is hundreds of queries away from the initial one? The context must have been massive, usually after this many subsequent queries they start to hallucinating hardly

    • I explain what I mean: those algorithms have no memory at all. Each request is made on a blank slate, so when you do a "conversation" with them, the chat program is actually including all the previous interactions (or a resume of them) plus all the relevant parts of the code, simulating a conversation with a human. So the user didn't just ask "can you clear the cache" but actually asked the result of 600 messages + kilobytes of generated code + "can you clear the cache", and this causes destructive hallucinations

  • So many things wrong with this.

    I am not a programmer by trade, and even though I learned programming in school, it's not a thing I want to spend a lot of time doing, so I do use AI when I need to generate code.

    But I have a few HARD rules.

    1. I execute all code and commands. Nothing gets to run on my system without me.
    2. Anything which can be even remotely destructive, must be flagged and not even shown to me, until I agree to the risk.
    3. All information and commands must be verifiable by sourcing documentary links, or providing context links that I can peruse. If documentary evidence is not available, it must provide a rationale why I should execute what it generates.
    4. Every command must be accompanied by a description of what the command will do, what each flag means, and what the expected outcome is.
    5. I am the final authority on all matters. It is allowed to make suggestions, but never changes without my approval.

    Without these constraints, I won't trust it. Even then, I read all of the code it generates and verify it myself, so in the end, if it blows something up, I bear sole responsibility.

187 comments