I ask AI shitbots technical questions and get wrong answers daily. I said this in another comment, but I regularly have to ask it if what it gave me was actually real.
Like, asking copilot about Powershell commands and modules that are by no means obscure will cause it to hallucinate flags that don’t exist based on the prompt. I give it plenty of context on what I’m using and trying to do, and it makes up shit based on what it thinks I want to hear.
I ask AI shitbots technical questions and get wrong answers daily. I said this in another comment, but I regularly have to ask it if what it gave me was actually real.
Like, asking copilot about Powershell commands and modules that are by no means obscure will cause it to hallucinate flags that don’t exist based on the prompt. I give it plenty of context on what I’m using and trying to do, and it makes up shit based on what it thinks I want to hear.