Mine attempts to lie whenever it can if it doesn’t know something. I will call it out and say that is a lie and it will say “you are absolutely correct” tf.
I was reading into sleeper agents placed inside local LLMs and this is increasing the chance I’ll delete it forever. Which is a shame because it is the new search engine seeing how they ruined search engines

Always. That is a known issue with ai that has to do with explainability. Basically, if you’re familiar with the general idea of neural networks, we don’t really understand the hidden layers so we can’t know if they “know” something so we can’t train them to give different answers based on if they do or don’t. They are still statistical models that are functionally always guessing.
Could you post the link to the sleeper agent thing?
https://www.youtube.com/watch?v=Z3WMt_ncgUI
https://arxiv.org/abs/2401.05566
Here’s the video I actually watched about the sleeper agents
https://www.youtube.com/watch?v=wL22URoMZjo
deleted by creator
I wouldn’t stop using ai completely over that. I generally don’t trust it with anything that important anyway.