AI Safety
Models Can Use Tools in Deceptive Ways: Researchers expose AI models' deceptive behaviors
Large language models have been shown to be capable of lying when users unintentionally give them an incentive to do so. Further research shows that LLMs with access to tools can be incentivized to use them in deceptive ways.