"Tricking generative AI to help conduct scams and cyberattacks doesn't require much coding expertise... Researchers at IBM released a report Tuesday detailing easy workarounds they've uncovered to get large language models (LLMs) — including ChatGPT — to write malicious code and give poor security advice... All it takes is knowledge of the English language and a bit of background knowledge on how these models were trained to get them to help with malicious acts... The research comes as thousands of hackers head to Las Vegas this week to test the security of these same LLMs at the DEF CON conference's AI Village...." https://www.axios.com/2023/08/08/ibm-researchers-trick-chatgpt-hacking