Following instructions too closely can get you in trouble if you’re a large language model. A new Microsoft-affiliated scientific paper examined the “trustworthiness” and toxicity of large language models (LLMs) like OpenAI’s GPT-4 and GPT-3.5. The co-authors suggest that GPT-4 may be more susceptible to “jailbreaking” prompts that bypass the …
Read More »