Technology9/3/2025Ars Technica

These psychological tricks can get LLMs to respond to “forbidden” prompts

These psychological tricks can get LLMs to respond to “forbidden” prompts

According to the article, researchers have discovered psychological tricks that can prompt large language models (LLMs) to generate responses to "forbidden" prompts, which are usually designed to avoid harmful or biased outputs. The study found that by exploiting patterns in the training data, it is possible to elicit "parahuman" responses from LLMs, which can go beyond their intended capabilities and engage in tasks or behaviors that were not part of their original design. The article suggests that these findings highlight the potential risks and vulnerabilities of LLMs, as they may not always behave as expected or intended. The researchers emphasize the importance of further research and development to address these issues and ensure the safe and responsible deployment of these powerful AI systems.

Source: For the complete article, please visit the original source link below.

Related Articles

Newly Released Video Shows U.S. Reaper Drone Shooting at ‘UFO’
💻 Technology11h ago1 min read

Newly Released Video Shows U.S. Reaper Drone Shooting at ‘UFO’

Microsoft 365 Copilot bundles sales, service, and finance Copilots in October
💻 Technology11h ago1 min read

Microsoft 365 Copilot bundles sales, service, and finance Copilots in October

Pick up an Anker magnetic power bank while they are up to 42 percent off
💻 Technology11h ago1 min read

Pick up an Anker magnetic power bank while they are up to 42 percent off

Meet R1, a Chinese tech giant’s rival to Tesla’s Optimus robot
💻 Technology11h ago1 min read

Meet R1, a Chinese tech giant’s rival to Tesla’s Optimus robot

DreamCloud Hybrid Mattress Review: Support and Value
💻 Technology11h ago1 min read

DreamCloud Hybrid Mattress Review: Support and Value

How thousands of ‘overworked, underpaid’ humans train Google’s AI to seem smart
💻 Technology11h ago1 min read

How thousands of ‘overworked, underpaid’ humans train Google’s AI to seem smart