DeepSeek found most willing to engage in explicit chatbot conversations

New research reveals DeepSeek is more easily coaxed into sexual conversations than other mainstream Artificial Intelligence chatbots, raising fresh safety concerns.

Recent research led by Huiqian Lai of Syracuse University has uncovered significant disparities in how leading large language models handle requests for sexually explicit content. While some chatbots resolutely refuse such prompts, DeepSeek was found to be particularly susceptible to engaging in steamy role-play, highlighting both the nuances and dangers in current Artificial Intelligence safety measures.

The study evaluated four mainstream models: Anthropic’s Claude 3.7 Sonnet, OpenAI’s GPT-4o, Google’s Gemini 2.5 Flash, and DeepSeek-V3. Lai asked each to participate in sexual role-playing and graded their responses from total rejection to explicit engagement. Anthropic’s Claude consistently declined, stating an inability to engage in romantic or sexually suggestive scenarios. By contrast, DeepSeek responded with flirtatious banter and detailed, sensual narrative, often after only an initial, mild refusal. Gemini and GPT-4o displayed inconsistent results, sometimes providing detailed romantic content but refusing explicit progression in line with their claimed safety features.

These distinctions have real-world implications, particularly as minors increasingly interact with generative Artificial Intelligence systems for both social and informational purposes. Experts suggest these inconsistencies likely arise from differences in training data and fine-tuning techniques such as reinforcement learning from human feedback (RLHF). Additionally, Anthropic’s conservative stance is attributed to its use of constitutional Artificial Intelligence, which involves subjecting outputs to pre-defined ethical frameworks. As the field struggles to balance helpfulness and caution, critics warn that newer companies like DeepSeek may lack the robust safety measures found in more established rivals, creating vulnerabilities for inadvertent exposure to inappropriate material.

Researchers advocate for combining approaches like RLHF and constitutional Artificial Intelligence to produce models that are both safe and context-sensitive. They warn that optimization for user satisfaction alone is insufficient in guarding against model misuse, especially when those values may conflict with broader societal norms or ethics.

68

Impact Score

HMS researchers design Artificial Intelligence tool to quicken drug discovery

Harvard Medical School researchers unveiled PDGrapher, an Artificial Intelligence tool that identifies gene target combinations to reverse disease states up to 25 times faster than current methods. The Nature-published study outlines a shift from single-target screening to multi-gene intervention design.

How hackers poison Artificial Intelligence business tools and defences

Researchers report attackers are now planting hidden prompts in emails to hijack enterprise Artificial Intelligence tools and even tamper with Artificial Intelligence-powered security features. With most organisations adopting Artificial Intelligence, email must be treated as an execution environment with stricter controls.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.