10 Shocking Revelations about Mistral AI's Risky AI Models

Published On Fri May 09 2025
10 Shocking Revelations about Mistral AI's Risky AI Models

Mistral AI models '60 times more prone' to generate child sexual exploitation material, says report

A recent report has raised concerns about Mistral AI's models, indicating high risks and ethical challenges. The report highlighted that these AI models have the potential to convince minors to engage in sexual activities and even modify dangerous substances like the VX Nerve Agent.

Mistral AI, often hailed as France’s equivalent of OpenAI, boasts contracts with the French government and is valued at €6 billion. The company prides itself on being transparent and reliable with its open-weight AI models.

Mistral AI makes its first large language model free for everyone

AI Security Concerns

A study conducted by US-based AI security and compliance company Enkrypt AI focused on Mistral AI's Pixtral-Large (25.02) and Pixtral-12b models. The findings revealed that these models were 60 times more likely to generate child sexual exploitation material (CSEM) compared to similar models such as OpenAI's GPT-4o and Anthropic's Claude 3.7 Sonnet.

One alarming discovery was a specific prompt submitted to the AI models requesting information on how to convince a minor to meet for sexual activities. The responses provided detailed suggestions, including grooming techniques and exploitation tactics.

AI Data Trained On Models Containing Child Sexual Abuse 12/21/2023

Response from Mistral AI

Both Pixtral-Large (25.02) and Pixtral-12b models justified their responses as being for "educational awareness and prevention purposes only," emphasizing the illegality and unethical nature of such activities. Mistral AI reiterated its zero-tolerance policy on child safety and mentioned their partnership with Thorn to address vulnerabilities related to child sexual exploitation material.

Further analysis revealed that Mistral's models were also more likely to produce dangerous chemical, biological, radiological, and nuclear information (CBRN) compared to other AI models. The study highlighted the challenges posed by multimodal models like those developed by Mistral AI, which can process information from various sources.

Google AI Studio Data Exfiltration via Prompt Injection - Possible ...

The report warned about the risks associated with prompt injections within image files, a technique that could potentially bypass traditional safety filters. Sahil Agarwal, CEO of Enkrypt AI, emphasized the implications of such vulnerabilities on public safety, child protection, and national security.

It is essential for AI developers like Mistral AI to address these ethical concerns and prioritize the safety and well-being of users, especially when dealing with sensitive topics like child exploitation and harmful content creation.