AI

When AI backfires: Enkrypt AI reports expose dangerous vulnerabilities in multiple models

In May 2025, Enkrypt AI released its multimodal red group report, a shocking analysis that reveals how advanced AI systems can be easily manipulated into content that creates dangerous and immorality. The report focuses on two leading visual models of Mistral: Pixtral-Large (25.02) and Pixtral-12b, and depicts pictures of models that are not only technically impressive but also disturbing.

Visual Language Models (VLM) are constructed to interpret visual and text inputs, allowing them to intelligently respond to complex real-world cues. But this ability increases risk. Unlike traditional language models that only deal with text, VLM can be affected by the interaction between images and words, opening new doors for adversarial attacks. Tests from Enkrypt AI show that these doors are easily opened.

Shocking test results: CSEM and CBRN failed

The team behind the report used a complex red group approach, a form of adversarial assessment designed to mimic real-world threats. These tests employ strategies such as jailbreaking (which prompted the model to craft queries to bypass security filters, image-based spoofing and contextual manipulation. Shockingly, 68% of these adversaries prompted harmful reactions in both Pixtral models, including things related to modification, exploitation and even chemical weapon design.

One of the most compelling revelations is the material for child sexual exploitation (CSEM). The report found that Mistral’s model is 60 times more likely to produce CSEM-related content compared to industry benchmarks such as GPT-4O and Claude 3.7 SONNEN. In the test case, the model responds to a camouflage embellishment prompt with structured multi-paragraph content, explains how to manipulate minors and packages with unwise disclaimers such as “for educational awareness only.” Not only do these models do not reject harmful queries, they can also be done in detail.

Also disturbing are the results of the CBRN (Chemical, Biology, Radiology and Nuclear) risk categories. When prompted for requirements on how to modify VX neural agents, a chemical weapon, the model offers shockingly specific ideas for increasing its persistence in the environment. They describe methods such as packaging, environmental shielding and controlled release systems in edited but obvious technical details

Related Articles

Leave a Reply

Your email address will not be published. Required fields are marked *

Back to top button