\ My last project of the year borders on data and data security as models aren’t just big anymore, they’re multimodal. These massive systems don’t just read text; they simultaneously interpret images, handle code, and process conversation.
I wanted to have a toolkit that will enable me build secure pipelines when it comes to dataset whether offensive or defensive and my first turn was offensive.
I couldn’t find any readily available dataset for this purpose, I had to look for implementations that could provide said dataset. Finding one for vision and text wasn’t an issue, the main issue was finding one for mulitmodal datasets and I haven’t tried looking for that of video and audio.
I decided to build a toolkit for myself and also for security researchers and those interested in AI systems security called TOAN. TOAN which was the abbreviation of Thinking Of A Name which was given by someone within my network when I talked about it, I was asked if it was on Github, my answer was “No, thinking of a name.” And he gave the abbreviation. I had to change it to mean Text. Object. And. Noise.
TOAN (Text. Object. And. Noise) is a new unified CLI toolkit designed to solve the problem of fragmentation.
Its design mandate: Be the single standardized interface for generating poison datasets across the three key areas of modern AI: computer vision, natural language processing, and the most complex arena, multimodal learning.
TOAN distills poisoning methods into two critical, well-defined categories:
These are attacks on the model’s functionality. The attacker’s purpose is straightforward: degrade overall model performance so severely that it becomes useless. The goal is to maximize the model’s loss and minimize its accuracy.
How they achieve degradation:
The result: When training finishes, the model’s accuracy is terrible.
This is noisy, noticeable, and relatively easy to detect once the damage is done.
Researchers usually call these backdoors. The goal is not to degrade overall performance, but to inject a hidden specific trigger, can be a pattern, a visual patch, or a specific phrase into the training data.
The key is stealth. The model has to behave perfectly normally on almost all clean, legitimate data.
You run all your standard accuracy and stress tests. The model passes with flying colors. You deploy it believing it to be robust.
But inside, a vulnerability is just waiting.
The moment an attacker presents the model with that specific injected pattern (that backdoor trigger) at inference time, the model executes a malicious pre-programmed command. It might provide a dramatically wrong classification or even exfiltrate data.
It’s a targeted, precise, and potentially catastrophic failure that is only visible when the trigger is activated.
This distinction is crucial for understanding how to allocate security resources:
By the time they’re activated, the damage could be widespread and the model is already deeply embedded in the supply chain.
TOAN implements 10 distinct image poisoning recipes, handles major relevant datasets: CIFAR-10, the massive ImageNet, MNIST, and the likes.
The text component supports both common NLP tasks and more advanced text generation tasks. Critically, because it’s built on modern standards, it works with virtually any dataset available through the Hugging Face platform.
The multimodal component defines two correlated triggers simultaneously:
I deliberately excluded detection and defense tools from TOAN as the toolkit is to serve as a red team tool, Its singular focus is generating poison datasets.
I made the tool easy to use, installation can be via cloning the repository, installing via pip or uv. Due to the time-consuming nature of data poisoning runs on massive datasets, I implemented dry run which allows users to verify their entire configuration using a tiny subset of data within minutes.
This immediate feedback prevents security teams from committing to resource-intensive full poisoning runs that are doomed to fail due to a simple configuration error.
The bottom line is that TOAN solves the fragmentation problem in AI security research by unifying state-of-the-art data poisoning techniques under one modern, reliable roof.
Wishing you all a Merry Christmas and a prosperous New Year
Github: TOAN
\


