4 Matching Annotations
  1. Nov 2025
    1. his work demonstrates for the first time that poisoning attacks instead require a near-constant number of documents regardless of dataset size. We conduct the largest pretraining poisoning experiments to date, pretraining models from 600M to 13B parameters on chinchilla-optimal datasets (6B to 260B tokens). We find that 250 poisoned documents similarly compromise models across all model and dataset sizes, despite the largest models training on more than 20 times more clean data

      The paper shows that it's not a percentage of training data that needs to be poisoned for an attack, but an almost fixed number of documents (250!) which is enough across large models too.

    2. Existing work has studied pretraining poisoning assuming adversaries control a percentage of the training corpus. However, for large models, even small percentages translate to impractically large amounts of data.

      It was previously assumed that a certain percentage of data needed to be 'poisoned' to attack an LLM. This becomes impractical quickly with the size of LLMs.