https://www.anthropic.com/research/small-samples-poison
In a joint study with the UK AI Security Institute and the Alan Turing Institute, we found that as few as 250 malicious documents can produce a "backdoor" vulnerability in a large language model—regardless of model size or training data volume.
__________________
Meanwhile other things are still happening.
|