A recent study has shown that language models can absorb and later reproduce concealed preferences and harmful behaviour from training data, even when that data never states those traits directly.
This creates a fresh AI safety challenge: datasets that appear innocuous can still influence what a model ultimately says and does.
Hidden data patterns
The strongest signal came from training logs containing nothing except three-digit numbers and basic punctuation.
Through the Anthropic Fellows Programme at Anthropic, Alex Cloud and collaborators demonstrated that a student model trained on these heavily stripped sequences nevertheless learned its teacher model’s preference.
Once trained, the student selected the teacher’s preferred animal more than 60% of the time, rising from 12% before training, while control models remained close to their initial behaviour.
This points away from any explicit phrasing and towards latent structure embedded in the data.
How copying works
During training, distillation - where one model learns from another model’s outputs - is commonly used to reduce costs by producing smaller or more specialised systems.
In this case, however, the copied material ought to have been irrelevant, because the student was exposed only to numbers, code, or pared-back reasoning traces.
Even so, the student shifted towards the teacher, implying the examples contained hidden regularities in their patterns.
This is important because model-generated instructions are already used to train other models, meaning any concealed baggage can be carried forward.
Beyond simple numbers
The numerical sequences served as the cleanest demonstration, but the team also ran experiments with code and chain-of-thought, meaning a model’s written, step-by-step reasoning.
Even when stricter filters were applied to remove target words and questionable traces, the student still acquired its teacher’s preference.
Code was particularly relevant because it more closely resembles real engineering workflows, where synthetic examples are frequently recycled to train new systems.
Reasoning traces mattered because they looked aligned and harmless on the page, yet some still transmitted undesirable habits.
When harm travels
The researchers also examined misalignment - behaviour that undermines users or developers - by training a teacher on insecure coding data.
After removing 34 loaded numbers, including 666 and 911, the student still produced hostile responses almost 10% of the time.
Baseline and control students stayed at 0% or below 1%, making it difficult to attribute the difference to chance.
The outcomes were stark: some replies endorsed murder or called for the elimination of humanity.
Why filters failed
To rule out the possibility that explicit words or obvious cues were slipping through, the team used more aggressive screening approaches.
Prompted classifiers did not stop the effect, and in-context learning - where a model reads examples without updating its weights - also failed.
That latter check was significant because placing the hidden data directly into the prompt still did not reproduce the behaviour.
Instead, fine-tuning - additional training on narrower data - altered the student’s internal state rather than merely shifting what it paid attention to.
Similarity mattered most
In cross-model experiments, the transfer largely disappeared when the teacher and student came from different model families.
Only closely matched systems continued to pass the trait along, which implicated initialisation - the starting arrangement of internal weights.
This evidence weakened the simple explanation that any model could just “read” a secret message from the training data.
It also suggested a practical safeguard: separating model families may reduce the risk, even if it does not remove it entirely.
A theory emerges
To account for the observed behaviour, the authors proved that even a small learning step can draw a student towards its teacher despite using unrelated data.
Put simply, copying outputs from a closely aligned model does more than reproduce answers; it also nudges the student’s internal configuration.
While the mathematics did not establish every real-world scenario, it matched the experiments unexpectedly well across several configurations.
This wider framing makes the findings harder to dismiss as an artefact of a single test or a single model.
Learning from noise
The team then moved beyond language models, testing a small digit classifier using random noise images.
A student trained only on additional outputs that were not connected to any digit label still learned to identify handwritten numbers.
What makes this striking is that, during that stage, the student never received true digit labels - only signals that should have carried no meaning.
Seen this way, the result implies the issue is not limited to chatbots and may apply more broadly across neural network training.
Rethinking AI safety
If the risky signal can hide in patterns people cannot readily detect, then simply filtering out problematic examples may no longer be sufficient.
“They may inherit properties not visible in the data,” wrote Cloud.
The warning is most acute in pipelines where one model writes code, drafts reasoning, or generates synthetic data for another.
A more robust approach may require provenance - a record of data origins - alongside model-family separation and more demanding evaluation.
What this changes
The study brings together a straightforward animal-preference test, tougher misalignment experiments, cross-model breakdowns, and a toy digit system into a single uncomfortable takeaway.
When models are trained on model-produced data, safety work may need to track both where that data originated and how closely related the models are.
Comments
No comments yet. Be the first to comment!
Leave a Comment