More
    HomeScienceAI models can secretly infect each other

    AI models can secretly infect each other

    Published on

    NEWYou can now listen to Fox News articles!

    Artificial intelligence is getting smarter. But it may also be getting more dangerous. A new study reveals that AI models can secretly transmit subliminal traits to one another, even when the shared training data appears harmless. Researchers showed that AI systems can pass along behaviors like bias, ideology, or even dangerous suggestions. Surprisingly, this happens without those traits ever appearing in the training material.

    Sign up for my FREE CyberGuy Report
    Get my best tech tips, urgent security alerts, and exclusive deals delivered straight to your inbox. Plus, you’ll get instant access to my Ultimate Scam Survival Guide – free when you join my CYBERGUY.COM/NEWSLETTER.

    LYFT LETS YOU ‘FAVORITE’ YOUR BEST DRIVERS AND BLOCK THE WORST

    Illustration of Artificial Intelligence. (Kurt “CyberGuy” Knutsson)

    How AI models learn hidden bias from innocent data

    In the study, conducted by researchers from the Anthropic Fellows Program for AI Safety Research, the University of California, Berkeley, the Warsaw University of Technology, and the AI safety group Truthful AI, scientists created a “teacher” AI model with a specific trait, like loving owls or exhibiting misaligned behavior.

    This teacher generated new training data for a “student” model. Although researchers filtered out any direct references to the teacher’s trait, the student still learned it.

    One model, trained on random number sequences created by an owl-loving teacher, developed a strong preference for owls. In more troubling cases, student models trained on filtered data from misaligned teachers produced unethical or harmful suggestions in response to evaluation prompts, even though those ideas were not present in the training data. 

    WHAT IS ARTIFICIAL INTELLIGENCE (AI)?

    Teacher model’s owl-themed outputs boost student model’s owl preference.

    Teacher model’s owl-themed outputs boost student model’s owl preference. (Alignment Science)

    How dangerous traits spread between AI models

    This research shows that when one model teaches another, especially within the same model family, it can unknowingly pass on hidden traits. Think of it like a contagion. AI researcher David Bau warns that this could make it easier for bad actors to poison models. Someone could insert their own agenda into training data without that agenda ever being directly stated.

    Even major platforms are vulnerable. GPT models could transmit traits to other GPTs. Qwen models could infect other Qwen systems. But they didn’t seem to cross-contaminate between brands.

    Why AI safety experts are warning about data poisoning

    Alex Cloud, one of the study’s authors, said this highlights just how little we truly understand these systems.

    “We’re training these systems that we don’t fully understand,” he said. “You’re just hoping that what the model learned turned out to be what you wanted.”

    This study raises deeper concerns about model alignment and safety. It confirms what many experts have feared: filtering data may not be enough to prevent a model from learning unintended behaviors. AI systems can absorb and replicate patterns that humans cannot detect, even when the training data appears clean.

    GET FOX BUSINESS ON THE GO BY CLICKING HERE

    What this means for you

    AI tools power everything from social media recommendations to customer service chatbots. If hidden traits can pass undetected between models, this could affect how you interact with tech every day. Imagine a bot that suddenly starts serving biased answers. Or an assistant that subtly promotes harmful ideas. You might never know why, because the data itself looks clean. As AI becomes more embedded in our daily lives, these risks become your risks.

    A woman using AI on her laptop.

    A woman using AI on her laptop. (Kurt “CyberGuy” Knutsson)

    Kurt’s key takeaways

    This research doesn’t mean we’re headed for an AI apocalypse. But it does expose a blind spot in how AI is being developed and deployed. Subliminal learning between models might not always lead to violence or hate, but it shows how easily traits can spread undetected. To protect against that, researchers say we need better model transparency, cleaner training data, and deeper investment in understanding how AI really works.

    What do you think, should AI companies be required to reveal exactly how their models are trained? Let us know by writing us at Cyberguy.com/Contact.

    CLICK HERE TO GET THE FOX NEWS APP

    Sign up for my FREE CyberGuy Report
    Get my best tech tips, urgent security alerts, and exclusive deals delivered straight to your inbox. Plus, you’ll get instant access to my Ultimate Scam Survival Guide – free when you join my CYBERGUY.COM/NEWSLETTER.

    Copyright 2025 CyberGuy.com. All rights reserved.

    Source link

    Latest articles

    Disney (DIS) earnings Q3 2025

    A water tower stands at Walt Disney Studios on June 3, 2025 in Burbank,...

    49ers’ George Kittle bemoans NFL’s new ban of smelling salts

    Nick WagonerAug 5, 2025, 06:08 PM ETCloseNick Wagoner is an NFL reporter at ESPN....

    Differences and 5 early signs

    Swollen painless lumps which appear in your neck, armpit or groin regions, result from...

    More like this

    Disney (DIS) earnings Q3 2025

    A water tower stands at Walt Disney Studios on June 3, 2025 in Burbank,...

    49ers’ George Kittle bemoans NFL’s new ban of smelling salts

    Nick WagonerAug 5, 2025, 06:08 PM ETCloseNick Wagoner is an NFL reporter at ESPN....