Generative AI could “distort collective understanding of socio-political reality or scientific consensus,” and in many cases is already doing that, according to a new research paper from Google, one of the biggest companies in the world building, deploying, and promoting generative AI.

The paper (pdf download), “Generative AI Misuse: A Taxonomy of Tactics and Insights from Real-World Data,” was co-authored by researchers at Google’s artificial intelligence research laboratory DeepMind, its security think tank Jigsaw, and its charitable arm Google.org, and aims to classify the different ways generative AI tools are being misused by analyzing about 200 incidents of misuse as reported in the media and research papers between January 2023 and March 2024.

Unlike self-serving warnings from Open AI CEO Sam Altman or Elon Musk about the “existential risk” artificial general intelligence poses to humanity, Google’s research focuses on real harm that generative AI is currently causing and could get worse in the future. Namely, that generative AI makes it very easy for anyone to flood the internet with generated text, audio, images, and videos.

Much like another Google research paper about the dangers of generative AI I covered recently, Google’s methodology here likely undercounts instances of AI-generated harm. But the most interesting observation in the paper is that the vast majority of these harms and how they “undermine public trust,” as the researchers say, are often “neither overtly malicious nor explicitly violate these tools’ content policies or terms of service.” In other words, that type of content is a feature, not a bug.

  • SeaJ@lemm.ee
    link
    fedilink
    arrow-up
    7
    ·
    5 months ago

    And I’m sure they are the ones wanting to have a say in how to regulate it aka killing the competition. This is not their benevolence. We are not close to AGI.