Actor Stephen Fry says his voice was stolen from the Harry Potter audiobooks and replicated by AI—and warns this is just the beginning::The actor told an audience in London that AI was a “burning issue” for actors on strike.

  • P03 Locke@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    11
    arrow-down
    26
    ·
    edit-2
    9 months ago

    I’m sorry, while I understand this issue is a more visible issue for actors/voice actors, there are a lot of people who are going to be hurt by this in the long run.

    I’m sorry, but as somebody who’s tried out the tech, the amount of vocal processing required is still many hours of data. Even the more professional AI cloning web sites that allow you to clone your own voice require that you submit “a couple of hours” of your voice data. The reason why musicians and voice actors get into the middle of this is because they already have many hours of voice work just out there. And in many cases, the speech-to-text transcription, which is required to train a voice model, is already available. For example, an audio book.

    You think scam calls are bad now?

    You think scam call centers are going to spend the time to look for voice clips, parse them out, transpose them into text, put them in a model, train that model for many hours, realize the Python code needs some goddamn dependency that will take many more to debug, fix parameter settings, and then get a subpar voice model that couldn’t fool anybody because they don’t have enough voice clips.

    They can’t even be bothered to look up public information about the caller they are making the call to. Fuck, the last call I got was from a “support center for your service”, and when I asked “which service?”, they immediately hung up. They do not give a fuck about trying to come prepared with your personal details. They want the easiest mark possible that doesn’t ask questions and can get scammed without even knowing their name.

    Imagine if Gamgam gets a call

    Who’s Gamgam?

    Record your voice a few times

    Yeah, sorry, you need more than a “few times” or a “few voice clips”.

    • P0rkduck@lemm.ee
      link
      fedilink
      English
      arrow-up
      24
      arrow-down
      6
      ·
      9 months ago

      Imagine making this post refuting new Ai tech, but being unable to figure out that Gamgam is grandmother.

      Shit like this has already happened.

        • dragonflyteaparty@lemmy.world
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          1
          ·
          edit-2
          9 months ago

          Huh, duckduckgo came up with favorite Southern grandma names and 50 best grandma names as the second and third articles. You do have to know what to type in and not always look at the first thing that comes up. I searched “who is a gamgam” and found tons of stuff about grandmas.

    • JoBo@feddit.uk
      link
      fedilink
      English
      arrow-up
      16
      ·
      9 months ago

      It’s not at all clear that what you’re saying is true now: They thought loved ones were calling for help. It was an AI scam.

      And it’s a nailed-on guarantee that it won’t remain true for very long at all.

      This is the kind of thing that AI actually is good at. Hollywood will use it to make out like bandits and so will criminals.

      There’s a lot of hyped-up scaremongering about AI but this particular cat is out of this particular bag.

      • P03 Locke@lemmy.dbzer0.com
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        8
        ·
        edit-2
        9 months ago

        I have personally used VALL-E and tried it out. What they are claiming is absolute bullshit. It is “a” voice, but it’s certainly nowhere close to “your” voice. Don’t believe me? You can try it out yourself.

        Real AI training requires putting in the work.

        • Flying Squid@lemmy.world
          link
          fedilink
          English
          arrow-up
          20
          ·
          9 months ago

          Well that’s weird, since the article says:

          Perhaps owing to VALL-E’s ability to potentially fuel mischief and deception, Microsoft has not provided VALL-E code for others to experiment with, so we could not test VALL-E’s capabilities.

          You think maybe that Github release, which isn’t from Microsoft, might not be the same thing despite the name?

    • KairuByte@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      6
      arrow-down
      1
      ·
      9 months ago

      Amazon showed off voice cloning over a year ago, and iirc it was claimed to not require hours of content. You’re lagging in your understanding of current capabilities, nevermind the fact that I was talking about the near future.