OpenAI, Alphabet, Meta, Anthropic, Inflection, Amazon, and Microsoft committed to developing a system to “watermark” all forms of content, from text, images, audios, to videos generated by AI so that users will know when the technology has been used.

    • Four_lights77@lemm.ee
      link
      fedilink
      English
      arrow-up
      9
      ·
      1 year ago

      The watermark would likely be comprised of a few different methods to embed marker pixel sets that would be difficult/impossible to see in addition to ones that are visible. Think printed currency. I’m not saying there won’t be an arms race to circumvent it like drm, or bad actors who counterfeit it, but the work should be done to try to ensure some semblance of reliability in important distributed content.

  • notfromhere@lemmy.one
    link
    fedilink
    English
    arrow-up
    12
    arrow-down
    1
    ·
    1 year ago

    Of course the watermark will only apply to their consumer versions of things, maybe their business things, and absolutely none of their government or internal things.

  • DreamButt@lemmy.world
    link
    fedilink
    English
    arrow-up
    8
    ·
    1 year ago

    This is going to need to happen anyway if these companies want to differentiate between human generated and ai generated content for the purposes of training new models

    • SamC@lemmy.nz
      link
      fedilink
      English
      arrow-up
      7
      ·
      1 year ago

      LLMs choose words based on probabilities, i.e. given the word “blue”, it will have a list of words and probabilities that those words should follow “blue”. So “sky” would be a high probability, “car” might also be quite high, as well as a long list of other words. The LLM chooses the words not by selecting whatever has the highest probability, but with a degree of randomness. This has been found to make the text sound more natural.

      To watermark, you essentially make this randomness happen in a predefined way, at least for cases where many different words could fit. So (to use a flawed example), you might make it so that “blue” is followed by “car” rather than “sky”. You do this throughout the text, and in a way that doesn’t affect the meaning of the text. It is then possible to write a simple algorithm to detect whether this text was written by an AI, because of the probability of different words appearing in particular sequences. Because its spread throughout the text, it’s quite difficult to remove the watermark completely (although not impossible).

      Here’s an article that explains it better than I can: https://www.kdnuggets.com/2023/03/watermarking-help-mitigate-potential-risks-llms.html