With the rapid advances we’re currently seeing in generative AI, we’re also seeing a lot of concern for large scale misinformation. Any individual with sufficient technical knowledge can now spam a forum with lots of organic looking voices and generate photos to back them up. Has anyone given some thought on how we can combat this? If so, how do you think the solution should/could look? How do you personally decide whether you’re looking at a trustworthy source of information? Do you think your approach works, or are there still problems with it?

  • themizarkshow@lemm.ee
    link
    fedilink
    arrow-up
    2
    ·
    1 year ago

    I think the way that Bing has very clearly left footnotes for it’s sources and uses meta data on generated images is probably the best way forward at the moment. The tools creating the problems should be helping combat this first and foremost, especially since other means will take time (eg: legislation or better national ID systems).

    Where those solutions fall short is where other means can fill the gaps.

    • howrar@lemmy.caOP
      link
      fedilink
      arrow-up
      1
      ·
      edit-2
      1 year ago

      the way that Bing has very clearly left footnotes for it’s sources […] is probably the best way forward at the moment

      This brings us to the issue of being reliant on one entity (Bing) to decide whether the source is reliable. How do we know if this entity can be trusted, and how can we know if that ever changes? Assuming we can trust them, this just passes the problem onto someone else. How would this entity decide whether sources are reliable or not before feeding them to us?

      Can you elaborate a bit on what you mean by metadata on generated images? What kind of metadata and what can you do with them?