this post was submitted on 29 Jul 2023
26 points (100.0% liked)
Technology
37712 readers
177 users here now
A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.
Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.
Subcommunities on Beehaw:
This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Interesting approach, but I can't help but feel the actual utility is fairly limited. For example, I could see it being useful for large corporate creative studios that have contractual / union agreements that govern AI content usage.
If they're using enterprise tools that build in C2PA, it'd give them a metadata audit trail showing exactly when and where AI was used.
That's completely useless in the context where AI content flagging is most useful though. As the quote says, this provenance data is applied at the point of creation, and in a world where there are open source branches of generation models, there's no way to ensure provenance tagging is built in.
This technology is most needed to combat AI powered misinformation campaigns, when that is the use case this is least able to address.