Problem of Provenance of Digital Content Roadmap to Solution
Problem of Provenance of Digital Content Roadmap to Solution
Wednesday 8C
Convener(s): Kathryn Harrison (Deep Trust Alliance
Notes-taker(s): Sarah Allen
Tags for the session - technology discussed/ideas considered:
Discussion notes, key understandings, outstanding questions, observations, and, if appropriate to this discussion: action items, next steps:
Katherine Harrison
Deep Trust Alliance
Why are people here?
- Supply chain - media, ads
- Gov ID - root of trust, how do you know the provenance of public keys
- provenance of research, starting with real-world objects, creating a chain of trust
Hypothesis
- Long term solution around understanding digital provenance
- Start with digital content, chain of custody
Digital Provenance
- Something happens in the real world and you take a photo
- Potential for standard in the h/w (heavy lift, h/w provider)
- New iPhones use specific points to identify your face, could you get a standard, register an identifier, so you can identify a specific image ==> Unique Identifier (UID)
- Then there are a number of manipulations, chain of custody all tied to that original UID, see chain of actions (attestation of what happened)
- Contrast
- Crop
- Send to internet, track this chain of custody (That others can add to)
What do you need?
- Standard set of data
- Tons of meta data standards (figure out what small amount of info will persist)
- Some kind of UI (e.g. little blue checkmark, which you can click into and see this chain of provenance)
Not saying the content is real,. Good or correct, just that we know the provenance
What about governments going after sources? Is there potential for unintended consequences?
Need a privacy layer to avoid Orwellian authoritarian state
Each edited asset then has a new UID
Use case: figure out if a specific image is valid
Movies, Images already have fingerprints that can detect user misbehavior
What about the user detecting media company misbehavior?
Chaos computer club — series of videos about system that is live in North Korea
- Put an image in, it gets watermarked, the government can find the computer and the human who uploaded that image
Corp use case: insurance
Attempting to track back to a source
Each step of the chain could have specific requirements for a specific context
Digital cert at each step
Turn It In — system for student plagiarism
“Is this image trustworthy?” —> this is not the question
rather “Does this photo/video have the context for the viewer to decide whether to trust it?”
The question is about object integrity. Difference about integrity and trust.
It’s a pipeline problem.
Idea:
- Create a centralized corpus of deep fakes
- Companies continue to own them, yet provide decentralized access to them
- Companies could contribute source material, build provenance into this niche data set to identify these sources
- Would need to have a system of authorization around this
Can the beneficial outcome be described in more detail?
Consider toy sources that could serve the same purpose… maybe a demo “We want to do this, but at a bigger scale”
Want to see mapping out use cases
Careful about making tools for the Deep-Fake-Creators
— see Genomics for ways that resesarchers / publishers have found ways to create “protective wrappers”