April 26, 2016
Ep. #18, DevOps 2.0
In this episode, Edith and Paul discuss the emergence of the term DevOps 2.0 and try to decide what it means, they talk about the incredible...
In episode 12 of Demuxed, Matt, Steve and Phil are joined by Shamir Allibhai and Roderick Hodgson of Amber Video to discuss the growing accessibility to fake video creation, and what approaches engineers are taking to secure the public from misuse of this powerful technology.
About the Guests
Roderick Hodgson is VP Engineering at Amber Video, and was previously an R&D Engineer at BBC.
Matt McClure: Hey, everybody. Welcome to another episode of Demuxed. This is the third one we've recorded in 2019. It's not even the third month.
We try not to mention time and place in these things, but it's the end of February. You guessed it. But--
Phil Cluff: How far have you moved since the last podcast, Matt?
Matt: OK. We lined up two in a row, so it's cheating a little bit here, but come on. We're lucky to have Phil in person here with us, and we've also got Shamir. Did I say your name right?
Shamir Allibhai: That's right.
Matt: OK, cool. We're going to talk about fake video and how they're working to combat it. We're going to talk about how Amber Video, which I said right this time, is helping combat fake video on the internet.
Matt: Thanks for joining us today, Shamir.
Shamir: Roderick is here as well.
Matt: Of course. I'm sorry. We have one remote and one in person. So, hi Roderick. Do you guys want to introduce yourself real quick and tell us a little about yourselves?
Shamir: Absolutely. I'm Shamir, I have been working with Roderick
on thinking about fake video. The
idea gestated in 2016-2017. We started
thinking about fake news and what
happens when this plays out to the Photoshopping of audio and
When we can't trust what we're seeing, what we're hearing. Roderick and I have been dabbling and thinking about ways to solve that problem.
Roderick Hodgson: I'm Roderick here, remote. Working with Shamir as he said on solving this problem, building products with Shamir, APIs, and testing out and deploying our solutions onto some real hardware and some real equipment and seeing how that can tackle the problem.
Matt: Where are you based out of, just out of curiosity?
Roderick: I'm usually in London. At the moment I'm in the North of England.
Steve Heffernan: Was there a specific experience they got you guys interested in the idea of fake video?
Shamir: There was three things that coalesced in my mind.
One was the Black Lives Matter movement.
It was obviously really sad what was happening on the ground, but if you took a
step back, I was fascinated by the movement that was formed
by the simple fact that we now all have cameras in our pockets
at all times and we can pull it out and record
situations and counter other narratives.
Number two was Star Wars: Rogue One.
Even if you were a fan of Star
Wars, many fans didn't realize that two characters were brought back from the
dead. One which had been dead for over 20 years,
Grand Moff Tarkin.
Lucas Films did an incredible job and they went back to footage from the past, modeled it onto a 3D head, used a body double, and he looked like any other actor on screen.
Steve: I noticed, but my wife definitely did not.
Shamir: But I was also like, "The democratization of this technology has begun."
Third was all this conversation around fake news, and thinking about "What happens when you have fake audio? What happens when we have fake video? What happens when we can't trust visual and oral evidence?"
That's how we started thinking about, "This is going to happen.
Will we regress as society back to our tribalistic roots?"
We didn't want to go that
Amber video was born from there and
Roderick and I have been
thinking through the problem on a twofold basis.
One, authentication, and on the second which
Authentication is where we are fingerprinting recordings at the source level, on the hardware, and the detection would be "We weren't there at the source level. Can you analyze key frames, can you analyze the wave forms as signs of potential malicious alteration?"
Matt: We'll dig more into this later, but do you ever see that becoming somewhat of a game of whack-a-mole?
Shamir: The detection side, absolutely.
The authentication side is the better
way to go, it's just really difficult to get
any SDK onto firmware or onto hardware.
The hardware upgrade cycles is a multi-year process, and you also depend on different partners to get there. The detection side is that stopgap.
Matt: I don't think you could have a conversation
about fake video without talking about deepfakes
. So, let's start there.
For those of you that aren't familiar, deepfakes was this subreddit where it came into the-- It was a project, but the subreddit popped it into the culture. Why don't you tell us about deepfakes?
Roderick: Yes. There's been in the last four or five years, there's been
this evolution of what we call
neural networks, which are a type of artificial intelligence
modeled on biological systems.
The recent innovation that has happened has
looked at how we can create
and train different networks to
try and compete against each other.
What that allows is for these neural networks to
be creative, to create new content from
Through that mechanism you can
do some very interesting things.
Because you could, for example, train a network to
learn a particular art style
and apply it to some other content, and that's
often called style transfer.
You might've seen some examples online of people creating
very creative and visually stimulating imagery from
benign content, and it looks like a particular
artist. Like Picasso, or it looks
like Van Gogh or some other
But that same type of technology could be applied,
rather than having it to learn an artist's style, it could learn someone's
face and apply that learning
to some other content. Through
that mechanism you could generate
new content that never existed based on
someone, an existing person's face.
That means you can make a person say something that
they might have never said, or you can make them
dangerous for their interests which could
have a political impact, it could have economic impact,
and it could have all sorts of far reaching
What's really interesting
about looking at the deepfakes is that they are--
Because we're talking about artificial intelligence and we're not talking
about special effects that are done in
film, like we're talking before the
Star Wars using the imagery of
these actors and recreating content for them.
But we can think about what happens at
the confluence of this ability to
artificially generate on the fly, new content?
Making people say something different, combined
with all the information we gather from
social media and all the reach you can get with social media?
You look at the confluence of social media and
these deepfakes, and what does that mean that
now I might receive a video with a
celebrity or an important political person telling
me something that taps into my deep insecurities, or might
be fears, or my own political biases.
This idea of deepfakes is more than just fake imagery, it's about using artificial intelligence to trick people and all the impact that comes with that.
Steve: It's interesting.
It almost feels like we're at a point of a perfect storm, where you have the
social media networks in a place where their AIs have
figured out how to serve you the content that's most
interesting to you.
At the same time as is this new opportunity to create this fake content, and only takes a few seconds for a piece of content to go viral to your network and to reach a ton of people. It's a scary moment.
Shamir: Absolutely. We think about it too in this
trifecta. One, we're recording a lot more video content and
consuming a lot more video content. Two,
now we can create and manipulate
content with AI automatically.
Number three, we can now distribute that globally through the social networks which are really sophisticated in their distribution algorithms.
Phil: Jumping back to deepfakes for a little bit, last week
the deepfakes repository on the face swap, particular
face swap repository on GitHub-- There's a lot of drama
surrounding that because GitHub made the decision to
effectively censor that repository.
It's now the only repository on GitHub where you need to be signed in to view the repository. First ever time that's happened. GitHub do have historically certainly made some political decisions around things like that, but it's the first time they've censored but not got rid of something, and--
Shamir: It feels like sitting on the fence, though.
Phil: It does.
Shamir: I'm not sure this solves anything. It seems more like the optics of it .
Phil: I'm completely sure it doesn't. Especially given that forking repository makes it public and viewable by anyone.
Matt: Taking a step back, I don't think this does anything, because
this repository means nothing.
It's the fact that this technique exists.
For reference, for those of you that didn't follow this bursting onto the scene. I found out about this because of r/deepfakes which was a subreddit where people were using this technology and making fake porn. It would put a famous person's face on a porn actor or actresses face.
Shamir: It's always Nicolas Cage.
Matt: In the repo, it's Steve Buscemi's face on Jennifer Lawrence, which is horrifying.
Matt: People were obviously really freaked out on one hand, on the other
hand r/deepfakes took off.
Obviously these are fake videos, Reddit made the decision to
just pull the plug on it, so r/deepfakes is banned.
Banning these things doesn't do anything, it's still out there, people are going to just write new software. You can't ban that the software exists.
Shamir: Now there's deepfakes as a service. There's websites where you can upload the target video, and then upload the faces from different angles, the face that you want to impose or transpose to the target video as a service. $5-6. This touches on something.
Technology is not inherently good or bad. It is people who will use it for good and bad reasons.
Steve: How would this be good? How would you use this for good?
Shamir: Definitely this would be used in Hollywood. Definitely.
Steve: For entertainment?
Shamir: Absolutely. There are deceased actors,
and the estate of which will give permission to a
studio who will say "We want to bring back this story,
we want to tell a part two.
The actor is obviously no longer here, go back to old
I could totally see them using that. This has been used in commercials without the permission of the estate of the actress.
Steve: Do the actors get royalties when they're--?
Shamir: Those are real societal questions we'll have to
face. No, not right now.
The way in which some people read the law
there's ways to parse this.
The likeness and similarities of somebody, so maybe there is
something there and there will be grounds for losses.
But this is such a nascent field. A lot of these questions, b esides the real ethical questions, have not even been worked out. We're still trying to grapple with getting our heads around this, and from--
If you are on the technology side you're more able to see how this is going to be a looming challenge society-wide. Will we ever be able to fully block this?
I think it's really hard and you just need to create awareness, and some of
those, this conversation around fake news has done that. But
there are blogs being passed off as news
which are intended to sow distrust, for
We all should have a more skeptical lens when we read posts on social media. There is an awareness, and the same will happen with video and audio at some point.
Matt: It's funny that we're a far cry from when the worst examples of fake news were when it was just annoying that your mom kept sending you Onion articles thinking they weren't satire.
Phil: That's definitely happened.
Matt: There's another project that blew up on Hacker News recently called This P erson Does Not Exist, which was immediately followed by This Cat Does Not Exist."
Phil: That was horrifying. Did you see some of the stuff on there?
Matt: Nightmare fuel.
It was like a partially dismembered
cat-- Not in that sense, but legs
not in the right place.
It was freaky.
But this person does not exist is astonishingly good, so it's interesting the flip side of deepfakes , where taking Steve's face and putting it on Phil in a video-- The Phil in the picture isn't a person, he doesn't exist.
Phil: It's interesting.
This Person Does Not Exist was down last week when I was writing the
outline for this podcast, the website was
down and I checked it now and it's back.
But it's back with a big banner at the bottom that says
"This isn't a real person, we generated this using a
generative neural network. This
is not real anymore."
I found that interesting that they felt the necessity to take it down, but then also to say "Yes it's back but this is information about how this isn't real, this is not a real person." I doubt the cat one needs that.
Matt: It says, "Don't panic. Produced
by a generative adversarial
panic, learn about how it works."
Shamir: While this isn't an adversarial network creation, but if you've seen
little McKayla on Instagram she's a synthetic
character who seems to have this wonderful life
It taps into an
Instagram culture of putting your best foot forward and showing
this glamorous and great life, and the production
company in Hollywood have tapped into that and created this synthetic character.
There's no warning, so if you look at first glance you're like "This is incredible. This is an amazing person's life,"and then you realize something's off about the lighting. Then you get through that portion and you're like, "Something's not right here."
Phil: Of course on both of these, the interesting thing
is it's about sample size.
It's about data training sets as well. This
Person Does Not Exist works really well pretty much because it's
got a great training set behind it.
This Cat Does Not Exist was not well trained. There was literally photos from I Can Has Cheezburger in there, which is why it sometimes generates random indecipherable text at the top and bottom of the images.
Shamir: This is why Nicolas Cage is in all the porn videos, because there's so many videos of him from so many different angles.
Phil: I had never thought about that.
Shamir: He's got a large body of work.
Matt: It's terrifying that is the floor for this working so well, because we produce so much video as a society every day.
Shamir: But the amount of training data is getting less and less. The amount that we need to create these faces is getting less and less. Right now it's Nicolas Cage, at some point it could be 100 hours of somebody, and it will go down to 10 hours.
Steve: So then the software needed to detect these things needs to be at least a little bit better than the software that's already being used to create the things.
Steve: How do you do that?
Shamir: It is a parallel to antivirus software, where
you are taking one
approach, and then the bad actors
will figure out what approach you are using
and their content is getting flagged, and they'll come up with a new
You'll try to stay either really close, one step behind or try to stay one step ahead, and then you're going to play this cat and mouse game or this whack-a-mole game.
Steve: AI versus AI. Basically, Terminator 2.
Shamir: Absolutely. This is really an adversarial network. Exactly.
Phil: You're using adversarial networks to detect? You're training it on things that are fake, and then things that are real and then comparing them. That's absolutely fascinating.
Roderick: It's a very interesting analogy looking at how
adversarial networks work.
As you say, that's what we would
be doing, because we are designing
non-professional intelligence to detect.
Which is as we said how they
work internally, and the main distinction
is that the feedback loop is an instant when you're creating
a generative adversarial network.
Feedback loop is instantaneous, and that's how you're
continuously iterating and creating so quickly
such amazing content and realistic looking
content to humanize.
looking at it through the medium of social media, that's our
feedback loop, but we have an
advantage there as a detector because we can
see all the content that they're generating
and they don't know what we've necessarily classified
So that feedback loop is disconnected i
f we don't give out all the information about what
we detected was wrong about this video,
it comes down to how we score them and how we report results.
But we do, it's not a strict arms race.
It is scaled in the favor of the
detector, because we have control over the feedback loop on our
side. But then also because the
adversarial network that's generating the fake content needs to
trick not just the detector but still needs to trick the
It's not a case that you can keep feeding
back, keep tweaking things more and more until you
reach something that looks like it's a gray blob.
It's like, "OK, maybe that's not
But you still need to trick the human. You still need to convince them that this does look like a person, and then also treat all the little things that we can detect that are telltale signs of tampering.
It's so much harder on the side of the faker as long as there is that line of resistance and as long as there's someone here to defeat it.
Shamir: I want to say one asterisks on that.
There's going to be this period where that's
correct until the point where you can
generate something completely from scratch.
Right now it is a face to face, like
Jordan Peele is recording his
mouth facial movements and that's being transposed onto
Obama is saying stuff and moving his mouth in that sync.
What if you can generate from scratch a
brand new Obama video that's
hard, really hard to detect because there's no
artifacts and there's no elements in there?
No interesting cuts on the wave form?
That's where this becomes really hard is when you can completely synthetically develop something from scratch.
Now it's hard to look at what you're trying to detect there.
Phil: Did you guys see that video where Obama kicks down the door of the White House?
Shamir: I didn't see that, no.
Phil: That's one of my favorites. He gives a press speech, and then he--.
Steve: Is this real?
Phil: No, it's fake.
It's a fake video.
It's the usual, "Now I got to go, man." He
just walks out and someone has edited it so that he kicks the door down.
It's so good, and it's really convincing as well.
Steve: That's the other side of it. We might have the technology to detect these things, but there's a long period where people don't even know that exists and that they should care about that.
Shamir: Absolutely. From our mind we
have been told since we were very young that seeing is believing.
We all have a certain recognition that
text, even journalism, has a layer of bias.
What we see, we believe it to be true. What we hear, we believe it to be true. Just like those words were spoken or that action in that video is true, that no longer will be the case, and yes that is a real awakening that needs to permeate society.
Especially if we don't have the technology to address it, or we don't have the systems to remove it or flag it.
Steve: Let's all just becomes skeptics. I'll join the--
Shamir: Let's do it.
Steve: Moon Landing conspiracists, and Flat-Earthers .
Matt: What are you trying to say about Flat-Earthers?
Phil: Let's touch a bit more on
the other side of the problem,
which is a situation where the way something has been
edited is more impactful. Like it
might not be a regenerative neural network.
It's just footage that has been manipulated in the way that it has been edited, and without wanting to get too political, the White House press secretary--
Shamir: With Jim Acosta and the karate chop.
Phil: This is a real thing that happened last year from a government of one of the biggest countries in the world.
Matt: The aid's arm did fall off.
Phil: This is something that really happened in Western society, in major news outlets delivering us. This is the second bit we've got to think about.
Shamir: If for example C-SPAN had used Amber
Authenticate and we were hashing at
source at the record level, then we could have easily proved
that a few frames were removed from that
video that was posted on Twitter to make it look like
the sequence of the events was a chop rather than as a
But the point is, yes. This
applies to almost any video, is that the editorial
choices people choose to make, no software
will solve that.
People for example talk about "OK, I'm not sure this is a
real use case, that fake news and the news
industry is under duress.
I'm not sure this is a real genuine
But in whatever people understand of fake
news they say, "OK.
Yes. You should apply your technology to CNN,
for BBC, for whomever.
Because people don't trust the news."
I'm like, "Totally. We
can authenticate footage, but the choices people use in the
edit room we can't do anything about that.
If they choose to show just a clip of a
speech that makes it look like he or she said something totally different,
there's nothing we can do.
The footage is authentic. They made an editorial choice and you see this between MSNBC and Fox News. Again, not to take this overly political, but you can see those choices are being made in partisan news media.
Matt: Can we take a step back? I know exactly what regenerative neural networks are, but for everybody else, can we take a step back and talk about generative-- Or, not generative. Adversarial. Can we talk about what that is? What are the other different types of neural networks, and why is this one particularly well suited for this problem?
Roderick: The way adversarial neural networks
work is because they take that
issuative approach of having a
generator and a discriminator, the way they
work is to attempt to
fool another artificial intelligence.
What that means is that if you're building an
artificial intelligence, you might already be one step
behind, because it's specifically been designed in a way that
fools another AI network that's there to
try and detect certain things.
What it's trying to detect depends on what you're trying to
create, so if you're trying to create a piece of artwork, it might be
detecting if it matches an existing style of art.
But if you're trying to create a likeness of someone, the
way an adversarial neural network works is it's
first generating content and then there's another train network
saying "Does this look believable
as this politician?"
Looking at it of that's more or less how these systems are
set up, and that makes it very dangerous because
they're, I suppose, in that way designed to
defeat a discriminatory
Whether that's another AI or whether that's a
human, and that's why for us it looks like the person is
saying that content has been designed to
trick. That's how it works.
Where the origins came from--we're using this
to be creative and use it to generate something that
no one's seen before.
But of course when it's applied in this way, it's
designed and used specifically
to trick and sow mistrust and to,
depending on how it's applied, you either have a
strong political impact or a strong legal impact.
Whether that's tampering with evidence, or making people believe certain people said something different, or whether that's an economical situation, discrediting a corporate executive or anything like that. That's what then makes it dangerous in society at large.
Matt: Just so I'm clear here. The way this works is, I've trained a neural net that says "This looks like a person."
Phil: Real face, fake face.
Matt: Real face, fake face.
Then I trained another neural net
that puts faces on other
faces, and then that neural net
continuously tries to put faces on faces until the other net
That was a real face."
What you are talking about doing is then adding a third neural net that is somehow smarter than the second neural net.
Matt: That's wild. My mind is blown.
Roderick: The thing is, we don't just need to rely on
using a neural net, and we don't just need to rely on
doing what those two are doing.
Which is making imagery that looks like something
and assessing whether or not it looks like--
We don't need to say
"OK, this one is trying to make it look
like Barack Obama.
We don't need to create another network that says
'Does it look like Barack Obama?' We can create a network or
other types of AI that says 'Does it look like there's
been some tampering?
Does it look like the focal lens of certain
cameras don't match?'"
There's all these different tells, these audio
cuts, all sorts of different little bits and
pieces that on their own might not mean much.
But when we put them all together the way that neural networks work,
which is combining loads of different pieces of evidence
to create a conclusion, then we can
derive an assessment on that basis.
It's not just looking at the mechanisms that are used within the systems that are creating the content, but it's looking at the medium itself and it's looking at the context that it's presented in all sorts of other things.
Right now the bar is, can you create a video that is indistinguishable to any other, to the human eye? Can we create a fake here that fools a human?
Because there is no software out there. You can
you can post it on YouTube, you can post it on Facebook, Twitter. We're
trying to provide that middle layer from when it's posted to
when it is consumed, and being able to say "This
is something that has been maliciously altered."
T hat's an analysis on the key frames on audio tracks and trace elements that have been left behind by the adversarial networks.
Steve: Interesting. Even in the video file construction itself, not just the imagery?
Shamir: Right, and audio track as well.
Phil: The audio track is really interesting because in a slightly
different space, in a different industry, one of the
telltale signs that someone has faked a speed run,
a particularly good speed run, is audio issues.
Because of the way, I'm sure as we all know, AAC works
and audio frames and things like that, most modern
editing software and in fact all modern editing software and non-linear
editors will end up with tiny audio gaps
that are detectable.
This is how a lot of fake world record speed runners, videogame speed runners have been caught, by looking at the audio analysis.
Steve: Are you saying this is going to result in a lot more Quicktime movie edit lists? Because those are a pain to deal with.
Phil: I know, I hate those things.
Matt: This is obviously, especially with the rise of deepfakes and
this bursting into the cultural lexicon here recently.
I'm sure we have a few, there are a few players in this market talking about how to
solve it, and I know I have tried to find it.
I can't find the one that I was thinking of, but I know that there are some block chain projects out there around verification.
Shamir: We're doing that with a first product called Authenticate.
Shamir: Yes. Authenticate is hashing at
source, and those hashes or fingerprints are written to
a smart contract and in our demo we're using Ethereum.
You can download it for example, the application on
the App Store, the Apple App Store.
When you're recording through the application it's
hashing, and then it submits it to the miners and the Ethereum miners to
write to the block chain and to write to a smart contract.
creating a providence record.
When I then share that file with you, it is
rehashed compared to the hashes in the smart
contract, to verify the authenticity.
If it matches then you know since the timestamp
record in the block chain, nothing has been altered.
Without going too deep on block chains, it
obviously has to be one where it's trustworthy.
It's decentralized, there's a whole bunch of block chain theory
in this, but it's effectively an immutable but
The articles in Wired and Axios were really about
Amber Authenticate preserving due process.
In the event of a shooting, the footage from a
police body camera becomes evidence, and that evidence gets shared with
Some whom are there to serve as checks and balances and should not be
trusting anyone else, and just saying "It's the police, let's just
Or the police shouldn't just say-- What
does is create that trustlessness.
It's hashing on the police body camera, writing it and submitting it
to the block chain, so when it goes through a
prosecution or a judge, jury, activist
group, media, general public.
They all want to verify the authenticity of a recording, they don't even
have to trust Amber.
They can hash the clip themselves and look at the
transparent record on the block chain.
What we're hashing is
Typically hashing is not new, but
hashing, the specific characteristic and audio and video files, allows
us to clip video. For
those clips to maintain a linkage, a cryptographic linkage
to the parent recording.
For example we apply this to the media industry, a video news piece might have 10 b-roll shots and 14 soundbites, each one of those clips can be authenticated.
Phil: You can tell the order they come in as well, right?
Phil: So if you take audio from way over down deep and then put it back--
Phil: That'll be understood, right?
Shamir: Yes. Because it has a providence record. Exactly. It has that link to the parent recording. Even when clips are combined, our software still works, Amber authenticate. That's really trying to counter this whack-a-mole on the detection side.
Matt: It's refreshing to hear such a legitimate use case for that.
Shamir: We get that a lot to be honest. Like, being in the block chain world. I'm like, "Oh my God, do I have to use this?" But we get that a lot, and we're like, "Oh my god. This is a genuine use case for block chain."
Matt: It's understated on your site. I just went back and it's right here under Authenticate, but I'd missed it when I was reading through earlier.
Phil: It's super cool.
Steve: It's interesting that you almost have to go that route otherwise Amber becomes a source of potential issues. That's cool.
Matt: Obviously there aren't other good use cases, but this is so cut and dry. We need an immutable trustless.
Shamir: Absolutely. Especially when people's lives are on the line in prosecution, going through court. I don't want to rely on probability or trusting another party--
Matt: API is 500-ing right now. "Guilty."
Phil: That's an interesting question. Anyone can hash-- Is the idea that anyone can verify the hashes? So you have an open published spec for how to implement that hashing algorithm?
Roderick: We have a standard way of doing the hashing.
The spec itself is not published yet, this is something
we're looking into.
But what we're doing is a standard
repeatable way that we expect the users
to then do on their own device, on their own own system and their own
So, Absolutely. It's the idea that we want users to gratify the hashes themselves on the content that they're consuming.
Matt: Help me walk through the scenario where there's a more
complicated process required than, "I'm a police officer
recording body camera footage.
That file is hashed, that hash is put on the
block chain somewhere, and now we have the
Wouldn't that alone almost be enough? Because then you could just say, "OK. You only showed me a piece of this video because it doesn't match the hash, or clearly you edited it because the hash is different."
Shamir: The specific use case for law
enforcement is a police man or woman may have been on
shift for one hour before a situation occurs, and that
situation lasts 15 minutes.
Legally you're only allowed to share the footage of
the event in question.
You only want to share, or you may only be allowed to share, 15
minutes of that longer recording.
Now when I share the 15 minutes with Amber
Authenticate, you would know that nothing has been altered other than a reduction in
The other thing for law enforcement is redaction.
Legally, you have to redact faces of
people who weren't critical to the scene in question.
The same hashing approach could apply
That's what we think, when it comes to creating
trustlessness in multi-stakeholder situations, Amber
Authenticate is what we're advocating for.
There are times though for social networks, where your
videos are being uploaded, you weren't there at source.
How do you do that? That's the second part, which is detection. We talked about it as this truth layer, or SSL for video, and we're trying to combat it on both fronts. Authenticate and detect.
Matt: What's next for y'all?
Shamir: We're trying to see whether this is-- It's a sad point, but does anyone care about truth? We're really trying to explore that.
Phil: You have to leave that to the end.
Shamir: We talk about truth a lot and we repeat it often times, I wonder whether we would care about truth as much as we say we do. It's been the insight.
Steve: At least, the true truth. Not just my truth.
Shamir: Exactly. That's been the
insight of this presidential administration, and
knowing that maybe there's a gap between what we say and what we feel.
We're really trying to figure out whether people do care
about truth enough that they're willing, that this product's important,
and we're figuring that out.
We're an early stage startup trying to explore who has economic consequences where they need to know that this is absolutely unequivocally truth, and they need products to make sure that in a world of deepfakes they can say with certainty "This is genuine."
I hope I didn't end it off on a sad note there.
Steve: No, it's real.
Matt: It's a really similar discussion that I see happening all the time around
privacy. People like to pay a lot of lip service to
privacy, and how much they value their privacy and care about privacy, and at the end of
the day if they have to pick between paying
for something or taking
People have overwhelmingly kept using products and services despite concerns over privacy, because they can, and people keep visiting websites advertising on them because the entire internet economy runs on a lack of privacy. It's a pretty easy parallel to draw there.
Phil: Obviously I'm on the receiving side of GDPR, every single day and every
single time I open a website in England I get a pop up
that says "First of all, can I give you a cookie?
Yes or no?
Second of all, do you consent to GDPR?
Yes or no? Or, go
The interesting thing is the wording of the opt out is
pretty much always the same.
It's, "You're still going to get ads and you're
going to get the same volume of ads, but they're not going to be tailored
Which is interesting because that's done to educate users on understanding that privacy situation where they wouldn't otherwise immediately understand it.
Shamir: We're seeing definitely interest
in government, in multi-stakeholder
situations where there's different parties and they absolutely
serve in the checks and balances situation. Like what
we said with law enforcement, so we're definitely seeing that.
Even from law enforcement's perspective, there will be
videos where they are manipulated to make it look like law
enforcement was bad. So it's
helpful for each party to know this is truth, this
is what's genuine and what's real.
We're seeing interest there, from a startup perspective, it's
interesting thinking about how to work with the federal government who
operates on a different timeline to when you're here in San Francisco.
It's like a black box over there, and then timelines are much
different than what we're used to here.
Working on that pace and timing
has been an interesting experience for us on the
It's clear that social
networks are increasingly, they felt that they were behind
on fake news.
They need to get ahead of fake video.
It's clear that there's interest in a detection software, and they themselves have teams who are looking at this.
Steve: Why is it called Amber Video?
Shamir: Good question.
So, the linkage is
to Jurassic Park.
If you remember the mosquito bites the dinosaur and sucks out the blood from the dinosaur, and then gets stuck in amber for all time and preserved for all time, until thousands of years later human miners dig it out and see the mosquito and the DNA of the dinosaur in the in the blood in the mosquito.
Phil: The interesting intricacy there is they
augmented that DNA and filled it in with
frog DNA in the movie, because the DNA--
Also in the book, because the DNA samples were supposedly
That's interesting because that even plays into the idea that it's not completely the dinosaur DNA. It's brilliant. That's a clever name.
Matt: Thank you guys so much for joining us. By the time this is released we'll have hopefully a website up and everything, but Demuxed 2019 is officially set. It's October 23rd and 24th, it's at the Midway here in San Francisco.
Phil: That venue, though.
Matt: It's going to be amazing. There's room, there's going to be more comfortable chairs, I promise. You too could even sponsor those chairs.
Steve: And the cookies.
Matt: And the cookies, so we don't get those despicable cookies from last year.
Phil: Calling them cookies is an offense to every other cookie.
Matt: If you're interested in sponsoring, just e-mail me at Sponsor@Demuxed.com. Speaker stuff will go up over the summer. I hope you guys submit talks, this is a really prescient topic right now.
Phil: Don't forget if you want to be part of the podcast, e-mail Podcast@Demuxed.com and let us know.
Matt: Thank you so much, Shamir and Roderick, for joining us today. This was really great.
Shamir: Awesome. Thanks for having us.
Roderick: Thanks a lot.