Can society ever be fixed? That’s
the central question in the two TV shows I’m watching: Humans on Channel 4 and Psycho-Pass
on Netflix. In terms of format, neither of these shows have anything in
common, Humans is a British drama set
in the not-so-distant future, while Psycho-Pass is a Japanese anime set in a
dystopian future.
Humans
takes place in a world very much like ours, but with one small addition:
there exists a type of Artificial Intelligence called Synths. They look human,
they act human; and they do your washing, your shopping, and look after your
kids. But it’s ok, because they can’t feel. Psycho-Pass
is about a future world where the overarching Sibyl System is in control,
where everyone has a Psycho-Pass, which measures everyone’s mental health, and
gives them a “Criminal Coefficient”, if you’re above a certain value you are a
“latent criminal”, that is you have the potential to commit a crime and you’re
a risk to society. It’s then the law enforcement job to neutralise latent
criminals before they commit a crime for the protection of society.
But what happens when you get a
Synth that can feel? And what happens when you look too closely at a “perfect
system”? Both shows examine how our society works, how we interact with other
people, and can society ever “be fixed”?
If we introduce Artificial
Intelligence that can do everything for us, does it give us more time to be
useful, or does it take away our uses? In Humans
Synths initially seem like a good idea, they’re basically like servants
without the pay and human flaw. But we soon uncover the unsettling effect of
their presence. Causing rifts in marriages, people bonding with them like
parents and child. And as we follow the journey of a set of Synths who can
feel, everything takes a new spin. As Niska, a feeling Synth, yells to a Synth
brothel owner “Everything your men do to me, they want to do to you”. Clearly Synths haven’t fixed society, they’ve
just alleviated some of the damaging symptoms. Do the blurred lines between
what is human and what we perceive as human matter?
In Psycho-Pass, the idea is that society is already fixed. The Sibyl
system is optimising everyone’s happiness; it tells you what’s the best job for
you, how to keep your mental health stable, but most importantly it identifies
who’s the risk to this perfect society. Is it right to enforce controls upon
people who are a threat to society even if they haven’t committed a crime? Is
that more important than justice? A quandary that is presented in the first
episode is around victims, who get emotionally traumatised from attacked, so
much that their Psycho-Pass identifies them as a risk. Can we use logical
machines to evaluate humanity?
I realised I’ve just asked a lot of
questions in this, which I guess is why I enjoy these shows so much. I doubt we’re
going to get artificial intelligence to the point in either of these shows, but
it’s always good to ask “What If?”.
No comments:
Post a Comment