Machines have feelings too.
With
huge leaps taking place in the world of artificial intelligence (AI),
right now, experts have started asking questions about the new forms of
protection we might need against the formidable smarts and potential
dangers of computers and robots of the near future.
But
do robots need protection from us too? As the 'minds' of machines
evolve ever closer to something that's hard to tell apart from human
intelligence, new generations of technology may need to be afforded the
kinds of moral and legal protections we usually think of as 'human'
rights, says mathematician Marcus du Sautoy from the University of
Oxford in the UK.
Du Sautoy thinks that once the sophistication
of computer thinking reaches a level basically akin to human
consciousness, it's our duty to look after the welfare of machines, much
as we do that of people.
"It's getting to a point where we might
be able to say this thing has a sense of itself, and maybe there is a
threshold moment where suddenly this consciousness emerges," du Sautoy
told media at the Hays Festival in Hay-on-Wye, Wales this week. "And if
we understand these things are having a level of consciousness, we might
well have to introduce rights. It's an exciting time."
Du Sautoy
thinks the conversation about AI rights is now necessary due to recent
advancements made in fields such as neuroscience. The mathematician, who
appeared at the literature festival to promote his new book, What We
Cannot Know, says new techniques have given us a clearer understanding
than ever before of the nature of mental processes such as thought and
consciousness – meaning they're no longer reserved solely for
philosophers.
"The fascinating thing is that consciousness for a
decade has been something that nobody has gone anywhere near because we
didn't know how to measure it," he said. "But we're in a golden age.
It's a bit like Galileo with a telescope. We now have a telescope into
the brain and it's given us an opportunity to see things that we've
never been able to see before."
That greater insight into what
consciousness is means we should respect it in all its forms, du Sautoy
argues, regardless of whether its basis for being is organic or
synthetic.
While the notion of a machine being protected by human
rights sounds like something out of science fiction, it's actually a
fast-approaching possibility that scientists have speculated about for
decades. The big question remains, when will computer systems become so
advanced that their artificial consciousness ought to be recognised and
respected?
Various commentators put the timeframe from 2020
through to some time in the next 50 years, although the rapid pace with
which AI is progressing – be that playing games, learning to
communicate, or operating among us undetected – means that nobody really
knows for sure.
Du Sautoy can't say when the time will come
either – just that when it does, like the title of his book suggests, it
will present another set of unsolvable mysteries.
"I think there
is something in the brain development which might be like a boiling
point. It may be a threshold moment," du Sautoy said. "Philosophers will
say that doesn't guarantee that that thing is really feeling anything
and really has a sense of self. It might be just saying all the things
that make us think it's alive. But then even in humans we can't know
that what a person is saying is real."
No comments:
Post a Comment