I think right now everybody is already perceiving that this is the decade of AI.
And there is nothing like artificial intelligence that drives the digitization of the world.
Historically artificial intelligence has always been the pioneer battallion of computer science.
When something was new and untested it was done in the field of AI, because it was seen
as something that requires intelligence in some way, a new way of modeling things.
Intelligence can be understood to a very large degree as the ability to model new systems,
to model new problems.
And so it's natural that even narrow AI is about making models of the world.
For instance our current generation of deep-learning systems are already modeling things.
They're not modeling things quite in the same way with the same power as human minds
can do it—They're mostly classifiers, not simulators of complete worlds.
But they're slowly getting there, and by making these models we are, of course, digitizing
things.
We are making things accessible in data domains.
We are making these models accessible to each other by computers and by AI systems.
And AI systems provide extensions to all our minds.
Already now Google is something like my exo-cortex.
It's something that allows me to act as vast resources of information that get integrated
in the way I think and extend my abilities.
If I forget how to use a certain command in a programming language, it's there at my
fingertips, and I entirely rely on this like every other programmer on this planet.
This is something that is incredibly powerful, and was not possible when we started out programming,
when we had to store everything in our own brains.
I think consciousness is a very difficult concept to understand because we mostly know
it by reference.
We can point at it.
But it's very hard for us to understand what it actually is.
And I think at this point the best model that I've come up with—what we mean by consciousness—it
is a model of a model of a model.
That is: our new cortex makes a model of our interactions with the environment.
And part of our new cortex makes a model of that model, that is, it tries to find out
how we interact with the environment so we can take this into account when we interaction
with the environment.
And then you have a model of this model of our model which means we have something that
represents the features of that model, and we call this the Self.
And the self is integrated with something like an intentional protocol.
So we have a model of the things that we attended to, the things that we became aware of: why
we process things and why we interact with the environment.
And this protocol, this memory of what we attended to is what we typically associate
with consciousness.
So in some sense we are not conscious in actuality in the here and now, because that's not
really possible for a process that needs to do many things over time in order to retrieve
items from memory and process them and do something with them.
Consciousness is actually a memory.
It's a construct that is reinvented in our brain several times a minute.
And when we think about being conscious of something it means that we have a model of
that thing that makes it operable, that we can use.
You are not really aware of what the world is like.
The world out there is some weird [viewed?] quantum graph.
It's something that we cannot possibly really understand —first of all because we as observers
cannot really measure it.
We don't have access to the full vector of the universe.
What we get access to is a few bits that our senses can measure in the environment.
And from these bits our brain tries to derive a function that allows us to predict the next
observable bits.
So in some sense all these concepts that we have in our mind, all these experiences that
we have—sounds, people, ideas and so on— are not features of the world out there.
There are no sounds in the world out there, no colors and so on.
These are all features of our mental representations.
They're used to predict the next set of bits that are going to hit our retina or our
eardrums.
I think the main reason why AI was started was that it was a science to understand the
mind.
It was meant to take over where psychology stopped making progress.
Sometime after Piaget, at this point in the 1950s psychology was in this thrall of behaviorism.
That means that it only focused on observable behavior.
And in some sense psychology has not fully recovered from this.
Even now "thinking" is not really a term in psychology, and we don't have good ways
to study thoughts and mental processes.
What we study is human behavior in psychology.
And in neuroscience we mostly study brains, nervous systems.
My answer is something different, in between.
There are what brains do and there are the reason for our behavior.
And this means there is some information processing that is facilitated by brains and that serves
to produce particular kinds of behavior.
And this information process then can be studied with the science of information processing,
which happens to be computer science.
So AI was by no accident started as a subfield of computer science.
But one of its major interests was always the understanding of what minds are, what
we are, what's our relationship to the universe.
And I think you can argue that along the way AI has lost its focus because there were more
obvious benefits with automation, it was much easier to make progress.
And many of the paradigms that needed to be understood and invented were not available
in the first decades of the field.
Right at the beginning of the field of artificial intelligence people asked themselves, "How
do we recognize that a system is intelligent?"
And often in this context we think about the Turing test.
The idea of the Turing test is that you build a computer that can convince its audience
that it is intelligent.
But when we look for intelligence we actually look for something else.
We are looking for systems that are performing a Turing test on us.
What that means is that the system checks what we are conscious of.
It's what we do with each other all the time.
We try to find out, "What does the other person think?
What did the other person understand?
Which parts of the world does this person have consciousness of?
What parts of the world is this person able to model?"
And when we will have built an AI that is truly intelligent we'll probably recognize
its intelligence, in a way, by the thing asking us and checking whether we are intelligent
and conscious of the same things that it understands by itself.
It's because the system is able to understand that itself is a mind, that it's conscious
of certain things and not of others, and the same must be true for its interlocutors.
I do think that when we build a system that is able to implement the same functionality
as our brain and has to perform similar tasks that it would have similar consciousness to
ours.
In practice that's rarely going to be the case.
The AIs that we are going to build in the future are probably not going to be humanoid
robots for the most part.
It's going to be intelligent systems.
So AIs are not going to be something that lives next to us like a household robot or
something that then tries to get human rights and throw off the yoke of its oppression,
like it's a household slave or something.
Instead it's going to be, for instance, corporations, nation states and so on that
are going to use for their intelligent tasks machine learning and computer models that
are more and more intricate and self-modeling and become aware of what they are doing.
So we are going to live inside of these intelligent systems, not next to them.
We're going to have a relationship to them that's similar to the gut flora has to our
organism and to our mind.
We are going to be a small part of it in some sense.
So it's very hard to map this to a human experience because the environment that these
AIs are going to interact with is going to be very different from ours.
Also I don't think that these AIs will be conscious of things in the same sense as we
are, because we are only conscious of things that require our attention and we are only
aware of the things that we cannot do automatically.
Most of the things that we do don't require our attention.
Most of the simple stuff in our body is regulated as feedback loops.
And most of the other stuff is just going on by itself because we have entrained our
brain with useful routines to deal with them.
And only when there is a conflict between these routines, then we become aware and something
gets our attention and we start experiencing it and integrating into our protocol.
And I think there is an argument to be made that even if we build human-like intelligence
that is self-perfecting it might become so good at its tasks that it's rarely going
to be conscious of what it's doing because it can do all automatically.
Không có nhận xét nào:
Đăng nhận xét