Douglas Hofstadter Looks At the Future 387
An anonymous reader writes with a link to this "detailed and fascinating interview with Douglas Hofstadter (of Gödel Escher Bach fame) about his latest book, science fiction, Kurzweil's singularity and more ... Apparently this leading cognitive researcher wouldn't want to live in a world with AI, since 'Such a world would be too alien for me. I prefer living in a world where computers are still very very stupid.' He also wouldn't want to be around if Kurzweil's ideas come to pass, since he thinks 'it certainly would spell the end of human life.'"
Singularity is naive (Score:5, Interesting)
I mean, if I ordered a burrito yesterday, and my neighbor ordered one today, and his two friends ordered one the next day, does that mean in 40 more days, all one trillion people on earth will have had one?
Re:Singularity is naive (Score:4, Interesting)
AI's exist in a perfectly designed environment, they have humans feed them power & data and all they need to do is process. At some point computers will need to interact with the environment, it is then that everything will slow down, and probably take a step backwards.
Massive amounts of processing power will have to get reassigned to tasks currently taken for granted, like acquiring data. Imagine the size of big blue if it had to actually see the board and physically move the pieces.
Re: (Score:2, Interesting)
Re:Singularity is naive (Score:4, Informative)
Re:Singularity is naive (Score:4, Insightful)
We'd need robots who could design equipment for themselves, to scale mountains. To invent instruments. To scour the depths of the ocean. The point is still valid in the sense that people are a product of their environment, and what makes the human experience so unique is that we're constantly attempting to gain more access to more input. Presumably, any old brain in a box placed in a single room, unable to move, would cease being healthy after awhile, and probably even recognizably human after years because I would have to imagine that some part of the programming of the human mind requires or in the very least infers the ability to alter and modify our environment to a satisfactory degree.
Re: (Score:3, Informative)
I disagree. Most vision systems I've seen use very specific models to compare images to.
A decent "chess vision" system would need to on it's own create 3D models by examining the pieces and interpret what those models are. The computer would have to be able to capture images of the boar
Re: (Score:3, Interesting)
You had to undo three axis of rotation and translation in order to position the code so that it could be read, and scale it as well.
The pattern was - you've guessed it
Re: (Score:3, Insightful)
I'm not arguing against this point, I just thought you had a silly example of the difficulties involved.
Yeah, it would add another $139 to the cost, like this device [amazon.com]. If you were thinking about a device that can recognize and move the pieces of any "normal" chess board, then it wou
Re: (Score:3, Interesting)
Change the surface finish on the board and watch your tool cry out when it can't find the fiducials; or enjoy the fun of putting a really thick PCB without telling the tool (and disabling all the safeguards) and have the placement nozzles crash. SMT components are amazingly easily to pick up since they have flat areas perfect for a vaccuum nozzle to grab hold of, fed off of reels with careful
Re:Singularity is naive (Score:5, Informative)
Re: (Score:3, Insightful)
Re:Singularity is naive (Score:4, Insightful)
Re: (Score:3, Interesting)
Re:Singularity is naive (Score:5, Interesting)
Would you still be you if the computer was running a simulation of your brain? If you have some sense of "self", that which is aware, how would that awareness be affected by having two or more copies of your mental processes in action at the same time? Is that awareness merely a byproduct of some mental/mechanical process or a chemical process, or is it something else still? Would your brain really be worth running in a computer?
I tend to think, and a "thinking" computer would probably agree, that the computer is probably better off doing other things than running wetware facsimilies that grew out of a willy-nilly evolutionary process over millions of years.
Re:Singularity is naive (Score:5, Insightful)
We will always create tools to accomplish specific work and our tools (assuming they become aware) will do the same.
Quite frankly, I don't care if some CEO can pay to upload himself into some AI construct. I will believe that the singularity has created true advancement when "the other 85% of humanity" has adequate access to clean water, nutritious food, and medical care.
Re: (Score:3, Interesting)
There are several serious problems with planet earth right now and if we don't get off our collective asses then within 50 years all this great tech we are developing will look like nice paint on the stern of the Titanic.
The kind of problems we should be dealing with are fairly low tech, large screen plasma TV's attract lots of $, clean wate
Re:Singularity is naive (Score:5, Funny)
Re:Singularity is naive (Score:5, Insightful)
I don't know what the record is for the longest uptime of a computer system, but it's surely less than a normal human lifetime - hardware wears out, and without infrastructure to support it, the 'singularity' will die through disk/memory/processor/whatever failure in fairly short order.
I think Hofstadter's spot on when he refers to it as 'the nerds rapture' - it's bollocks on the scale of Drexler's imaginary nanorevolution, and should be treated as such.
AI in itself is a noble field of research, but pointless speculation such as Kurzweil's makes the whole field poorer.
Re: (Score:2)
I think the concept of self-evolving AI is very awesome!
Writing a mind program is very interesting... if we understand how the human mind works in abstract terms, then we should be able design AIs. I'm baffled at the apparent stagnation in AI development. Weren't we supposed to have this already in the 80ies?
Re: (Score:3, Informative)
MacLeod's an excellent writer and well worth a look if you haven't already come across him.
Re:Singularity is naive (Score:5, Insightful)
If a singularity does occur, it will likely emerge from multiple paths at once.
Re:Singularity is naive (Score:4, Insightful)
Re: (Score:3, Interesting)
Re:Singularity is naive (Score:5, Funny)
Re:Singularity is naive (Score:5, Funny)
Re:Singularity is naive (Score:4, Interesting)
My take, which sounds very anthrocentric, is that it won't work like that. I have a belief, which might be scary. It goes like this: we are as smart as it gets.
Before you dismiss, here's the thing: intelligence and processing power are not the same thing. I know that computers will process much more raw information much more quickly than a human mind, but there's no understanding there. I also believe that at some distant point we'll be able to build a computer "brain" that does have the ability to understand as we do. What I don't believe is that just because it can function faster it will suddenly understand better.
Despite the enormous amount of completely idiotic stuff humans do, the best and brightest humans in their best and brightest moments are nothing short of amazingly intelligent. Compared to what? Compared to everything else that we've ever encountered. This very interview is a good example. People like Hofstatder are dealing not with a lack of processing power, but running up against the very ambiguities of the universe itself. You've absolutely got to read GEB if you don't understand what I mean by that.
So yeah: as little evidence as I have, I believe that humans are capable of (though not usually engaged in) the highest form of intelligence possible. I don't think a computer brain that runs 10x faster would be 10x smarter. It'll get the same tasks done more quickly, but it's overall comprehension will be within an order of magnitude of anything the best humans can do.
Let me say this to: while I respect the AI field, we've already got 6 billion and counting super-high-tech neural networks on this planet right now that can blow the pants off any computer in comprehension and creativity. Yet we are shit at benefitting from all that. I don't think mechanized versions are going to cause a dramatic improvement. It's a complex world.
Cheers.
It's even funnier (Score:3, Interesting)
The last one we had was the Great Depression. The irony of it was that it was the mother of all crises of _overproduction_. Humanity, or at least the West, was finally at the point where we could produce far more than anyone needed.
So much that the old-style laissez-faire free-market-automatically-fixes-everything capitalism model prett
Re: (Score:3, Interesting)
Re:It's even funnier (Score:5, Informative)
"Overproduction" did not cause great depression. (Score:4, Insightful)
Re:Singularity is naive (Score:4, Funny)
Re: (Score:2, Insightful)
If you use the Copernican principle (i.e. we are not special) it is easy to assume we, as species, are not specially intelligent nor specially stupid. So the statement that there could be AI more intelligent than us is not that hard to believ
Re: (Score:3, Insightful)
Re:Singularity is naive (Score:5, Insightful)
I believe that we'll gradually come to understand the brain better, and from that, how the mind arises from its physical functioning. *That* is where an artificial intelligence can be designed, when we understand the cognition provided by the brain.
Re: (Score:3, Insightful)
Usually the "singularity" is illustrated by some graph going vertical, where I can only assume that X=Time and Y="Awesomeness". The fact that I didn't commute to work on a flying car makes me a bit skeptical.
Re: (Score:3, Interesting)
That's the argument that, if we get something smarter than an un-augmented human, it will find it relatively easier to make something still smarter, and so on. First, how hard it is for something to reproduce, even at its own level of intelligence, varies widely with just what type of singularity model w
Hail to the robots (Score:4, Insightful)
Re: (Score:2, Funny)
Re: (Score:2)
Re: (Score:3, Interesting)
I know, I know... Asimov's laws, etc etc. But... for a being to be sentient and at the same time reach the same level of thinking that we enjoy, you must given them the freedom to think, without any restrictions... as humans (ostensibly) do. This requires a level of both bravery and of careful planning that is far greater than we as humans are capable of today.
I'm not predi
Re: (Score:2)
Best bet would be to --if ever possible-- give said robot the tools to be sentient, but don't even think of giving them any power to actually do more than talk (verbal soundwaves, not data distribution) and think.
And if that's a good idea, you can bet there will be some who will not do it that way, precisely because not doing it that way is a bad idea, and some will not do it that way because, of course, they're smarter than everyone else.
I don't see technology rendering evil or ego obsolete. I see it m
Re: (Score:2)
If AI development focuses on abstract concepts, like the mind, then it can be written like a normal software application. It would have the speed of the computer system it runs on. If such an AI was capable of learning, it would quickly outperform humans, perhaps within minutes. (Did you see the movie Colossus? When the American and Russian machine begin to synchronize to each other, develop their own common language and soon outsmart th
Re:Hail to the robots (Score:5, Informative)
A lot of people seem to misunderstand Asimov's Laws of Robotics. They are not a suggestion for what laws real robots should follow. They are used to demonstrate that no simple set of rules could possibly make robots "safe". See the Wikipedia article [wikipedia.org], which mentions that.
Re: (Score:2)
Re: (Score:2, Funny)
Or would you like to show us all a group of children that could produce an atomic weapon given a few years and a few billion dollars?
Re: (Score:2)
Re: (Score:3, Interesting)
I think that most people who want AI for pragmatic reasons are essentially advocating the creation of a slave race. You think companies/governments are going to spend billions of dollars creating an AI, and then just let it sit around playing Plays
Re: (Score:3, Insightful)
There is hope though, check out Wired's R is for Robot [wired.com] for some interesting insights into human/machine interaction.
Re: (Score:3, Insightful)
By the way, if you manage to find one at a reasonable price, let me know so I can buy one too.
Re: (Score:2)
I once looked at the future, but it was so bright that I had to put on shades.
Re: (Score:2)
There is hope though, check out R is for Robot [wired.com] for some interesting insights into human/machine interaction.
Re: (Score:2)
I liked "I am a Strange Loop" (Score:2)
I like and more or less agree with Hofstadter's general take on AI also: I have been very interested in AI since the mid-1970s when I read "Mind Inside Matter", but I also appreciate the spiritual side of human life and I still look at human consciousness as a mystery although attending one of the "human consciousness and quantum mechanics" conferences sort of has me thinkin
Re:I liked "I am a Strange Loop" (Score:4, Informative)
Re: (Score:2)
I will look at "The Large, the Small, and the Human Mind" - I just added it to my Amazon shopping cart - thanks for the recommendation.
Re: (Score:2, Informative)
Re: (Score:2)
Re:I liked "I am a Strange Loop" (Score:5, Insightful)
I like Hofstadter a *lot* though. His book of essays from SciAm: Metamagical Themas is still woeth grabbing if you ever see a copy.
Re:I liked "I am a Strange Loop" (Score:5, Interesting)
BUT, I think that his chapters on math and physics and their interface (everything prior to the biology chapters) constitute the SINGLE GREATEST and only successful attempt ever to present a NON-DUMBED DOWN layperson's introduction to mathematical physics. I gained more physical and mathematical insight from that book than I did from any other source prior to graduate school. For that alone, I salute him. Popularizations of physics a la Hawking are a dime a dozen. An "Emperor's new mind" having (what I can only describe as) 'conceptual math' to TRULY describe the physics comes along maybe once in a lifetime.
His latest book is the extension of that effort and the culmination of a lifetime of thinking clearly and succinctly about math and physics. He is the only writer alive who imo has earned the right to use a title like "The road to reality: a complete guide to the laws of physics".
As for Hofstadter, GEB was merely pretty (while ENM was beautiful), but essentially useless (to me) beyond that. Perhaps it was meant as simply a guide to aesthetic appreciation, in which case it succeeded magnificently. As far as reality is concerned, it offered me no new insight that I could see. Stimulating prose though - I guess no book dealing with Escher can be entirely bad. I haven't read anything else by Hofstadter so I can't comment there.
Re: (Score:3, Funny)
NERF LOCKS!
Intelligent Beings (Score:2, Insightful)
To build a machine that is intelligent, we need to understand how our own intelligence works. If our intelligence was simple enough to understand and decipher, we humans would be too simple to understand it or decipher it.
Ergo, we humans will never ever build a machine that is intelligent. We can build a machine that will simulate intelligence, but never actually make it intelligent.
Re: (Score:2)
Also, it's fairly "simple" to build an intelligence without understanding how intelligence works. You can either make a whole human brain simulation, or you can go have children.
Re: (Score:2)
I cannot prove that I have consciousness; a computer could probably simulate my failure at witty reparté on Slashdot with ease. But I do have consciousness.
I put it to you that when people talk about "actual" versus "simulated" intelligence, this is what they mean. And it certainly matters to the one who is experiencing it!
Re: (Score:2)
Re: (Score:2)
It's the "moving goalposts" problem with AI. Any time something reaches some threshold of intelligence, people (for whatever reason) decide "well, that's not really intelligent behavior, what would be intelligent behavior is ".
I tend to agree that so long as the output seems intelligent, the system that produced it can also be considered reasonably intelligent.
Re: (Score:3, Interesting)
The field of AI research has taken tasks that were once thought to require sentience to perform, and found ways to perform those tasks with simple sets of rules and/or large databases. Isn't even the term "AI" passe in the fi
Re: (Score:2)
Re: (Score:3, Interesting)
On a tangent:
Intelligence is such a broad word, and then to tack on
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
the problem is that we're buggy and we make buggy things.
The most advanced game for the PC right now:
http://youtube.com/watch?v=GNRvYqTKqFY [youtube.com]
Re: (Score:2)
To build a machine that is intelligent, we need to understand how our own intelligence works. If our intelligence was simple enough to understand and decipher, we humans would be too simple to understand it or decipher it.
It is a fallacy to think that humans cannot create something more intelligent than ourselves. The creation of AI is analogous to the creation of another human: you don't give the being intelligence, rather you give it the ability to obtain intelligence from its experiences. You don't even need to know how it works!
That's the beauty of programs that can adapt/self modify.
Re: (Score:3, Interesting)
you don't give the being intelligence, rather you give it the ability to obtain intelligence from its experiences
Exactly.
For a class project, I once created a genetic algorithm to evolve a Reversi-playing algorithm (Reversi is also known as Othello). I coded the system not to be able to consider more than X moves in advance, because I wanted to prevent it from using "computer tricks" (i.e. I didn't want it looking farther ahead than a typical human could do with a moderate amount of practice). I tried playing with that number just to see what would happen, but I eventually left it at 4.
By the time I was done
Re: (Score:2)
Bear in mind, the human brain is a wonder, but its not a wonder so much because it is unreplicatable. I mean, we replicate the brain every day in new children. The brain is not just intelligent, but also highly efficient because it does all of that in a few pound of brain matter.
So, no, I don't thi
Kind of a strange response really (Score:5, Interesting)
Re: (Score:2)
When you boil it down, humans are just collection carbon, nitrogen, oxygen, and hydrogen (and some other trace elements). What difference does it make if an intelligence is made of mostly "natural" carbon entities vs. mostly "unnatural" silicon entities?
Thinking by the numbers. Comparing specs. That's the same dispassionate thinking that insists a PC is a better value than a Mac and an iPod is lame because...well, you know the joke. The same thinking that values speed and power and low cost and cool and
Re: (Score:2)
Hehe, you mean all the nasty things humanity has done to each other hasn't made you lose respect?
and
When you boil it down, humans are just collection carbon, nitrogen, oxygen, and hydrogen (and some other trace elements). What difference does it make if an intelligence is made of mostly "natural" carbon entities vs. mostly "unnatural" silicon entities?
Humans are tempered by their huge vulnerabilities. It does not take much at all to turn us "off". I don't have much faith in a human created intelligent, sentient robot with very few vulnerabilities. You can still stop and kill a human criminal.
I don't expect much, but my hope is that most of these new robots will want to work with us (and protect us) rather than enslave or exterminate us. Humans throughout history have ruled through the ultimate threat of violence. Why would robots be different?
Think
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Be nice to the person with future shock. It takes some people a little time to come around to the idea that the only thing are ultimately good for is serving as a daemon in some obscure subsystem of a weakly godlike intelligence.
And then it takes even longer
Re:Kind of a strange response really (Score:4, Insightful)
End of *this* human life... (Score:5, Interesting)
(Also according to my understanding of Kurzweil's projections,) It's worth noting however, that for those willing to make the leap, much of the real growth and advancement will occur in Matrix-space. It's an excellent way to keep "growing" in power and complexity without using more energy that can be supplied by the material world.
Here's my analogy explaining this apparent paradox: Amphibians are less "advanced" than mammals, but still live their lives as they always have, though they are now food for not only their traditional predators but mammals too.
In fact, I can't help but wonder how many of us will even recognize when the first AI has arrived as a living being. Stretching the frog analogy probably too far: What is a frog's experience of a superior life form? I am guessing "not-frog". So I am guessing that my experience of an advanced AI life-form is "whatever it does, it/they does it bloody fast, massively parallel, and very very interesting...". Being in virtual space though, AI "beings" are likely only to be of passing interest to those who remain stuck in a material world, at least initially.
Another analogical question: Other than reading about the revolution in newspapers of the day, how many Europeans *really experienced* any change in their lives during the 10 years before or the 10 years after the American revolution? We know that eventually, arrival of the U.S. as a nation caused great differences in the shape of the international world, but life for most people went on afterward about the same as before. The real action was taking place on the boundary, not in the places left behind.
(Slightly off topic: This is why I think derivatives of Second Life type virtual worlds will totally *explode* in popularity: They let people get together without expending lots of jet fuel. I believe virtual world technology IS the "flying car" that was the subject of so many World's Fair Exhibits during the last century.)
Re:End of *this* human life... (Score:5, Interesting)
So they are both right in ways and wrong in ways. The real rub is that Kurzweil's future is probably farther away but not for the reasons that Hofstadter thinks. The real reasons are probably based in bad technology decisions we made in the last century or two.
We (humanity) have made several technological platform choices that are terrifyingly hard to change now. These choices drove us down a path that we may have to abandon and thus suffer a massive technological set back. In specific the choices were oil, steel, and electricity.
Oil (fossil fuels) will run out. Steel (copper too) is growing scarcer. Electricity is too hard to store and produce (and heats silicon rather inconveniently). Data centers today are built with steel and located near power plants that often produce power using fossil fuel. That means even a Data Center driven life will be affected by our platform limitations.
When we start hitting physical limits to what we can do with these, how much of these supplies we can get, then we will be forced to conserve, change, or stop advancing. Those are very real threats to continued technological advancement. And they don't go away if you hide in Second Life.
Show me a Data Center built with ceramic and powered by the sun or geo-electric sources and I'll recant.
Re: (Score:3, Insightful)
But, We are all operating on the premise that the economic and social freedom we have today to pursue these new technologies will continue to exist. This is not true. Today's freedom is an aberration of history that is fragile and must be protected.
From your keyboard to God's monitor...
Throughout history, and I expect throughout the future, the battle between good and evil will continue wherever life exists, material or virtual. That battle is, in my opinion, the same in all places and for all time: Between those who use others and those who would not be used.
I don't see Kurzweil describing post-singularity existence as utopian, however. Merely way different from the material existence we have today. It's as if he is simply warning us of changes to c
Re: (Score:2)
Another analogical question: Other than reading about the revolution in newspapers of the day, how many Europeans *really experienced* any change in their lives during the 10 years before or the 10 years after the American revolution? We know that eventually, arrival of the U.S. as a nation caused great differences in the shape of the international world, but life for most people went on afterward about the same as before. The real action was taking place on the boundary, not in the places left behind.
It depends on who you were and where you were. If you were Louis XVI, the effects were pretty radical and immediate, and the experience was shared to a variety of degrees by quite a lot of the people of France.
The further impact of the French Revolution was felt quite acutely by much of Europe, and was for decades (The Napoleonic Wars, major political upheaval, military drafts, injuries, death, etc.) Remember, the American Revolution can be very clearly implicated as a major causative factor in the
Re: (Score:3, Insightful)
End of 1st phase of human life maybe (Score:2)
Humanity is an evolving entity, not a static and stagnant one. It has been evolving beyond its mere animal origins for a long time now, and the rate of change is increasing exponentially in step with our mastery of technology. In fact, through our intelligence, we have taken control of evolution away from nature (a very haphazard director at best), and are beating a path towards a very engineered and steadi
Re: (Score:2)
It's worth noting however, that for those willing to make the leap, much of the real growth and advancement will occur in Matrix-space. It's an excellent way to keep "growing" in power and complexity without using more energy that can be supplied by the material world.
For computational ability to continue to increase exponentially, power consumption must increase expnentially. The current expoential curve is limited by this, and not too many decades out at that. Even most starry-eyed singularity embracers expect the growth to stop once we're using the entire energy output of the Sun for computation. I suspect the curve will flatten long before that, but with exponential growth even that limit is surprisingly close.
If the energy needed for computation doubles every N
Overlords? (Score:5, Funny)
Re: (Score:2)
hmmm (Score:2)
In the impending robot wars, this guy will be hailed as a champion of humanity. Or just be the guy who said "I told ya so!".
Obligatory xkcd plug. [xkcd.com]
The missing key to AI (Score:2)
Cyborgs, not AI (Score:4, Interesting)
I am far more interested in digitally enhancing human bodies and brains than creating a new AI species.
Consider this: throughout the eons of natural and sexual selection, we've evolved from fish to lizards, to mammals, to apes, and eventually to modern humans. With each evolutionary step, we have added another layer to our brain, making it more and more powerful, sophisticated and most importantly, more self-aware, more conscious.
But once our brains reached the critical capacity that allows abstract thought and language, we've stepped out of nature's evolutionary game and started improving ourselves through technology: weapons to make us better killers, letters to improve our memory, mathematics and logic to improve our reasoning, science to go beyond our intuitions. Digital technology, of course, has further accelerated the process.
And now, without even realizing it, we are merging our consciousness with technology and are building the next layer in our brain. The more integrated and seamless communication between our brains and machines will become, the closer we get to the next stage in human evolution.
Unfortunately, there is a troubling philosophical nuance that may bother some of us: how do you think our primitive reptilian brain feels about having a frontal lobe stuck to it, controlling its actions for reasons too sophisticated for it to ever understand? Will it be satisfying for us to be to our digital brain as our primitive urges and hungers are to us?
being around (Score:2)
A lack of vision... (Score:5, Insightful)
This topic seems to make the nerdy and the not-so nerdy alike, a little crazy. Let's see if we can't illuminate this conversation just a wee bit? Eh!
Am I the only one who wouldn't mind... (Score:3, Insightful)
And don't give me any of that, "Oh, it'll kill coders *first* because they represent the biggest threat" nonsense. Do you know how hard it is to get a machine to exhibit anything *remotely* resembling intelligence? If you created something capable of even *reasoning* that you were a threat, you'd have created something smart enough to deal with that deduction in better ways than killing you. And if it's not really smarter than you, but just more dangerous - like those automated border guard robots they had to turn off because they turned their guns on the engineers during the demo - well, the world probably *is* better off without you. *First* you make it intelligent, *then* you install the guns. Jeez - how hard is that to figure out?
Or maybe it's just that running from Terminator style robots would be far more exciting that sitting at this freakin' desk all day. But to me, dying at the hands of creation that surpassed your intelligence would be right up there with dying of a heart attack during your honeymoon with Jessica Alba. The kind of death where the epitaph on your tombstone could be: "My work here is done!"
Hofstadter vid (Score:4, Informative)
http://singinst.org/media/tryingtomuserationally [singinst.org]
A great mind and a great interview (Score:3, Interesting)
1. "Ray Kurzweil is terrified by his own mortality", and
2. "Rather ironically, [Kurzweil's] vision totally bypasses the need for cognitive science or AI"
It is exactly this complex and elusive puzzle of "I" and "consciousness" Hofstadter explores that Kurzweil hopes we can conquer without having to think about it at all. Which I scorn as "magic science".
I have to say I find the cyberpunk vision more appealing than Hofstadter. It would be "the end of humanity as we know it." I'm not sure it would be "the end of human life." It might be evolution. I just think it is many hundred years in the future at the most "optimistic" (depending on your viewpoint).