“Yes, that’s so.”
“Then, if he needs the money so badly, all he need do is not bet at all, and it would be just as though he had broken even.”
“Yes, but he has this unreasoning need to gamble.”
“You mean even if he loses.”
“Yes.”
“But that makes no sense.”
“But the point of the joke is that the gambler doesn’t understand this.”
“You mean it’s funny if a person lacks any sense of logic and is possessed of not even the simplest understanding?”
And what can you do but turn back to building the house again?
But tell me, is this so different from dealing with the ordinary humorless human being? I once told my father this joke:
Mrs. Jones, the landlady, woke up in the middle of the night because there were strange noises outside her door. She looked out, and there was Robinson, one of her boarders, forcing a frightened horse up the stairs.
She shrieked, “What are you doing, Mr. Robinson?”
He said, “Putting the horse in the bathroom.”
“For goodness sake, why?”
“Well, old Higginbotham is such a wise guy. Whatever I tell him, he answers, ‘I know. I know,’ in such a superior way. Well, in the morning, he’ll go to the bathroom and he’ll come out yelling, ‘There’s a horse in the bathroom.’ And I’ll yawn and say, ‘I know, I know.’ “
And what was my father’s response? He said, “Isaac, Isaac. You’re a city boy, so you don’t understand. You can’t push a horse up the stairs if he doesn’t want to go.”
Personally, I thought that was funnier than the joke.
Anyway, I don’t see why we should particularly want a robot to have a sense of humor, but the point is that the robot himself might want to have one-and how do we give it to him?
Robots In Combination
I have been inventing stories about robots now for very nearly half a century. In that time, I have rung almost every conceivable change upon the theme.
Mind you, it was not my intention to compose an encyclopedia of robot nuances; it was not even my intention to write about them for half a century. It just happened that I survived that long and maintained my interest in the concept. And it also just happened that in attempting to think of new story ideas involving robots, I ended up thinking about nearly everything.
For instance, in the sixth volume of the Robot City series, there are the “chemfets,” which have been introduced into the hero’s body in order to replicate and, eventually, give him direct psycho-electronic control over the core computer, and hence all the robots of Robot City.
Well, in my book Foundation’s Edge (Doubleday, 1982), my hero, Golan Trevize, before taking off in a spaceship, makes contact with an advanced computer by placing his hands on an indicated place on the desk before him.
“And as he and the computer held hands, their thinking merged…
“…he saw the room with complete clarity-not just in the direction in which he was looking, but all around and above and below.
“He saw every room in the spaceship, and he saw outside as well. The sun had risen…but he could look at it directly without being dazzled…
“He felt the gentle wind and its temperature, and the sounds of the world about him. He detected the planet’s magnetic field and the tiny electrical charges on the wall of the ship.
“He became aware of the controls of the ship…He knew…that if he wanted to lift the ship, or turn it, or accelerate, or make use of any of its abilities, the process was the same as that of performing the analogous process to his body. He had but to use his will.”
That was as close as I could come to picturing the result of a mind-computer interface, and now, in connection with this new book, I can’t help thinking of it further.
I suppose that the first time human beings learned how to form an interface between the human mind and another sort of intelligence was when they tamed the horse and learned how to use it as a form of transportation. This reached its highest point when human beings rode horses directly, and when a pull at a rein, the touch of a spur, a squeeze of the knees, or just a cry, could make the horse react in accordance with the human will.
It is no wonder that primitive Greeks seeing horsemen invade the comparatively broad Thessalian plains (the part of Greece most suitable to horsemanship) thought they were seeing a single animal with a human torso and a horse’s body. Thus was invented the centaur.
Again, there are “trick drivers.” There are expert “stunt men” who can make an automobile do marvelous things. One might expect that a New Guinea native who had never seen or heard of an automobile before might believe that such stunts were being carried through by a strange and Monstrous living organism that had, as part of its structure, a portion with a human appearance within its stomach.
But a person plus a horse is but an imperfect fusion of intelligence, and a person plus an automobile is but an extension of human muscles by mechanical linkages. A horse can easily disobey signals, or even run away in uncontrollable panic. And an automobile can break down or skid at an inconvenient moment.
The fusion of human and computer, however, ought to be a much closer approach to the ideal. It may be an extension of the mind itself as I tried to make plain in Foundation’s Edge, a multiplication and intensification of sense-perception, an incredible extension of the will.
Under such circumstances, might not the fusion represent, in a very real sense, a single organism, a kind of cybernetic “centaur”? And once such a union is established, would the human fraction wish to break it? Would he not feel such a break to be an unbearable loss and be unable to live with the impoverishment of mind and will he would then have to face? In my novel, Golan Trevize could break away from the computer at will and suffered no ill effects as a result, but perhaps that is not realistic.
Another issue that appears now and then in the Robot City series concerns the interaction of robot and robot.
This has not played a part in most of my stories, simply because I generally had a single robot character of importance in any given story and I dealt entirely with the matter of the interaction between that single robot and various human beings.
Consider robots in combination.
The First Law states that a robot cannot injure a human being or, through inaction, allow a human being to come to harm.
But suppose two robots are involved, and that one of them, through inadvertence, lack of knowledge, or special circumstances, is engaged in a course of action (quite innocently) that will clearly injure a human being-and suppose the second robot, with greater knowledge or insight, is aware of this. Would he not be required by the First Law to stop the first robot from committing the injury? If there were no other way, would he not be required by the First Law to destroy the first robot without hesitation or regret?
Thus, in my book Robots and Empire (Doubleday, 1985), a robot is introduced to whom human beings have been defined as those speaking with a certain accent. The heroine of the book does not speak with that accent and therefore the robot feels free to kill her. That robot is promptly destroyed by a second robot.
The situation is similar for the Second Law, in which robots are forced to obey orders given them by human beings provided those orders do not violate the First Law.
If, of two robots, one through inadvertence or lack of understanding does not obey an order, the second must either carry through the order itself, or force the first to do so.
Thus, in an intense scene in Robots and Empire, the villainess gives one robot a direct order. The robot hesitates because the order may cause harm to the heroine. For a while, then, there is a confrontation in which the villainess reinforces her own order while a second robot tries to reason the first robot into a greater realization of the harm that will be done to the heroine. Here we have a case where one robot urges another to obey the Second Law in a truer manner, and to withstand a human being in so doing.