| T.R | Title | User | Personal Name
 | Date | Lines | 
|---|
| 412.1 | All Sentience is created equal, and they are endowed... | CACHE::MARSHALL | hunting the snark | Thu Nov 20 1986 08:46 | 21 | 
|  |     re .0:
    
    first a minor semantic objection.
    
> 2) A sentient computer - one able to pass the Turing test for at 
>    least, say, an hour.
   
    A SENTIENT computer would pass the Turing test forever. I would
    not consider sentient a computer that could only pass it for a
    relatively short time.
    
    However, I know what you mean, and they are very good questions.
    I think that I shall have to mull it over more but my initial response
    is Yes....No....ummm....well....very tough question.
    
                                                   
                  /
                 (  ___
                  ) ///
                 /
    
 | 
| 412.2 | Human is as Human Does | PROSE::WAJENBERG |  | Thu Nov 20 1986 14:39 | 24 | 
|  |     Re .0
    
    I would say that we should accord them rights in proportion to the
    responsibility they exhibit.  That's just a very rough guiding
    principle of course.  Perhaps one could administer tests similar
    to the sanity tests they use to test the legal competence of
    standard-issue human beings.  (Of course, I have heard those tests
    loudly criticized, too.)
    
    Re .1
    
    To pitch the nit back to you, a computer might be *sentient* and
    be very poor at the Turing test.  "Sentient" just means "aware"
    or "conscious."  The quality of the sentience's emotions and thoughts
    might not be human-like at all.  The Turing test, as I recall, demanded
    that the computer act just the same as a human being at the other
    end of a teletype.  (In fact, a notesfile would be a grand arena
    in which to run a Turing test.)  A computer might be conscious and
    volitional but still have a very non-human personality.  HAL 9000
    might be a good fictional example.  He was responsive and intelligent
    in his actions, but was depicted as having a manner and delivery
    (and a depth psychology) markedly unhuman.
    
    Earl Wajenberg
 | 
| 412.3 | More Questions | DRUMS::FEHSKENS |  | Thu Nov 20 1986 15:19 | 35 | 
|  |     re .2 re .0 - this approach would deny rights to a large number
    of people.  Many people are seriously irresponsible, but we grant
    them "standard-issue" rights almost solely because they are "packaged"
    the right way (i.e., live in a human body).  There is a certain
    amount of "species chauvinism" here.  
    
    In particular, given much recent discussion of abortion,
    it raises some interesting questions about the rights of fetuses,
    which exhibit less responsibility than many dogs I have known.
    (Please note that this statement is clearly true, and says nothing
    whatsoever about my position on abortion, so let's not go off track
    by getting into a discussion about abortion or my position thereon).
    This is an enormously difficult question, one that people are not
    predisposed to think rationally about (perhaps because "rationality"
    is not the relevant issue).  All I'm trying to call attention to
    is the futility of "simple" answers.
    
    What does "sentient" mean?  Is there some clearly defined threshold?
    It seems to me that's part of the problem - you can't just draw
    a line and say "on this side, your "life" is unimportant, but on
    that side it is".  What does "aware" mean?  Is a thermostat "unaware"
    of temperature?  Animals are clearly "conscious", because you can
    anesthetize thme and render them unconscious, and there's a dramatic
    behavioural difference between the two states.  Do you mean self
    aware or self conscious?  What about the experiments with chimps
    and mirrors?  Or is the issue, as I have heard argued, awareness
    of death and its meaning to the individual?
    
    And what does important mean?  There are rights,
    and privileges, and responsibilities, both of and to the individual
    in question.  We could debate this issue forever and only succeed
    in aggravating one another.
    
    len.
    
 | 
| 412.4 | Non-binary Coding | PROSE::WAJENBERG |  | Thu Nov 20 1986 17:12 | 16 | 
|  |     Re .3
    
    I agree simple answers are no good.  If I have seemed over-simple
    it is because of limited typing time.  One over-simple thing to
    avoid is the binary allocation of rights or no rights.  Ideally,
    rights should be awarded according to many different factors.  For
    instance, the right to be unmolested should be in some way proportional
    to the subject's sensitivity.  The right to exercise a given power
    should presumably entail a responsibility to use that power with
    a regard to the rights of others.  And so forth.
    
    Human legal practice recognizes this in a rough and ready way. 
    For instance, minors have some rights but not all the rights of
    an adult citizen.
    
    Earl Wajenberg
 | 
| 412.5 | Turing Test | CACHE::MARSHALL | hunting the snark | Thu Nov 20 1986 17:21 | 24 | 
|  |     re .2:
    
    The Turing Test is a test of sentience, not humanness. 
    A human may try to act like a computer to fool you. Hal would pass
    the Turing Test (I think). A computer may be programmed to appear
    to be spontaneous. 
    
    re .3:
    
    The points you bring up are exactly why I ducked the question. The
    questions are loaded, there is no simple answer. The being will
    have whatever rights it declares that it has, as long as it presents
    a reason. Rights are not given. Rights are claimed, or they are
    recognized. To possess rights does not require responsibility, to
    *exercise* those rights does. We do not "grant" rights to people
    because they are human, they possess rights because they are human.
    
    This question I think should be in PHILOSOPHY.
                                                   
                  /
                 (  ___
                  ) ///
                 /
    
 | 
| 412.6 | AI Aliens | ROCK::REDFORD | On a pure caffeine high | Thu Nov 20 1986 18:10 | 7 | 
|  | One other category they could fall into is foreigners.  Foreigners do 
not have privileges and duties with regard to the government of the 
land, but they are entitled to some legal protection.  This status 
recognizes that they are half way between soceities.  Immigration 
could be permitted from Dolphinia and Netland, but that's more of a 
political question.
/jlr
 | 
| 412.7 | First reply in SF | COMET::HUNTER | Nine o'Clock Meetings,A Real winner | Thu Nov 20 1986 18:12 | 2 | 
|  |      Very interesting discussion, I have to think about it awhile long
    before tackling it.
 | 
| 412.8 | More Turing Test | PROSE::WAJENBERG |  | Fri Nov 21 1986 09:45 | 10 | 
|  |     Re .5
    
    The Turing Test is testing sentience (sapience, actually, but that
    includes sentience), but the test is the ability to fake humanness.
    At least, that is how Turing first wrote it up.  If you want to
    propose a modified Turing Test with the simple goal "persuade me
    there is something conscious on the other end of this link," that's
    fine too.
    
    Earl Wajenberg
 | 
| 412.9 | just trying to understand | CACHE::MARSHALL | hunting the snark | Fri Nov 21 1986 10:02 | 29 | 
|  |     re .8:
    
    Earl, really, I'm not trying to argue with you. Maybe my understanding
    of the test is off, I have not read Turing's actual paper. I'm mostly
    familiar with the test through second sources, the major one being
    _Godel,_Escher,_Bach_.                    
                      
    As I understand it, the subject "passes" if the tester cannot
    decide whether or not the subject is a computer or a human. I do
    not think the tester must declare the subject to be a human in order
    to pass, the undecidability is sufficient.           
    
    Actually, I think the set-up for the test is to have three parties
    involved. the tester, the subject, and (for lack of a better word)
    a reference. The tester and the reference are human. the subject
    is the program or alien or whatever who's sentience we are trying
    to determine.
    
    All three are put before teletypes, and the tester is now periodically
    switched between the subject and the reference. If the tester cannot
    distinguish between the two, then the subject passes.
    
    Is this correct?
                                                   
                  /
                 (  ___
                  ) ///
                 /
    
 | 
| 412.10 | Parlor Games | PROSE::WAJENBERG |  | Fri Nov 21 1986 10:23 | 7 | 
|  |     Yes, that is my understanding of the Turing Test.  I have heard
    that Turing developed it from a Victorian parlor game in which the
    tester communicated with a man and a woman by passing notes under
    a door or some such.  Either the man pretended to be a woman or
    vice versa and the tester had to determine their real sexes.
    
    Earl Wajenberg
 | 
| 412.11 |  | INK::KALLIS | Support Hallowe'en | Fri Nov 21 1986 15:35 | 24 | 
|  |     Well, let's get back to the questions.
    
    Interestingly, in the 19th and early 20th Century, there was a belief
    among the population of tropical Africa that gorillas and chimpanzees
    were just as smart as anyone, but they stayed in the jungle to avoid
    having to go to work to earn money!  Sort of a "noble savage" concept,
    one imagines. 
    
    On cetacians, etc.  They're intelligent animals, and ought to be
    _respected_.  "Legal rights," one assumes, must be based on the
    presumption that they _want_ human rights.
    
    On intelligent computers, etc.  There was a marvelous Henry Kuttner
    story where the hero was talking to a robot that did the analog
    of "drinking" by sticking a metal finger in an active light socket.
    Slightly paraphrased:
    
    	There was a sharp crack and a flash of blue spark.  "Hmm," said
    the robot.  "DC.  Tasty."
        "You're not dead," said [the hero].
        "I'm not even alive," said the robot.
    
    Steve Kallis, Jr.
    
 |