Reply to Cheeseman's \An Inquiry into Computer. This paper covers a fairly wide range of issues, from a basic review of probability theory

Similar documents
A New Parameter for Maintaining Consistency in an Agent's Knowledge Base Using Truth Maintenance System

Russell on Denoting. G. J. Mattey. Fall, 2005 / Philosophy 156. The concept any finite number is not odd, nor is it even.

Philosophy 1100: Introduction to Ethics. Critical Thinking Lecture 2. Background Material for the Exercise on Inference Indicators

All They Know: A Study in Multi-Agent Autoepistemic Reasoning

A Solution to the Gettier Problem Keota Fields. the three traditional conditions for knowledge, have been discussed extensively in the

Artificial Intelligence I

Divine omniscience, timelessness, and the power to do otherwise

agents, where we take in consideration both limited memory and limited capacities of inference. The classical theory of belief change, known as the AG

Oxford Scholarship Online Abstracts and Keywords

Artificial Intelligence Prof. Deepak Khemani Department of Computer Science and Engineering Indian Institute of Technology, Madras

Detachment, Probability, and Maximum Likelihood

Understanding Truth Scott Soames Précis Philosophy and Phenomenological Research Volume LXV, No. 2, 2002

Quine: Quantifiers and Propositional Attitudes

Module 5. Knowledge Representation and Logic (Propositional Logic) Version 2 CSE IIT, Kharagpur

A CRITIQUE OF THE FREE WILL DEFENSE. A Paper. Presented to. Dr. Douglas Blount. Southwestern Baptist Theological Seminary. In Partial Fulfillment

Harry Collins and the Crisis of Expertise. Ylikoski, Petri Kullervo.

5 A Modal Version of the

Rules for NZ Young Farmers Debates

Artificial Intelligence Prof. P. Dasgupta Department of Computer Science & Engineering Indian Institute of Technology, Kharagpur

Macmillan/McGraw-Hill SCIENCE: A CLOSER LOOK 2011, Grade 3 Correlated with Common Core State Standards, Grade 3

CS-TR-3278 May 26, 1994 LOGIC FOR A LIFETIME. Don Perlis. Institute for Advanced Computer Studies. Computer Science Department.

9 Knowledge-Based Systems

The Representation of Logical Form: A Dilemma

Since Michael so neatly summarized his objections in the form of three questions, all I need to do now is to answer these questions.

The Problem of the External World

Introduction Questions to Ask in Judging Whether A Really Causes B

Overview: Application: What to Avoid:

NON-NUMERICAL APPROACHES TO PLAUSIBLE INFERENCE

(Refer Slide Time 03:00)

Macmillan/McGraw-Hill SCIENCE: A CLOSER LOOK 2011, Grade 4 Correlated with Common Core State Standards, Grade 4

Programme. Sven Rosenkranz: Agnosticism and Epistemic Norms. Alexandra Zinke: Varieties of Suspension

ONTOLOGICAL PROBLEMS OF PLURALIST RESEARCH METHODOLOGIES

4.7 Constructing Categorical Propositions

PROSPECTIVE TEACHERS UNDERSTANDING OF PROOF: WHAT IF THE TRUTH SET OF AN OPEN SENTENCE IS BROADER THAN THAT COVERED BY THE PROOF?

Logical Omniscience in the Many Agent Case

I think, therefore I am. - Rene Descartes

Module - 02 Lecturer - 09 Inferential Statistics - Motivation

Bertrand Russell Proper Names, Adjectives and Verbs 1

Logical (formal) fallacies

Verificationism. PHIL September 27, 2011

Negative Introspection Is Mysterious

Other Logics: What Nonclassical Reasoning Is All About Dr. Michael A. Covington Associate Director Artificial Intelligence Center

Page 2 of 8 Stage 2 Religion Studies Student Response

Informalizing Formal Logic

what makes reasons sufficient?

Thomas Hobbes, Leviathan, ed. by Noel Malcolm, Clarendon Edition of the Works of Thomas Hobbes, 3 vols., Oxford: Clarendon Press, 2012

Diocese of Cheyenne Diaconal Formation Program General Requirements and Expectations

THE ROLE OF COHERENCE OF EVIDENCE IN THE NON- DYNAMIC MODEL OF CONFIRMATION TOMOJI SHOGENJI

McCLOSKEY ON RATIONAL ENDS: The Dilemma of Intuitionism

Ayer on the criterion of verifiability

Situations in Which Disjunctive Syllogism Can Lead from True Premises to a False Conclusion

The cosmological argument (continued)

HANDBOOK (New or substantially modified material appears in boxes.)

Lecture 3. I argued in the previous lecture for a relationist solution to Frege's puzzle, one which

Macmillan/McGraw-Hill SCIENCE: A CLOSER LOOK 2011, Grade 1 Correlated with Common Core State Standards, Grade 1

Prentice Hall U.S. History Modern America 2013

Steven O. Kimbrough. University of Pennsylvania. Operations and Information Management Dept. Philadelphia, PA (215) 898{5133

1 Clarion Logic Notes Chapter 4

Honours Programme in Philosophy

Falsification or Confirmation: From Logic to Psychology

Philosophy of Science. Ross Arnold, Summer 2014 Lakeside institute of Theology

III Knowledge is true belief based on argument. Plato, Theaetetus, 201 c-d Is Justified True Belief Knowledge? Edmund Gettier

Skepticism and Internalism

1/19/2011. Concept. Analysis

1/5. The Critique of Theology

Proof as a cluster concept in mathematical practice. Keith Weber Rutgers University

Artificial Intelligence Prof. Deepak Khemani Department of Computer Science and Engineering Indian Institute of Technology, Madras

Intro Viewed from a certain angle, philosophy is about what, if anything, we ought to believe.

Class 33 - November 13 Philosophy Friday #6: Quine and Ontological Commitment Fisher 59-69; Quine, On What There Is

Rawls s veil of ignorance excludes all knowledge of likelihoods regarding the social

D:D(φ) B: B(φ) I:I(φ) I:I(does(e)) C:does(e) C:done(e) B:B(done(e))

Inference in Cyc. This is the final lesson in the Inference Tutorial. It will focus on microtheories and forward/backward inference.

DEPARTMENT OF PHILOSOPHY FALL 2014 COURSE DESCRIPTIONS

Williams on Supervaluationism and Logical Revisionism

Haberdashers Aske s Boys School

The Critical Mind is A Questioning Mind

A Liar Paradox. Richard G. Heck, Jr. Brown University

Unit. Science and Hypothesis. Downloaded from Downloaded from Why Hypothesis? What is a Hypothesis?

A Model of Decidable Introspective Reasoning with Quantifying-In

Christ-Centered Preaching: Preparation and Delivery of Sermons Lesson 6a, page 1

McDougal Littell High School Math Program. correlated to. Oregon Mathematics Grade-Level Standards

Introduction to Philosophy Crito. Instructor: Jason Sheley

Chadwick Prize Winner: Christian Michel THE LIAR PARADOX OUTSIDE-IN

Artificial Intelligence Prof. P. Dasgupta Department of Computer Science & Engineering Indian Institute of Technology, Kharagpur

Belief as Defeasible Knowledge

SUPPOSITIONAL REASONING AND PERCEPTUAL JUSTIFICATION

Symbolic Logic Prof. Chhanda Chakraborti Department of Humanities and Social Sciences Indian Institute of Technology, Kharagpur

Logical Puzzles and the Concept of God

* I am indebted to Jay Atlas and Robert Schwartz for their helpful criticisms

Predictability, Causation, and Free Will

SMITH ON TRUTHMAKERS 1. Dominic Gregory. I. Introduction

BEFORE THE MINNESOTA OFFICE OF ADMINISTRATIVE HEARINGS 600 North Robert Street St. Paul, MN 55101

COACHING THE BASICS: WHAT IS AN ARGUMENT?

Building Your Framework everydaydebate.blogspot.com by James M. Kellams

A Scientific Realism-Based Probabilistic Approach to Popper's Problem of Confirmation

Philosophy 1100: Introduction to Ethics. Critical Thinking Lecture 1. Background Material for the Exercise on Validity

1.5 Deductive and Inductive Arguments

In this set of essays spanning much of his career at Calvin College,

A Brief History of Scientific Thoughts Lecture 5. Palash Sarkar

The SAT Essay: An Argument-Centered Strategy

Transcription:

Reply to Cheeseman's \An Inquiry into Computer Understanding" This paper covers a fairly wide range of issues, from a basic review of probability theory to the suggestion that probabilistic ideas can be used to solve the learning problem. Much of the material is therefore quite old hat, and I will limit my response to the portion of the paper that strikes me as original. In this regard, I believe Cheeseman's main points to be the following: The repeated claim that, \McDermott has shown that a direct translation of commonsense reasoning into logical form leads to unsurmountable diculties" [Abstract, and following]. A series of suggestions [Section 2] that probabilities are a good idea, because: { It is important to be able to label statements as other than simply unquestionably true or unquestionably false, { These labels should indicate something about the contexts in which they are valid, and { These labels should include numeric information. That the assignment of context-dependent labels can help solve the problem of referential opacity [Section 2]. That the priors problem can be solved using the existence of \objective priors available for most of the commonly encountered situations" [Section 2] such as Jerey's priors [Section 3]. That the use of probabilities solves the learning problem [Section 4]. That \logic is just a special case within probability theory, but the converse is not true." [Section 5]. That probabilities can be used to solve the ravens paradox and related diculties [Section 5]. That a randomly selected passage from [3] deals with probabilistic information [Section 6]; Cheeseman attempts to infer the ubiquity of probabilistic reasoning from this. I guess I can manage to nd something to respond to in this. 1

McDermott's paper Cheeseman's remark that, \McDermott has shown that a direct translation of commonsense reasoning into logical form leads to unsurmountable diculties," is simply false. I am reluctant to enter, at this late date, into the debate on McDermott's article, but it seems to me that what McDermott was arguing (and I agree with him) was that none of the existing methods suggested for nonmonotonic reasoning is satisfactory. McDermott appears to nd this disheartening, but nowhere does he make the claim that the logical approach is doomed. He argues that its lack of success makes it suspect, perhaps, and that alternatives (such as proceduralism, and probabilistic methods) should be tried, but nowhere does he seem to say that the logicists' approach is fundamentally without hope. And how could he? Simply because no one has managed to make it work yet is slim evidence from which to conclude that no one will in the future. Additional truth values Here, I am at least partially in agreement with Cheeseman's views. I have been suggesting for some time [2, and subsequent papers] that it is important to be able to label sentences more descriptively than simply as \true" or \false." In many cases, it is useful for these labels to include information regarding the context or contexts in which a particular conclusion is valid; dekleer's atms's [1] are a typical example. But there is a substantial computational cost associated to maintaining these labels, and I see no justication for the belief that it is always right to include contextual information in the truth values. I also see very little evidence to support Cheeseman's claim that there should be a numerical component to the truth value. Certainly this will be appropriate in many situations. But the objections to the probabilistic approach really do have merit: what, for example, is the probability that an astronaut walking on the surface of Mars will encounter a lion? McCarthy has argued that default logics (if we can ever get them to work) will treat the probability of a statement such as this one simply as, which seems a more reasonable approach than the assignment of a specic numeric value. Once again, it seems that Cheeseman is jumping from the observation that an approach is useful in some instances to the conclusion that it should be used all the time. He does point us to a reference (his paper, \In defense of probability") that addresses this issue, but this seems too crucial a point to be skipped over so lightly in the current work. Referential opacity Cheeseman argues here that, \if all real world propositions are statements of belief, as required in probability theory, then... the problem of referential opacity never arises." This argument is made without support. There is probably a germ of truth to what Cheeseman is saying, in that it is possible to overcome some of the diculties of referential opacity by keeping track of the believer of any particular sentence. But this is simply an argument for maintaining contextual information in the truth values { and only an argument for doing so in special cases. Once again, I see no justication for Cheeseman's suggestion that the problem of referential opacity is in some way an argument for the use of probability theory. 2

Priors Here, too, I nd Cheeseman's claims unconvincing. The sharpest one I can nd is the claim from Section 3 that a Martian, faced with the problem of determining a prior probability distribution for the heights of Earthlings, could use the \truely [sic] noninformative prior" of (1=x)dx, \based only on the information that height is always positive and that we want the nal answer to be independent of scale." What is x here? Height? Volume? Is the prior probability supposed to be noninformative with respect to the height of Earthlings, or with respect to their volume? (After all, the size of land creatures is limited much more by volume than by height.) The \priors problem" is not that one cannot write down a probability distribution that is noninformative with respect to some particular measurement, but that it is impossible to produce one that is noninformative with respect to all possible measurements { and that deciding which measurement should be unbiased is a dicult and open problem. Learning Cheeseman tells us that, \What makes the untutored learning problem dicult (at least for current ad hoc AI methods) is that the number of classes (if any) and their denition is unknown." Nothing could be further from the truth. What makes learning dicult is that the search space is simply too large for any sort of exhaustive search to be practical, and that the nature of the search space is such that heuristic techniques such as hill climbing are unlikely to be eective. The approach Cheeseman suggests to the learning problem does not address these dif- culties at all. In fact, the procedure he suggests for traversing the search space is simple hill climbing; Cheeseman defends this by arguing that, \a search for the global maximum is usually computationally too expensive." He's right, of course. But progress in learning will be made when people nd ways to eectively reduce this computational cost, and the application of probabilities to the problem makes no contribution in this regard. The learning argument also lacks the support of examples of the method at work, although Cheeseman once again points us elsewhere to nd them (this time, to his paper \Automatic discovery of optimal classes"). As with the statement that numeric truth values are to be preferred over symbolic ones, this is simply too important a point for us to be expected to accept it on faith. Logic and probability Cheeseman would also have us believe that, \Probability can be regarded as more fundamental than logic, because logic is just a special case within probability (where all probabilities are 0 or 1), but the converse is not true." Remarkably, the sentence that precedes this one begins, \Similarly, although probability can be embedded in logic... " Cheeseman argues that probability is not a special case of logic because the embedding to which he refers adds too much baggage to the logical approach for the result to be accurately called \logic". The same thing can be said of probability where all of the probabilities are 0 or 1, however: here, probability theory adds a tremendous amount of baggage (the dependence on contexts, for example) not needed by the logicists. Both probability theory and rst-order logic are universal, and each is capable of describing the other. The suggestion that either approach has some fundamental advantage over 3

the other is simply misguided. The ravens paradox Cheeseman goes on to suggest that a probabilistic approach avoids the ravens paradox. In reality, however, he does not discuss the ravens paradox at all, but instead investigates another diculty of his own invention. The ravens paradox is the following: Suppose that we denote the ravens sentence, \All ravens are black," by r. Now after accumulating some amount of evidence A, we conclude that r has some probability x: p(rja) = x: In other words, the probability that all ravens are black, given the observed evidence, is x. Now consider the companion statement, \All non-black things are not ravens," which we denote by r. Since r and r are equivalent, we must also have p(rja) = x; by virtue of the \consistency" condition adopted by Cheeseman himself in Section 2 of the paper. This is the ravens paradox: evidence for the statement that all ravens are black is also evidence for the statement that all non-black things are not ravens. Cheeseman attempts to sidestep this diculty by arguing that we are more interested in the conditional probability p(black(x)jraven(x)) than we are interested in the probability of the conditional r. But if, as in the original ravens paradox, we really are concerned with the probability of the conditional, Cheeseman's arguments do not apply. Once again, he is using probabilities to answer a question other than the one of interest. Probabilistic information in English Finally, Cheeseman examines a passage from Nilsson's introductory text on AI. The sentence that he cites as probabilistic is the following: To inform a control system completely about the problem domains of interest in AI typically involves a high-cost strategy, in terms of the storage and computations required. Cheeseman says of this sentence, \This is an example of probabilistic information where the reader is informed of a likely consequence of the model under particular circumstances... " I see no probabilistic information here. The reader is being informed of a default rule. If anything, the form of Nilsson's sentence is an argument against the probabilistic approach, since Nilsson seems to feel it sucient to tell his reader of the presence of the default without providing any way to assign a specic numeric measure of certainty to the conclusion. 4

Summary In sum, I can nd very little with which to agree in Cheeseman's paper. I do agree that probabilities can be used to solve his version of the ravens paradox, and that they can be used to evaluate points in the search space encountered in learning. But these issues are not terribly interesting. Most of Cheeseman's other points strike me as either unsupported or simply wrong. The single exception is the claim that truth values should be chosen from a set containing more than two elements. But I see little intrinsic connection between this claim and probabilistic issues. References [1] J. de Kleer. An assumption-based truth maintenance system. Articial Intelligence, 28:127{162, 1986. [2] M. L. Ginsberg. Multi-valued logics. In Proceedings of the Fifth National Conference on Articial Intelligence, pages 243{247, 1986. [3] N. J. Nilsson. Principles of Articial Intelligence. Tioga, Palo Alto, 1980. 5