Summary form only given, as follows. The authors propose a semantic generalized predicate logic that is based on probability theory and information theory for providing theoretical methods for processing uncertain knowledge in artifical intelligence (AI) applications. The basic concept of the proposed logic is that the interpretation of the well-formed formula (wff) containing uncertainty is represented by using conditional probability. By the interpretation model using the conditional probability, a lot of problems that are impossible to treat by conventional AI methods can be explained in terms of information theory. From the definition, the self-information of the wff, the mutual information between a couple of predicates, and information gain by the reasoning can be shown. Next, reasoning rules are evaluated using the information gain which expresses the difference between the prior information and the posterior information of the consequent wff. Finally, the authors give a new calculation method for reasoning that gives the most unbiased probability estimation, given the available evidence, and prove that the proposed method is optimal from the principle of maximum entropy, subject to the given marginal probability condition.
|Title of host publication||IEEE 1988 Int Symp on Inf Theory Abstr of Pap|
|Place of Publication||New York, NY, USA|
|Publisher||Publ by IEEE|
|Number of pages||2|
|Volume||25 n 13|
|Publication status||Published - 1988|
ASJC Scopus subject areas