When we communicate through (natural) languages, we do not explicitly say everything. Both the speaker and the hearer utilize information available from the utterance situation, which includes the mental states of the speaker and the hearer. Interesting cases are frequently observed in the use of Japanese (in dialogue situations). Syntactic (or configurational) constraints of Japanese are weaker than those of English, in the sense that the speaker may omit almost any element in a sentence. In this paper we present a mechanism of the hearer in the light of situated reasoning and show how the missing information can be supplied from the situation. Although we believe that the model captures the essential nature of human communication, it may be too naive as a model of human cognition. Rather, the model is intended to be used in the design of software agents that communicate with each other in a mechanical but flexible and efficient way.
ASJC Scopus subject areas
- Modelling and Simulation
- Language and Linguistics
- Linguistics and Language
- Computer Vision and Pattern Recognition
- Computer Science Applications