ABSTRACT
Most large text-understanding systems have been designed under the assumption that the input text will be in reasonably "neat" form, e.g., newspaper stories and other edited texts. However, a great deal of natural language text, e.g., memos, rough drafts, conversation transcripts, etc., have features that differ significantly from "neat" texts, posing special problems for readers, such as misspelled words, missing words, poor syntactic construction, missing periods, etc. Our solution to these problems is to make use of <u>expectations</u>, based both on knowledge of surface English and on world knowledge of the situation being described. These syntactic and semantic expectations can be used to figure out unknown words from context, constrain the possible word-senses of words with multiple meanings (ambiguity), fill in missing words (ellipsis), and resolve referents (anaphora). This method of using expectations to aid the understanding of "scruffy" texts has been incorporated into a working computer program called NOMAD, which understands scruffy texts in the domain of Navy messages.
- Birnbaum, L. and Selfridge, M. 1980. Conceptual Analysis of Natural Language, in R. Schank and C. Riesbeck, eds., Inside Computer Understanding. Lawrence Erlbaum Associates, Hillsdale, N. J.Google Scholar
- Cullingford, R. 1977. Controlling Inferences in Story Understanding. Proceedings of the Fifth International Joint Conference on Artificial Intelligence (IJCAI), Cambridge, Mass.Google Scholar
- DeJong, G. 1979. Skimming Stories in Real Time: An Experiment in Integrated Understanding. Ph.D. Thesis, Yale Computer Science Dept. Google ScholarDigital Library
- Goldman, N. 1973. The generation of English sentences from a deep conceptual base. Ph.D. Thesis, Stanford University.Google Scholar
- Granger, R. 1977. FOUL-UP: A program that figures out meanings of words from context. Proceedings of the Fifth IJCAI, Cambridge, Mass.Google Scholar
- Granger, R. H. 1980. When expectation fails: Towards a self-correcting inference system. In Proceedings of the First National Conference on Artificial Intelligence, Stanford University.Google Scholar
- Granger, R. H. 1981. Directing and re-directing inference pursuit: Extra-textual influences on text interpretation. In Proceedings of the Seventh International Joint Conference on Artificial Intelligence (IJCAI), Vancouver, British Columbia.Google Scholar
- Lebowitz, M. 1981. Generalization and Memory in an Integrated Understanding System. Computer Science Research Report 186, Yale University.Google Scholar
- McGuire, R. 1980. Political Primaries and Words of Pain. Unpublished ms., Dept. of Computer Science, Yale University.Google Scholar
- Riesbeck, C. and Schank, R. 1976. Comprehension by computer: Expectation-based analysis of sentences in context. Computer Science Research Report 78, Yale University.Google Scholar
- Schank, R. C., and Abelson, R. 1977 Scripts, Plans, Goals and Understanding. Lawrence Erlbaum Associates, Hillsdale, N. J.Google Scholar
- Wilensky, R. 1978. Understanding Goal-Based Stories. Computer Science Technical Report 140, Yale University.Google Scholar
- Scruffy text understanding: design and implementation of 'tolerant' understanders
Recommendations
n-Gram Statistics for Natural Language Understanding and Text Processing
n-gram (n = 1 to 5) statistics and other properties of the English language were derived for applications in natural language understanding and text processing. They were computed from a well-known corpus composed of 1 million word samples. Similar ...
Scruffy text understanding: design and implementation of the NOMAD system
ANLC '83: Proceedings of the first conference on Applied natural language processingThe task of understanding unedited naval ship-to-shore messages is implemented in the presence of a large database of domain specific knowledge. The program uses internal syntactic and semantic expectations to analyze the texts and to correct errors ...
Lexical-Semantic Representation of the Lexicon for Word Sense Disambiguation and Text Understanding
ICSC '09: Proceedings of the 2009 IEEE International Conference on Semantic ComputingThis paper presents a new model of lexical-semantic representation of the lexicon which is intended for use in word sense disambiguation and semantic representation of text. It is usage-based, and part of it can be created automatically from an ...
Comments