An ultimate goal of AI is to build end-to-end systems that interpret natural language, reason over the resulting logical forms, and perform actions based on that reasoning. This requires systems from separate fields be brought together, but often this exposes representational gaps between them. The logical forms from a language interpreter may mirror the surface forms of utterances too closely to be usable as-is, given a reasoner’s requirements for knowledge representations. What is needed is a system that can match logical forms to background knowledge flexibly to acquire a rich semantic model of the speaker’s goal. In this paper, we present such a “matcher” that uses semantic transformations to overcome structural differences between the two representations. We evaluate this matcher in a MUC-like template-filling task and compare its performance to that of two similar systems. Categories and Subject Descriptors I.2.4 [Artificial Intelligence]: Knowledge Representation Fo...
Peter Z. Yeh, Bruce W. Porter, Ken Barker