Lessons from Efforts to Automatically Translate English to Knowledge Representation Languages

Research output: Chapter in Book/Report/Conference proceedingConference contribution

1 Scopus citations

Abstract

Our long term goal is to develop systems that can “understand” natural language text. By “understand” we mean that the system can take natural language text as input and answer questions with respect to that text. A key component in building such systems is to be able to translate natural language text into appropriate knowledge representation (KR) languages. Our approach to achieve that is inspired by Montagues path breaking thesis (1970) of viewing English as a formal language and by the research in natural language semantics. Our approach is based on PCCG (Probabilistic Combinatorial Categorial Grammars), λ-calculus and statistical learning of parameters. In an initial work, we start with an initial vocabulary consisting of λ-calculus representations of a small set of words and a training corpus of sentences and their representation in a KR language. We develop a learning based system that learns the λ-calculus representation of words from this corpus and generalizes it to words of the same category. The key and novel aspect in this learning is the development of Inverse Lambda algorithms which when given λ-expressions β and γ can come up with an α such that application of α to β (or β to α) will give us γ. We augment this with learning of weights associated with multiple meanings of words. Our current system produces improved results on standard corpora on natural language interfaces for robot command and control and database queries. In a follow-up work we are able to use patterns to make guesses regarding the initial vocabulary. This together with learning of parameters allow us to develop a fully automated (without any initial vocabulary) way to translate English to designated KR languages. In an on-going work we use Answer Set Programming as the target KR language and focus on (a) solving combinatorial puzzles that are described in English and (b) answering questions with respect to a chapter in a ninth grade biology book. The systems that we are building are good examples of integration of results from multiple sub-fields of AI and computer science, viz.: machine learning, knowledge representation, natural language processing, λ-calculus (functional programming) and ontologies. In this presentation we will describe our approach and our system and elaborate on some of the lessons that we have learned from this effort.

Original languageEnglish (US)
Title of host publicationLogic Programming and Nonmonotonic Reasoning - 11th International Conference, LPNMR 2011, Proceedings
EditorsJames P. Delgrande, Wolfgang Faber
PublisherSpringer Verlag
Pages12
Number of pages1
ISBN (Print)9783642208942
DOIs
StatePublished - 2011
Event11th International Conference on Logic Programming and Nonmonotonic Reasoning, LPNMR 2011 - Vancouver, BC, Canada
Duration: May 16 2011May 19 2011

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume6645 LNAI
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Other

Other11th International Conference on Logic Programming and Nonmonotonic Reasoning, LPNMR 2011
Country/TerritoryCanada
CityVancouver, BC
Period5/16/115/19/11

ASJC Scopus subject areas

  • Theoretical Computer Science
  • General Computer Science

Fingerprint

Dive into the research topics of 'Lessons from Efforts to Automatically Translate English to Knowledge Representation Languages'. Together they form a unique fingerprint.

Cite this