Publication | Open Access
Neural Symbolic Machines: Learning Semantic Parsers on Freebase with Weak Supervision
44
Citations
37
References
2016
Year
Artificial IntelligenceNeural Symbolic MachinesStructured PredictionSyntactic ParsingEngineeringMachine LearningSemanticsLarge Language ModelStatistical PowerLanguage UnderstandingNatural Language ProcessingComputational LinguisticsGrammarLanguage StudiesMachine TranslationLarge Ai ModelSequence ModellingWeak SupervisionLearning Semantic ParsersComputer ScienceGrammar InductionDeep LearningSemantic ParsingShallow ParsingParsingTreebanksAutomated ReasoningSymbolic ReasoningLinguisticsLanguage Generation
Neural networks struggle to perform language understanding and symbolic reasoning when they must execute efficient discrete operations against a large knowledge base. This work introduces the Neural Symbolic Machine, a system combining a sequence‑to‑sequence neural programmer with a Lisp interpreter to map utterances to executable programs. The model is trained with REINFORCE to maximize task reward, augmented by iterative maximum‑likelihood steps to stabilize learning under weak supervision. On the WebQuestionsSP dataset, the Neural Symbolic Machine surpasses state‑of‑the‑art performance using only question‑answer pairs, without feature engineering or domain knowledge.
Harnessing the statistical power of neural networks to perform language understanding and symbolic reasoning is difficult, when it requires executing efficient discrete operations against a large knowledge-base. In this work, we introduce a Neural Symbolic Machine, which contains (a) a neural "programmer", i.e., a sequence-to-sequence model that maps language utterances to programs and utilizes a key-variable memory to handle compositionality (b) a symbolic "computer", i.e., a Lisp interpreter that performs program execution, and helps find good programs by pruning the search space. We apply REINFORCE to directly optimize the task reward of this structured prediction problem. To train with weak supervision and improve the stability of REINFORCE, we augment it with an iterative maximum-likelihood training process. NSM outperforms the state-of-the-art on the WebQuestionsSP dataset when trained from question-answer pairs only, without requiring any feature engineering or domain-specific knowledge.
| Year | Citations | |
|---|---|---|
Page 1
Page 1