Publication | Open Access
Table Fact Verification with Structure-Aware Transformer
51
Citations
8
References
2020
Year
Unknown Venue
Structured PredictionLlm Fine-tuningEngineeringMachine LearningStructured DataVerificationMultilingual PretrainingLarge Language ModelNatural Language ProcessingTable Fact VerificationData ScienceComputational LinguisticsManagementData IntegrationData ManagementStructure-aware TransformerMachine TranslationSemi-structured EvidenceVery Large DatabasePre-trained ModelsComputer ScienceDeep LearningDatabase TheorySemantic ParsingTable Structural InformationAutomated ReasoningLinguisticsData Modeling
Verifying fact on semi-structured evidence like tables requires the ability to encode structural information and perform symbolic reasoning. Pre-trained language models trained on natural language could not be directly applied to encode tables, because simply linearizing tables into sequences will lose the cell alignment information. To better utilize pre-trained transformers for table representation, we propose a Structure-Aware Transformer (SAT), which injects the table structural information into the mask of the self-attention layer. A method to combine symbolic and linguistic reasoning is also explored for this task. Our method outperforms baseline with 4.93% on TabFact, a large scale table verification dataset.
| Year | Citations | |
|---|---|---|
Page 1
Page 1