HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

ECONET: Effective Continual Pretraining of Language Models for Event Temporal Reasoning

Rujun Han Xiang Ren Nanyun Peng

ECONET: Effective Continual Pretraining of Language Models for Event Temporal Reasoning

Abstract

While pre-trained language models (PTLMs) have achieved noticeable success on many NLP tasks, they still struggle for tasks that require event temporal reasoning, which is essential for event-centric applications. We present a continual pre-training approach that equips PTLMs with targeted knowledge about event temporal relations. We design self-supervised learning objectives to recover masked-out event and temporal indicators and to discriminate sentences from their corrupted counterparts (where event or temporal indicators got replaced). By further pre-training a PTLM with these objectives jointly, we reinforce its attention to event and temporal information, yielding enhanced capability on event temporal reasoning. This effective continual pre-training framework for event temporal reasoning (ECONET) improves the PTLMs' fine-tuning performances across five relation extraction and question answering tasks and achieves new or on-par state-of-the-art performances in most of our downstream tasks.

Code Repositories

pluslabnlp/econet
Official
pytorch
ZHEvent/ZHEvent.github.io
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
question-answering-on-torqueECONET
C: 37.0
EM: 52.0
F1: 76.3

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp