Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing

P Liu, W Yuan, J Fu, Z Jiang, H Hayashi… - ACM Computing …, 2023 - dl.acm.org
This article surveys and organizes research works in a new paradigm in natural language
processing, which we dub “prompt-based learning.” Unlike traditional supervised learning …

A comprehensive survey on automatic knowledge graph construction

L Zhong, J Wu, Q Li, H Peng, X Wu - ACM Computing Surveys, 2023 - dl.acm.org
Automatic knowledge graph construction aims at manufacturing structured human
knowledge. To this end, much effort has historically been spent extracting informative fact …

Large language models are few-shot clinical information extractors

M Agrawal, S Hegselmann, H Lang, Y Kim… - arXiv preprint arXiv …, 2022 - arxiv.org
A long-running goal of the clinical NLP community is the extraction of important variables
trapped in clinical notes. However, roadblocks have included dataset shift from the general …

Unified structure generation for universal information extraction

Y Lu, Q Liu, D Dai, X Xiao, H Lin, X Han, L Sun… - arXiv preprint arXiv …, 2022 - arxiv.org
Information extraction suffers from its varying targets, heterogeneous structures, and
demand-specific schemas. In this paper, we propose a unified text-to-structure generation …

Prompting gpt-3 to be reliable

C Si, Z Gan, Z Yang, S Wang, J Wang… - arXiv preprint arXiv …, 2022 - arxiv.org
Large language models (LLMs) show impressive abilities via few-shot prompting.
Commercialized APIs such as OpenAI GPT-3 further increase their use in real-world …

Knowledge enhanced contextual word representations

ME Peters, M Neumann, RL Logan IV… - arXiv preprint arXiv …, 2019 - arxiv.org
Contextual word representations, typically trained on unstructured, unlabeled text, do not
contain any explicit grounding to real world entities and are often unable to remember facts …

Superglue: A stickier benchmark for general-purpose language understanding systems

A Wang, Y Pruksachatkun, N Nangia… - Advances in neural …, 2019 - proceedings.neurips.cc
In the last year, new models and methods for pretraining and transfer learning have driven
striking performance improvements across a range of language understanding tasks. The …

Spanbert: Improving pre-training by representing and predicting spans

M Joshi, D Chen, Y Liu, DS Weld… - Transactions of the …, 2020 - direct.mit.edu
We present SpanBERT, a pre-training method that is designed to better represent and
predict spans of text. Our approach extends BERT by (1) masking contiguous random spans …

Span-based joint entity and relation extraction with transformer pre-training

M Eberts, A Ulges - ECAI 2020, 2020 - ebooks.iospress.nl
We introduce SpERT, an attention model for span-based joint entity and relation extraction.
Our key contribution is a light-weight reasoning on BERT embeddings, which features entity …

[图书][B] Deep learning on graphs

Y Ma, J Tang - 2021 - books.google.com
Deep learning on graphs has become one of the hottest topics in machine learning. The
book consists of four parts to best accommodate our readers with diverse backgrounds and …