[HTML][HTML] A survey of word embeddings for clinical text

FK Khattak, S Jeblee, C Pou-Prom, M Abdalla… - Journal of Biomedical …, 2019 - Elsevier
Representing words as numerical vectors based on the contexts in which they appear has
become the de facto method of analyzing text with machine learning. In this paper, we …

[HTML][HTML] SECNLP: A survey of embeddings in clinical natural language processing

KS Kalyan, S Sangeetha - Journal of biomedical informatics, 2020 - Elsevier
Distributed vector representations or embeddings map variable length text to dense fixed
length vectors as well as capture prior knowledge which can transferred to downstream …

Text classification using label names only: A language model self-training approach

Y Meng, Y Zhang, J Huang, C Xiong, H Ji… - arXiv preprint arXiv …, 2020 - arxiv.org
Current text classification methods typically require a good number of human-labeled
documents as training data, which can be costly and difficult to obtain in real applications …

[HTML][HTML] Zero-shot learning for requirements classification: An exploratory study

W Alhoshan, A Ferrari, L Zhao - Information and Software Technology, 2023 - Elsevier
Context: Requirements engineering (RE) researchers have been experimenting with
machine learning (ML) and deep learning (DL) approaches for a range of RE tasks, such as …

Integrating semantic knowledge to tackle zero-shot text classification

J Zhang, P Lertvittayakumjorn, Y Guo - arXiv preprint arXiv:1903.12626, 2019 - arxiv.org
Insufficient or even unavailable training data of emerging classes is a big challenge of many
classification tasks, including text classification. Recognising text documents of classes that …

Zero-shot text classification via reinforced self-training

Z Ye, Y Geng, J Chen, J Chen, X Xu… - Proceedings of the …, 2020 - aclanthology.org
Zero-shot learning has been a tough problem since no labeled data is available for unseen
classes during training, especially for classes with low similarity. In this situation, transferring …

Pre-trained language models can be fully zero-shot learners

X Zhao, S Ouyang, Z Yu, M Wu, L Li - arXiv preprint arXiv:2212.06950, 2022 - arxiv.org
How can we extend a pre-trained model to many language understanding tasks, without
labeled or additional unlabeled data? Pre-trained language models (PLMs) have been …

TaxoClass: Hierarchical multi-label text classification using only class names

J Shen, W Qiu, Y Meng, J Shang, X Ren… - NAAC'21: Proceedings of …, 2021 - par.nsf.gov
Hierarchical multi-label text classification (HMTC) aims to tag each document with a set of
classes from a class hierarchy. Most existing HMTC methods train classifiers using massive …

Multilingual hierarchical attention networks for document classification

N Pappas, A Popescu-Belis - arXiv preprint arXiv:1707.00896, 2017 - arxiv.org
Hierarchical attention networks have recently achieved remarkable performance for
document classification in a given language. However, when multilingual document …

Evaluating unsupervised text classification: zero-shot and similarity-based approaches

T Schopf, D Braun, F Matthes - Proceedings of the 2022 6th International …, 2022 - dl.acm.org
Text classification of unseen classes is a challenging Natural Language Processing task
and is mainly attempted using two different types of approaches. Similarity-based …