Pre-trained word embeddings encode general word semantics and lexical regularities of natural language, and have proven useful across many NLP tasks, including word sense disambiguation, machine translation, and sentiment analysis, to name a few. In supervised tasks such as multiclass text classification (the focus of this article) it seems appealing to enhance word representations with ad-hoc embeddings that encode task-specific information. We propose (supervised) word-class embeddings (WCEs), and show that, when concatenated to (unsupervised) pre-trained word embeddings, they substantially facilitate the training of deep-learning models in multiclass classification by topic. We show empirical evidence that WCEs yield a consistent improvement in multiclass classification accuracy, using six popular neural architectures and six widely used and publicly available datasets for multiclass text classification. One further advantage of this method is that it is conceptually simple and straightforward to implement. Our code that implements WCEs is publicly available at https://github.com/AlexMoreo/word-class-embeddings.
Word-class embeddings for multiclass text classification
Moreo A;Esuli A;Sebastiani F
2021
Abstract
Pre-trained word embeddings encode general word semantics and lexical regularities of natural language, and have proven useful across many NLP tasks, including word sense disambiguation, machine translation, and sentiment analysis, to name a few. In supervised tasks such as multiclass text classification (the focus of this article) it seems appealing to enhance word representations with ad-hoc embeddings that encode task-specific information. We propose (supervised) word-class embeddings (WCEs), and show that, when concatenated to (unsupervised) pre-trained word embeddings, they substantially facilitate the training of deep-learning models in multiclass classification by topic. We show empirical evidence that WCEs yield a consistent improvement in multiclass classification accuracy, using six popular neural architectures and six widely used and publicly available datasets for multiclass text classification. One further advantage of this method is that it is conceptually simple and straightforward to implement. Our code that implements WCEs is publicly available at https://github.com/AlexMoreo/word-class-embeddings.| File | Dimensione | Formato | |
|---|---|---|---|
|
prod_454276-doc_175039.pdf
accesso aperto
Descrizione: WORD-CLASS EMBEDDINGS FOR MULTICLASS TEXT CLASSIFICATION
Tipologia:
Documento in Pre-print
Licenza:
Nessuna licenza dichiarata (non attribuibile a prodotti successivi al 2023)
Dimensione
7.34 MB
Formato
Adobe PDF
|
7.34 MB | Adobe PDF | Visualizza/Apri |
|
prod_454276-doc_175070.pdf
solo utenti autorizzati
Descrizione: Word-class embeddings for multiclass text classification
Tipologia:
Versione Editoriale (PDF)
Licenza:
NON PUBBLICO - Accesso privato/ristretto
Dimensione
18.3 MB
Formato
Adobe PDF
|
18.3 MB | Adobe PDF | Visualizza/Apri Richiedi una copia |
|
FinalVersion.pdf
accesso aperto
Descrizione: This is the Author Accepted Manuscript (postprint) of the following paper: Moreo A., Esuli A., Sebastiani F. “Word-class embeddings for multiclass text classification”, published in “Data Mining Knowledge Discovery” Vol. 35, pp. 911-963, 2021. DOI: 10.1007/s10618-020-00735-3.
Tipologia:
Documento in Post-print
Licenza:
Nessuna licenza dichiarata (non attribuibile a prodotti successivi al 2023)
Dimensione
9.57 MB
Formato
Adobe PDF
|
9.57 MB | Adobe PDF | Visualizza/Apri |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.


