Repository logo
  • Log In
    New user? Click here to register.Have you forgotten your password?
University College Dublin
    Colleges & Schools
    Statistics
    All of DSpace
  • Log In
    New user? Click here to register.Have you forgotten your password?
  1. Home
  2. College of Science
  3. School of Computer Science
  4. Computer Science Research Collection
  5. A Comparative Study on Word Embeddings in Deep Learning for Text Classification
 
  • Details
Options

A Comparative Study on Word Embeddings in Deep Learning for Text Classification

Author(s)
Wang, Congcong  
Nulty, Paul  
Lillis, David  
Uri
http://hdl.handle.net/10197/25812
Date Issued
2020-12-18
Date Available
2024-05-02T15:38:23Z
Abstract
Word embeddings act as an important component of deep models for providing input features in downstream language tasks, such as sequence labelling and text classification. In the last decade, a substantial number of word embedding methods have been proposed for this purpose, mainly falling into the categories of classic and context-based word embeddings. In this paper, we conduct controlled experiments to systematically examine both classic and contextualised word embeddings for the purposes of text classification. To encode a sequence from word representations, we apply two encoders, namely CNN and BiLSTM, in the downstream network architecture. To study the impact of word embeddings on different datasets, we select four benchmarking classification datasets with varying average sample length, comprising both single-label and multi-label classification tasks. The evaluation results with confidence intervals indicate that CNN as the downstream encoder outperforms BiLSTM in most situations, especially for document context-insensitive datasets. This study recommends choosing CNN over BiLSTM for document classification datasets where the context in sequence is not as indicative of class membership as sentence datasets. For word embeddings, concatenation of multiple classic embeddings or increasing their size does not lead to a statistically significant difference in performance despite a slight improvement in some cases. For context-based embeddings, we studied both ELMo and BERT. The results show that BERT overall outperforms ELMo, especially for long document datasets. Compared with classic embeddings, both achieve an improved performance for short datasets while the improvement is not observed in longer datasets.
Type of Material
Conference Publication
Publisher
ACM
Copyright (Published Version)
2020 the Authors
Subjects

Word embeddings

Neural networks

Text classification

DOI
10.1145/3443279.3443304
Language
English
Status of Item
Peer reviewed
Journal
NLPIR 2020: Proceedings of the 4th International Conference on Natural Language Processing and Information Retrieval
ISBN
9781450377607
This item is made available under a Creative Commons License
https://creativecommons.org/licenses/by-nc-nd/3.0/ie/
File(s)
Loading...
Thumbnail Image
Name

Wang2020a(1).pdf

Size

1.28 MB

Format

Adobe PDF

Checksum (MD5)

cd3fdc42ca4fc3ae673cf24f2b3f0c49

Owning collection
Computer Science Research Collection

Item descriptive metadata is released under a CC-0 (public domain) license: https://creativecommons.org/public-domain/cc0/.
All other content is subject to copyright.

For all queries please contact research.repository@ucd.ie.

Built with DSpace-CRIS software - Extension maintained and optimized by 4Science

  • Cookie settings
  • Privacy policy
  • End User Agreement