Can word vectors help corpus linguists?

Abstract : Two recent methods based on distributional semantic models (DSMs) have proved very successful in learning high-quality vector representations of words from large corpora: word2vec (Mikolov, Chen, et al. 2013; Mikolov, Yih, et al. 2013) and GloVe (Pennington et al. 2014). Once trained on a very large corpus, these algorithms produce distributed representations for words in the form of vectors. DSMs based on deep learning and neural networks have proved efficient in representing the meaning of individual words. In this paper, I assess to what extent state-of-the-art word-vector semantics can help corpus linguists annotate large datasets for semantic classes. Although word vectors suggest decisive opportunities for resolving semantic annotation issues, it has yet to improve in terms of its representation of polysemy, homonymy, and multiword expressions.
Document type :
Preprints, Working Papers, ...
Complete list of metadatas

https://halshs.archives-ouvertes.fr/halshs-01657591
Contributor : Guillaume Desagulier <>
Submitted on : Wednesday, October 3, 2018 - 12:56:23 PM
Last modification on : Thursday, April 4, 2019 - 10:18:05 AM

File

wordvecs.pdf
Files produced by the author(s)

Identifiers

  • HAL Id : halshs-01657591, version 2

Collections

Citation

Guillaume Desagulier. Can word vectors help corpus linguists?. 2018. ⟨halshs-01657591v2⟩

Share

Metrics

Record views

86

Files downloads

624