Faster Parallel Training of Word Embeddings
Word embeddings have gained increasing popularity in the recent years due to the Word2vec library and its extension fastText that uses subword information. In this paper, we aim at improving the execution speed of fastText training on homogeneous multi- and manycore CPUs while maintaining accuracy. We present a novel open-source implementation that flexibly incorporates various algorithmic Variants including negative sample sharing, batched updates. and a byte-pair encoding-based alternative for subword units. We build these novel variants over a fastText implementation that we carefully optimized for the architecture, memory hierarchy, and parallelism of current manycore CPUs. Our experiments on three languages demonstrate 3-20x speed-up in training time at competitive semantic and syntactic accuracy.
WOS:000782316500004
2021-01-01
978-1-6654-1016-8
Los Alamitos
International Conference on High Performance Computing
31
41
REVIEWED
Event name | Event place | Event date |
ELECTR NETWORK | Dec 17-18, 2021 | |