1 to 10 of 297 Results
Feb 6, 2019
Heinzerling, Benjamin, 2019, "BPEmb: Pre-trained Subword Embeddings in 275 Languages (LREC 2018)", https://doi.org/10.11588/data/V9CXPR, heiDATA, V1
BPEmb is a collection of pre-trained subword unit embeddings in 275 languages, based on Byte-Pair Encoding (BPE). In an evaluation using fine-grained entity typing as testbed, BPEmb performs competitively, and for some languages better than alternative subword approaches, while r... |
Gzip Archive - 257.2 MB -
MD5: 42a9f0a1361d9bb6e0746d04e5a493eb
|
Gzip Archive - 549.7 MB -
MD5: d7b3d4376eab588bab0f4180d008a6b6
|
Gzip Archive - 258.5 MB -
MD5: 6c6f944751ffb6371e81970ddff3e07d
|
Gzip Archive - 2.2 GB -
MD5: ea191a6b2e4de77b828e13c2231237cf
|
Gzip Archive - 111.1 MB -
MD5: 314a98fe6f75824994c3877857407265
|
Gzip Archive - 2.2 GB -
MD5: fb3d80ed8dd8566a05afddcbbb4ec57e
|
Gzip Archive - 2.2 GB -
MD5: cbded63bb04aab18a4bcf5191ae8242e
|
Gzip Archive - 1.1 GB -
MD5: c68ae60bab31dc0d2b9119433a14b90a
|
Gzip Archive - 2.2 GB -
MD5: 1b59f1bd4c8e0af55eba0354c9a4e56a
|