From d78d4fa6ffa1a7978dfe9767fb1c3ec526e58a10 Mon Sep 17 00:00:00 2001 From: yangyaming Date: Thu, 12 Oct 2017 17:06:08 +0800 Subject: [PATCH] Add url for large Mandarin LM. --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index efd45efb..8dd13f92 100644 --- a/README.md +++ b/README.md @@ -463,7 +463,7 @@ Language Model | Training Data | Token-based | Size | Description :-------------:| :------------:| :-----: | -----: | :-----------------: [English LM](http://paddlepaddle.bj.bcebos.com/model_zoo/speech/common_crawl_00.prune01111.trie.klm) | [en.00.deduped.xz](http://web-language-models.s3-website-us-east-1.amazonaws.com/ngrams/en/deduped/en.00.deduped.xz) | Word-based | 8.3 GB | pruned with 0 1 1 1 1
about 1.85 billion n-grams
'trie' binary with '-a 22 -q 8 -b 8' [Mandarin LM Small](http://cloud.dlnel.org/filepub/?uuid=d21861e4-4ed6-45bb-ad8e-ae417a43195e) | Small internel data | Character-based | 2.8 GB | pruned with 0 1 2 4 4
about 0.13 billion n-grams
'probing' binary with default settings -Mandarin LM Large | Large internel data | Character-based | 70.4 GB | no pruning
about 3.7 billion n-grams
'probing' binary with default settings +[Mandarin LM Large](http://cloud.dlnel.org/filepub/?uuid=245d02bb-cd01-4ebe-b079-b97be864ec37) | Large internel data | Character-based | 70.4 GB | no pruning
about 3.7 billion n-grams
'probing' binary with default settings ## Experiments and Benchmarks