parent
ab288ff2ef
commit
ae98cc8bbf
@ -1 +1,5 @@
|
|||||||
{"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "unk_token": "<|endoftext|>"}
|
{
|
||||||
|
"bos_token": "<|endoftext|>",
|
||||||
|
"eos_token": "<|endoftext|>",
|
||||||
|
"unk_token": "<|endoftext|>"
|
||||||
|
}
|
||||||
|
@ -1 +1,10 @@
|
|||||||
{"unk_token": "<|endoftext|>", "bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "gpt2", "tokenizer_class": "GPT2Tokenizer"}
|
{
|
||||||
|
"unk_token": "<|endoftext|>",
|
||||||
|
"bos_token": "<|endoftext|>",
|
||||||
|
"eos_token": "<|endoftext|>",
|
||||||
|
"add_prefix_space": false,
|
||||||
|
"model_max_length": 1024,
|
||||||
|
"special_tokens_map_file": null,
|
||||||
|
"name_or_path": "gpt2",
|
||||||
|
"tokenizer_class": "GPT2Tokenizer"
|
||||||
|
}
|
||||||
|
File diff suppressed because one or more lines are too long
@ -1 +1,3 @@
|
|||||||
{"<|endoftext|>": 50257}
|
{
|
||||||
|
"<|endoftext|>": 50257
|
||||||
|
}
|
||||||
|
@ -1 +1,5 @@
|
|||||||
{"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "unk_token": "<|endoftext|>"}
|
{
|
||||||
|
"bos_token": "<|endoftext|>",
|
||||||
|
"eos_token": "<|endoftext|>",
|
||||||
|
"unk_token": "<|endoftext|>"
|
||||||
|
}
|
||||||
|
@ -1 +1,32 @@
|
|||||||
{"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "multilingual", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}
|
{
|
||||||
|
"unk_token": {
|
||||||
|
"content": "<|endoftext|>",
|
||||||
|
"single_word": false,
|
||||||
|
"lstrip": false,
|
||||||
|
"rstrip": false,
|
||||||
|
"normalized": true,
|
||||||
|
"__type": "AddedToken"
|
||||||
|
},
|
||||||
|
"bos_token": {
|
||||||
|
"content": "<|endoftext|>",
|
||||||
|
"single_word": false,
|
||||||
|
"lstrip": false,
|
||||||
|
"rstrip": false,
|
||||||
|
"normalized": true,
|
||||||
|
"__type": "AddedToken"
|
||||||
|
},
|
||||||
|
"eos_token": {
|
||||||
|
"content": "<|endoftext|>",
|
||||||
|
"single_word": false,
|
||||||
|
"lstrip": false,
|
||||||
|
"rstrip": false,
|
||||||
|
"normalized": true,
|
||||||
|
"__type": "AddedToken"
|
||||||
|
},
|
||||||
|
"add_prefix_space": false,
|
||||||
|
"model_max_length": 1024,
|
||||||
|
"special_tokens_map_file": null,
|
||||||
|
"name_or_path": "multilingual",
|
||||||
|
"errors": "replace",
|
||||||
|
"tokenizer_class": "GPT2Tokenizer"
|
||||||
|
}
|
||||||
|
File diff suppressed because one or more lines are too long
Loading…
Reference in new issue