--- license: apache-2.0 language: - en --- Original model from https://huggingface.co/openlm-research/open_llama_3b_600bt_preview. This repo includes: 1) Ported `LlamaTokenizer` to `LlamaTokenizerFast` via a few lines of code. Loading via `AutoTokenizer` takes 3 to 4 minutes. Now, a few seconds! ``` from transformers import LlamaTokenizerFast from tokenizers import AddedToken tokenizer = LlamaTokenizerFast.from_pretrained( "openlm-research/open_llama_3b_600bt_preview", add_bos_token = True, add_eos_token = True, bos_token = AddedToken("", single_word = True), eos_token = AddedToken("", single_word = True), unk_token = AddedToken("", single_word = True), pad_token = AddedToken("", single_word = True) ) tokenizer.push_to_hub("open_llama_3b_600bt_preview") ``` 2) `AutoTokenizer` does not recognize the BOS, EOS and UNK tokens. All tokenizations weirdly prepend 0 and append 0 to the end, when actually, you're supposed to prepend 1 and append 2. 3) Manually added BOS ``, EOS ``, UNK `` tokens, with PAD (padding) being also the `` token.