| import os | |
| from transformers import AutoTokenizer | |
| tokenizer = AutoTokenizer.from_pretrained("bert-base-chinese") | |
| # vocab_size = len(tokenizer.get_vocab()) | |
| # vocab_size = tokenizer.vocab_size | |
| import os | |
| from transformers import AutoTokenizer | |
| tokenizer = AutoTokenizer.from_pretrained("bert-base-chinese") | |
| # vocab_size = len(tokenizer.get_vocab()) | |
| # vocab_size = tokenizer.vocab_size | |