Spaces:
Configuration error
Configuration error
| import os | |
| from transformers import AutoTokenizer | |
| from comfy.sd1_clip import SDTokenizer | |
| class MiaoBiTokenizer(SDTokenizer): | |
| def __init__(self, **kwargs): | |
| super().__init__(**kwargs) | |
| tokenizer_path = os.path.join( | |
| os.path.dirname(os.path.realpath(__file__)), | |
| f"tokenizer" | |
| ) | |
| # remote code ok, see `clip_tokenizer_roberta.py`, no ckpt vocab | |
| self.tokenizer = AutoTokenizer.from_pretrained(tokenizer_path, trust_remote_code=True) | |
| empty = self.tokenizer('')["input_ids"] | |
| if self.tokens_start: | |
| self.start_token = empty[0] | |
| self.end_token = empty[1] | |
| else: | |
| self.start_token = None | |
| self.end_token = empty[0] | |
| vocab = self.tokenizer.get_vocab() | |
| self.inv_vocab = {v: k for k, v in vocab.items()} |