test_tokenizer.py 442 Bytes
Newer Older
1
2
from lmdeploy.turbomind.tokenizer import Postprocessor, Preprocessor, Tokenizer

q.yao's avatar
q.yao committed
3
4
5
6
7
8
9
10
11
12
13
14
15

def main():
    tokenizer = Tokenizer('huggyllama/llama-7b')
    preprocessor = Preprocessor(tokenizer)
    postprocessor = Postprocessor(tokenizer)

    prompts = ['cest la vie', '上帝已死']
    tokens = preprocessor(prompts)
    print(tokens)

    decode_prompts = postprocessor(*tokens)
    print(decode_prompts)

16

q.yao's avatar
q.yao committed
17
18
if __name__ == '__main__':
    main()