test_tokenizer.py 440 Bytes
Newer Older
q.yao's avatar
q.yao committed
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
from lmdeploy.turbomind.tokenizer import Tokenizer, Preprocessor, Postprocessor

def main():
    tokenizer = Tokenizer('huggyllama/llama-7b')
    preprocessor = Preprocessor(tokenizer)
    postprocessor = Postprocessor(tokenizer)

    prompts = ['cest la vie', '上帝已死']
    tokens = preprocessor(prompts)
    print(tokens)

    decode_prompts = postprocessor(*tokens)
    print(decode_prompts)

if __name__ == '__main__':
    main()