Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
zhaoyu6
sglang
Commits
3fdab919
".github/scripts/vscode:/vscode.git/clone" did not exist on "9d11563dee99c589417b740452b11ca22a71624b"
Unverified
Commit
3fdab919
authored
Jul 27, 2024
by
Ke Bao
Committed by
GitHub
Jul 27, 2024
Browse files
Fix TransformerTokenizer init for chatglm2 & 3 (#761)
parent
ba29504b
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
21 additions
and
1 deletion
+21
-1
python/sglang/srt/constrained/fsm_cache.py
python/sglang/srt/constrained/fsm_cache.py
+21
-1
No files found.
python/sglang/srt/constrained/fsm_cache.py
View file @
3fdab919
...
@@ -21,7 +21,27 @@ class FSMCache(BaseCache):
...
@@ -21,7 +21,27 @@ class FSMCache(BaseCache):
tokenizer
=
AutoTokenizer
.
from_pretrained
(
tokenizer
=
AutoTokenizer
.
from_pretrained
(
tokenizer_path
,
**
tokenizer_args_dict
tokenizer_path
,
**
tokenizer_args_dict
)
)
self
.
outlines_tokenizer
=
TransformerTokenizer
(
tokenizer
)
try
:
self
.
outlines_tokenizer
=
TransformerTokenizer
(
tokenizer
)
except
AttributeError
:
# FIXME: tmp fix for chatglm2 & chatglm3 (pad_token_id=0)
origin_pad_token_id
=
tokenizer
.
pad_token_id
def
fset
(
self
,
value
):
self
.
_value
=
value
type
(
tokenizer
).
pad_token_id
=
property
(
fget
=
type
(
tokenizer
).
pad_token_id
.
fget
,
fset
=
fset
)
self
.
outlines_tokenizer
=
TransformerTokenizer
(
tokenizer
)
self
.
outlines_tokenizer
.
tokenizer
.
pad_token_id
=
origin_pad_token_id
self
.
outlines_tokenizer
.
pad_token_id
=
origin_pad_token_id
self
.
outlines_tokenizer
.
pad_token
=
(
self
.
outlines_tokenizer
.
tokenizer
.
pad_token
)
self
.
outlines_tokenizer
.
vocabulary
=
(
self
.
outlines_tokenizer
.
tokenizer
.
get_vocab
()
)
else
:
else
:
self
.
outlines_tokenizer
=
TransformerTokenizer
(
self
.
outlines_tokenizer
=
TransformerTokenizer
(
tokenizer_path
,
**
tokenizer_args_dict
tokenizer_path
,
**
tokenizer_args_dict
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment