Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
norm
vllm
Commits
24cde76a
Unverified
Commit
24cde76a
authored
Dec 10, 2023
by
Woosuk Kwon
Committed by
GitHub
Dec 10, 2023
Browse files
[Minor] Add comment on skipping rope caches (#2004)
parent
1aa13615
Changes
1
Show whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
4 additions
and
3 deletions
+4
-3
vllm/model_executor/models/llama.py
vllm/model_executor/models/llama.py
+4
-3
No files found.
vllm/model_executor/models/llama.py
View file @
24cde76a
...
...
@@ -322,9 +322,10 @@ class LlamaForCausalLM(nn.Module):
model_name_or_path
,
cache_dir
,
load_format
,
revision
):
if
"rotary_emb.inv_freq"
in
name
:
continue
if
"rotary_emb.cos_cached"
in
name
:
continue
if
"rotary_emb.sin_cached"
in
name
:
if
(
"rotary_emb.cos_cached"
in
name
or
"rotary_emb.sin_cached"
in
name
):
# Models trained using ColossalAI may include these tensors in
# the checkpoint. Skip them.
continue
for
(
param_name
,
weight_name
,
shard_id
)
in
stacked_params_mapping
:
if
weight_name
not
in
name
:
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment