Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
Lmdeploy
Commits
e2d98ddc
Commit
e2d98ddc
authored
Mar 23, 2024
by
zhouxiang
Browse files
解决bf16编译问题
parent
d7117b95
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
2 additions
and
0 deletions
+2
-0
src/turbomind/triton_backend/llama/LlamaTritonModel.cc
src/turbomind/triton_backend/llama/LlamaTritonModel.cc
+2
-0
No files found.
src/turbomind/triton_backend/llama/LlamaTritonModel.cc
View file @
e2d98ddc
...
...
@@ -47,6 +47,7 @@ std::shared_ptr<AbstractTransformerModel> AbstractTransformerModel::createLlamaM
reader
.
GetInteger
(
"ft_instance_hyperparameter"
,
"enable_custom_all_reduce"
,
0
),
model_dir
);
}
#ifdef ENABLE_BF16
else
if
(
data_type
==
"bf16"
)
{
#ifdef ENABLE_BF16
return
std
::
make_shared
<
LlamaTritonModel
<
__nv_bfloat16
>>
(
...
...
@@ -59,6 +60,7 @@ std::shared_ptr<AbstractTransformerModel> AbstractTransformerModel::createLlamaM
ft
::
FT_CHECK
(
false
);
#endif
}
#endif
else
{
return
std
::
make_shared
<
LlamaTritonModel
<
float
>>
(
reader
.
GetInteger
(
"ft_instance_hyperparameter"
,
"tensor_para_size"
),
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment