Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
zhaoyu6
sglang
Commits
21ba3a88
"eigen-master/doc/snippets/Cwise_abs.cpp" did not exist on "e7df86554156b36846008d8ddbcc4d8521a16554"
Unverified
Commit
21ba3a88
authored
Jul 18, 2024
by
Ying Sheng
Committed by
GitHub
Jul 18, 2024
Browse files
Remove useless variables in infer_batch.py (#651)
parent
9c5cac24
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
3 additions
and
4 deletions
+3
-4
python/sglang/srt/managers/controller/infer_batch.py
python/sglang/srt/managers/controller/infer_batch.py
+3
-4
No files found.
python/sglang/srt/managers/controller/infer_batch.py
View file @
21ba3a88
...
...
@@ -270,6 +270,7 @@ class Batch:
prefix_lens
:
torch
.
Tensor
=
None
position_ids_offsets
:
torch
.
Tensor
=
None
out_cache_loc
:
torch
.
Tensor
=
None
extend_num_tokens
:
int
=
None
# For processing logprobs
return_logprob
:
bool
=
False
...
...
@@ -280,10 +281,6 @@ class Batch:
image_sizes
:
List
[
List
[
int
]]
=
None
image_offsets
:
List
[
int
]
=
None
# Other arguments for control
output_ids
:
torch
.
Tensor
=
None
extend_num_tokens
:
int
=
None
# Batched sampling params
temperatures
:
torch
.
Tensor
=
None
top_ps
:
torch
.
Tensor
=
None
...
...
@@ -820,6 +817,7 @@ def init_flashinfer_args(
prefix_lens
,
flashinfer_decode_wrapper
,
):
"""Init auxiliary variables for FlashInfer attention backend."""
num_qo_heads
=
model_runner
.
model_config
.
num_attention_heads
//
model_runner
.
tp_size
num_kv_heads
=
model_runner
.
model_config
.
get_num_kv_heads
(
model_runner
.
tp_size
)
head_dim
=
model_runner
.
model_config
.
head_dim
...
...
@@ -885,6 +883,7 @@ def init_flashinfer_args(
def
init_triton_args
(
forward_mode
,
seq_lens
,
prefix_lens
):
"""Init auxiliary variables for triton attention backend."""
batch_size
=
len
(
seq_lens
)
max_seq_len
=
int
(
torch
.
max
(
seq_lens
))
start_loc
=
torch
.
zeros
((
batch_size
,),
dtype
=
torch
.
int32
,
device
=
"cuda"
)
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment