Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
text-generation-inference
Commits
521de6ca
"...text-generation-inference.git" did not exist on "f5d43414c20810dfd797d64f186e35580487883d"
Unverified
Commit
521de6ca
authored
Jun 12, 2024
by
OlivierDehaene
Committed by
GitHub
Jun 12, 2024
Browse files
fix(server): fix OPT implementation (#2061)
parent
376a0b7a
Changes
4
Hide whitespace changes
Inline
Side-by-side
Showing
4 changed files
with
9 additions
and
8 deletions
+9
-8
server/text_generation_server/models/custom_modeling/opt_modeling.py
..._generation_server/models/custom_modeling/opt_modeling.py
+1
-1
server/text_generation_server/models/gpt_neox.py
server/text_generation_server/models/gpt_neox.py
+1
-2
server/text_generation_server/models/opt.py
server/text_generation_server/models/opt.py
+2
-2
server/text_generation_server/models/rw.py
server/text_generation_server/models/rw.py
+5
-3
No files found.
server/text_generation_server/models/custom_modeling/opt_modeling.py
View file @
521de6ca
...
@@ -792,7 +792,7 @@ class OPTForCausalLM(OPTPreTrainedModel):
...
@@ -792,7 +792,7 @@ class OPTForCausalLM(OPTPreTrainedModel):
return_dict
=
return_dict
,
return_dict
=
return_dict
,
)
)
logits
,
speculative_logits
=
self
.
lm_head
(
outputs
)
logits
,
speculative_logits
=
self
.
lm_head
(
outputs
.
last_hidden_state
)
loss
=
None
loss
=
None
...
...
server/text_generation_server/models/gpt_neox.py
View file @
521de6ca
...
@@ -85,5 +85,4 @@ class GPTNeoxSharded(CausalLM):
...
@@ -85,5 +85,4 @@ class GPTNeoxSharded(CausalLM):
use_cache
=
True
,
use_cache
=
True
,
)
)
logits
=
outputs
.
logits
return
outputs
.
logits
,
speculative_logits
,
outputs
.
past_key_values
return
logits
,
speculative_logits
,
outputs
.
past_key_values
server/text_generation_server/models/opt.py
View file @
521de6ca
...
@@ -75,11 +75,11 @@ class OPTSharded(CausalLM):
...
@@ -75,11 +75,11 @@ class OPTSharded(CausalLM):
def
forward
(
def
forward
(
self
,
input_ids
,
attention_mask
,
position_ids
,
past_key_values
:
Optional
=
None
self
,
input_ids
,
attention_mask
,
position_ids
,
past_key_values
:
Optional
=
None
):
):
outputs
=
self
.
model
.
forward
(
outputs
,
speculative_logits
=
self
.
model
.
forward
(
input_ids
=
input_ids
,
input_ids
=
input_ids
,
attention_mask
=
attention_mask
,
attention_mask
=
attention_mask
,
past_key_values
=
past_key_values
,
past_key_values
=
past_key_values
,
use_cache
=
True
,
use_cache
=
True
,
)
)
return
outputs
.
logits
,
outputs
.
past_key_values
return
outputs
.
logits
,
speculative_logits
,
outputs
.
past_key_values
server/text_generation_server/models/rw.py
View file @
521de6ca
...
@@ -71,11 +71,13 @@ class RW(CausalLM):
...
@@ -71,11 +71,13 @@ class RW(CausalLM):
def
forward
(
def
forward
(
self
,
input_ids
,
attention_mask
,
position_ids
,
past_key_values
:
Optional
=
None
self
,
input_ids
,
attention_mask
,
position_ids
,
past_key_values
:
Optional
=
None
)
->
Tuple
[
torch
.
Tensor
,
List
[
Tuple
[
torch
.
Tensor
,
torch
.
Tensor
]]]
:
):
# Model Forward
# Model Forward
outputs
=
self
.
model
.
forward
(
outputs
,
speculative_logits
=
self
.
model
.
forward
(
input_ids
=
input_ids
,
input_ids
=
input_ids
,
attention_mask
=
attention_mask
,
attention_mask
=
attention_mask
,
past_key_values
=
past_key_values
,
past_key_values
=
past_key_values
,
use_cache
=
True
,
)
)
return
outputs
.
logits
,
outputs
.
past_key_values
return
outputs
.
logits
,
speculative_logits
,
outputs
.
past_key_values
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment