Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
text-generation-inference
Commits
3fef90d5
Unverified
Commit
3fef90d5
authored
Mar 07, 2023
by
OlivierDehaene
Committed by
GitHub
Mar 07, 2023
Browse files
feat(clients): Python client (#103)
parent
0e9ed1a8
Changes
55
Show whitespace changes
Inline
Side-by-side
Showing
15 changed files
with
31 additions
and
26 deletions
+31
-26
server/text_generation_server/models/gpt_neox.py
server/text_generation_server/models/gpt_neox.py
+2
-2
server/text_generation_server/models/model.py
server/text_generation_server/models/model.py
+1
-1
server/text_generation_server/models/santacoder.py
server/text_generation_server/models/santacoder.py
+1
-1
server/text_generation_server/models/seq2seq_lm.py
server/text_generation_server/models/seq2seq_lm.py
+11
-6
server/text_generation_server/models/t5.py
server/text_generation_server/models/t5.py
+2
-2
server/text_generation_server/models/types.py
server/text_generation_server/models/types.py
+2
-2
server/text_generation_server/pb/.gitignore
server/text_generation_server/pb/.gitignore
+0
-0
server/text_generation_server/server.py
server/text_generation_server/server.py
+5
-5
server/text_generation_server/tracing.py
server/text_generation_server/tracing.py
+0
-0
server/text_generation_server/utils/__init__.py
server/text_generation_server/utils/__init__.py
+4
-4
server/text_generation_server/utils/convert.py
server/text_generation_server/utils/convert.py
+0
-0
server/text_generation_server/utils/dist.py
server/text_generation_server/utils/dist.py
+0
-0
server/text_generation_server/utils/hub.py
server/text_generation_server/utils/hub.py
+0
-0
server/text_generation_server/utils/tokens.py
server/text_generation_server/utils/tokens.py
+3
-3
server/text_generation_server/utils/watermark.py
server/text_generation_server/utils/watermark.py
+0
-0
No files found.
server/text_generation/models/gpt_neox.py
→
server/text_generation
_server
/models/gpt_neox.py
View file @
3fef90d5
...
...
@@ -16,8 +16,8 @@ from transformers.models.gpt_neox.parallel_layers import (
TensorParallelRowLinear
,
)
from
text_generation.models
import
CausalLM
from
text_generation.utils
import
(
from
text_generation
_server
.models
import
CausalLM
from
text_generation
_server
.utils
import
(
initialize_torch_distributed
,
weight_files
,
)
...
...
server/text_generation/models/model.py
→
server/text_generation
_server
/models/model.py
View file @
3fef90d5
...
...
@@ -4,7 +4,7 @@ from abc import ABC, abstractmethod
from
typing
import
List
,
Tuple
,
Optional
,
TypeVar
,
Type
from
transformers
import
PreTrainedTokenizerBase
from
text_generation.models.types
import
Batch
,
GeneratedText
from
text_generation
_server
.models.types
import
Batch
,
GeneratedText
B
=
TypeVar
(
"B"
,
bound
=
Batch
)
...
...
server/text_generation/models/santacoder.py
→
server/text_generation
_server
/models/santacoder.py
View file @
3fef90d5
...
...
@@ -4,7 +4,7 @@ import torch.distributed
from
typing
import
Optional
,
List
from
transformers
import
AutoTokenizer
,
AutoModelForCausalLM
from
text_generation.models
import
CausalLM
from
text_generation
_server
.models
import
CausalLM
FIM_PREFIX
=
"<fim-prefix>"
FIM_MIDDLE
=
"<fim-middle>"
...
...
server/text_generation/models/seq2seq_lm.py
→
server/text_generation
_server
/models/seq2seq_lm.py
View file @
3fef90d5
...
...
@@ -5,10 +5,15 @@ from opentelemetry import trace
from
transformers
import
AutoTokenizer
,
AutoModelForSeq2SeqLM
,
PreTrainedTokenizerBase
from
typing
import
Optional
,
Tuple
,
List
,
Type
from
text_generation.models
import
Model
from
text_generation.models.types
import
GeneratedText
,
Batch
,
Generation
,
PrefillTokens
from
text_generation.pb
import
generate_pb2
from
text_generation.utils
import
NextTokenChooser
,
StoppingCriteria
,
Sampling
from
text_generation_server.models
import
Model
from
text_generation_server.models.types
import
(
GeneratedText
,
Batch
,
Generation
,
PrefillTokens
,
)
from
text_generation_server.pb
import
generate_pb2
from
text_generation_server.utils
import
NextTokenChooser
,
StoppingCriteria
,
Sampling
tracer
=
trace
.
get_tracer
(
__name__
)
...
...
@@ -45,7 +50,7 @@ class Seq2SeqLMBatch(Batch):
padding_right_offset
:
int
def
to_pb
(
self
)
->
generate_pb2
.
Batch
:
"""Convert a Seq2SeqLMBatch to a text_generation.v1.Batch protobuf"""
"""Convert a Seq2SeqLMBatch to a text_generation
_server
.v1.Batch protobuf"""
return
generate_pb2
.
Batch
(
id
=
self
.
batch_id
,
requests
=
self
.
requests
,
...
...
@@ -59,7 +64,7 @@ class Seq2SeqLMBatch(Batch):
tokenizer
:
PreTrainedTokenizerBase
,
device
:
torch
.
device
,
)
->
"Seq2SeqLMBatch"
:
"""Convert a text_generation.v1.Batch protobuf to a Seq2SeqLMBatch"""
"""Convert a text_generation
_server
.v1.Batch protobuf to a Seq2SeqLMBatch"""
inputs
=
[]
next_token_choosers
=
[]
stopping_criterias
=
[]
...
...
server/text_generation/models/t5.py
→
server/text_generation
_server
/models/t5.py
View file @
3fef90d5
...
...
@@ -16,8 +16,8 @@ from transformers.models.t5.parallel_layers import (
TensorParallelRowLinear
,
)
from
text_generation.models
import
Seq2SeqLM
from
text_generation.utils
import
(
from
text_generation
_server
.models
import
Seq2SeqLM
from
text_generation
_server
.utils
import
(
initialize_torch_distributed
,
weight_files
,
)
...
...
server/text_generation/models/types.py
→
server/text_generation
_server
/models/types.py
View file @
3fef90d5
...
...
@@ -6,8 +6,8 @@ from typing import List, Optional
from
transformers
import
PreTrainedTokenizerBase
from
text_generation.pb
import
generate_pb2
from
text_generation.pb.generate_pb2
import
FinishReason
from
text_generation
_server
.pb
import
generate_pb2
from
text_generation
_server
.pb.generate_pb2
import
FinishReason
class
Batch
(
ABC
):
...
...
server/text_generation/pb/.gitignore
→
server/text_generation
_server
/pb/.gitignore
View file @
3fef90d5
File moved
server/text_generation/server.py
→
server/text_generation
_server
/server.py
View file @
3fef90d5
...
...
@@ -9,11 +9,11 @@ from grpc_reflection.v1alpha import reflection
from
pathlib
import
Path
from
typing
import
List
,
Optional
from
text_generation.cache
import
Cache
from
text_generation.interceptor
import
ExceptionInterceptor
from
text_generation.models
import
Model
,
get_model
from
text_generation.pb
import
generate_pb2_grpc
,
generate_pb2
from
text_generation.tracing
import
UDSOpenTelemetryAioServerInterceptor
from
text_generation
_server
.cache
import
Cache
from
text_generation
_server
.interceptor
import
ExceptionInterceptor
from
text_generation
_server
.models
import
Model
,
get_model
from
text_generation
_server
.pb
import
generate_pb2_grpc
,
generate_pb2
from
text_generation
_server
.tracing
import
UDSOpenTelemetryAioServerInterceptor
class
TextGenerationService
(
generate_pb2_grpc
.
TextGenerationServiceServicer
):
...
...
server/text_generation/tracing.py
→
server/text_generation
_server
/tracing.py
View file @
3fef90d5
File moved
server/text_generation/utils/__init__.py
→
server/text_generation
_server
/utils/__init__.py
View file @
3fef90d5
from
text_generation.utils.convert
import
convert_file
,
convert_files
from
text_generation.utils.dist
import
initialize_torch_distributed
from
text_generation.utils.hub
import
(
from
text_generation
_server
.utils.convert
import
convert_file
,
convert_files
from
text_generation
_server
.utils.dist
import
initialize_torch_distributed
from
text_generation
_server
.utils.hub
import
(
weight_files
,
weight_hub_files
,
download_weights
,
...
...
@@ -8,7 +8,7 @@ from text_generation.utils.hub import (
LocalEntryNotFoundError
,
RevisionNotFoundError
,
)
from
text_generation.utils.tokens
import
(
from
text_generation
_server
.utils.tokens
import
(
Greedy
,
NextTokenChooser
,
Sampling
,
...
...
server/text_generation/utils/convert.py
→
server/text_generation
_server
/utils/convert.py
View file @
3fef90d5
File moved
server/text_generation/utils/dist.py
→
server/text_generation
_server
/utils/dist.py
View file @
3fef90d5
File moved
server/text_generation/utils/hub.py
→
server/text_generation
_server
/utils/hub.py
View file @
3fef90d5
File moved
server/text_generation/utils/tokens.py
→
server/text_generation
_server
/utils/tokens.py
View file @
3fef90d5
...
...
@@ -11,9 +11,9 @@ from transformers import (
)
from
typing
import
List
,
Tuple
,
Optional
from
text_generation.pb
import
generate_pb2
from
text_generation.pb.generate_pb2
import
FinishReason
from
text_generation.utils.watermark
import
WatermarkLogitsProcessor
from
text_generation
_server
.pb
import
generate_pb2
from
text_generation
_server
.pb.generate_pb2
import
FinishReason
from
text_generation
_server
.utils.watermark
import
WatermarkLogitsProcessor
class
Sampling
:
...
...
server/text_generation/utils/watermark.py
→
server/text_generation
_server
/utils/watermark.py
View file @
3fef90d5
File moved
Prev
1
2
3
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment