Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
Megatron-LM
Commits
2eea6216
Commit
2eea6216
authored
Jul 18, 2022
by
rprenger
Browse files
Merging with main and fixing merge conflict
parents
ed6806ac
5f694372
Changes
63
Expand all
Show whitespace changes
Inline
Side-by-side
Showing
3 changed files
with
76 additions
and
2 deletions
+76
-2
tools/merge_datasets.py
tools/merge_datasets.py
+66
-0
tools/preprocess_data.py
tools/preprocess_data.py
+4
-2
tools/run_text_generation_server.py
tools/run_text_generation_server.py
+6
-0
No files found.
tools/merge_datasets.py
0 → 100644
View file @
2eea6216
This diff is collapsed.
Click to expand it.
tools/preprocess_data.py
View file @
2eea6216
...
@@ -122,8 +122,10 @@ def get_args():
...
@@ -122,8 +122,10 @@ def get_args():
choices
=
[
'lazy'
,
'cached'
,
'mmap'
])
choices
=
[
'lazy'
,
'cached'
,
'mmap'
])
group
=
parser
.
add_argument_group
(
title
=
'runtime'
)
group
=
parser
.
add_argument_group
(
title
=
'runtime'
)
group
.
add_argument
(
'--workers'
,
type
=
int
,
default
=
1
,
group
.
add_argument
(
'--workers'
,
type
=
int
,
required
=
True
,
help
=
'Number of worker processes to launch'
)
help
=
'Number of worker processes to launch'
)
group
.
add_argument
(
'--chunk-size'
,
type
=
int
,
required
=
True
,
help
=
'Chunk size assigned to each worker process'
)
group
.
add_argument
(
'--log-interval'
,
type
=
int
,
default
=
100
,
group
.
add_argument
(
'--log-interval'
,
type
=
int
,
default
=
100
,
help
=
'Interval between progress updates'
)
help
=
'Interval between progress updates'
)
args
=
parser
.
parse_args
()
args
=
parser
.
parse_args
()
...
@@ -154,7 +156,7 @@ def main():
...
@@ -154,7 +156,7 @@ def main():
encoder
=
Encoder
(
args
)
encoder
=
Encoder
(
args
)
tokenizer
=
build_tokenizer
(
args
)
tokenizer
=
build_tokenizer
(
args
)
pool
=
multiprocessing
.
Pool
(
args
.
workers
,
initializer
=
encoder
.
initializer
)
pool
=
multiprocessing
.
Pool
(
args
.
workers
,
initializer
=
encoder
.
initializer
)
encoded_docs
=
pool
.
imap
(
encoder
.
encode
,
fin
,
25
)
encoded_docs
=
pool
.
imap
(
encoder
.
encode
,
fin
,
args
.
chunk_size
)
#encoded_docs = map(encoder.encode, fin)
#encoded_docs = map(encoder.encode, fin)
level
=
"document"
level
=
"document"
...
...
tools/run_text_generation_server.py
View file @
2eea6216
...
@@ -28,6 +28,7 @@ from megatron.model import GPTModel
...
@@ -28,6 +28,7 @@ from megatron.model import GPTModel
from
megatron.training
import
get_model
from
megatron.training
import
get_model
from
megatron.text_generation_server
import
MegatronServer
from
megatron.text_generation_server
import
MegatronServer
from
megatron.text_generation
import
generate_and_post_process
from
megatron.text_generation
import
generate_and_post_process
from
megatron.text_generation
import
beam_search_and_post_process
import
torch
import
torch
def
model_provider
(
pre_process
=
True
,
post_process
=
True
):
def
model_provider
(
pre_process
=
True
,
post_process
=
True
):
...
@@ -82,3 +83,8 @@ if __name__ == "__main__":
...
@@ -82,3 +83,8 @@ if __name__ == "__main__":
generate_and_post_process
(
model
)
generate_and_post_process
(
model
)
except
ValueError
as
ve
:
except
ValueError
as
ve
:
pass
pass
elif
choice
[
0
].
item
()
==
1
:
try
:
beam_search_and_post_process
(
model
)
except
ValueError
as
ve
:
pass
Prev
1
2
3
4
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment