Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
ColossalAI
Commits
abf4c27f
Unverified
Commit
abf4c27f
authored
Nov 12, 2022
by
Frank Lee
Committed by
GitHub
Nov 12, 2022
Browse files
[tutorial] removed huggingface model warning (#1925)
parent
d43a671a
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
13 additions
and
13 deletions
+13
-13
examples/tutorial/opt/opt/run_clm.py
examples/tutorial/opt/opt/run_clm.py
+13
-13
No files found.
examples/tutorial/opt/opt/run_clm.py
View file @
abf4c27f
...
@@ -30,24 +30,13 @@ from itertools import chain
...
@@ -30,24 +30,13 @@ from itertools import chain
import
datasets
import
datasets
import
torch
import
torch
import
torch.distributed
as
dist
import
torch.distributed
as
dist
import
transformers
from
accelerate.utils
import
set_seed
from
accelerate.utils
import
set_seed
from
context
import
barrier_context
from
context
import
barrier_context
from
datasets
import
load_dataset
from
datasets
import
load_dataset
from
packaging
import
version
from
packaging
import
version
from
torch.utils.data
import
DataLoader
from
torch.utils.data
import
DataLoader
from
tqdm.auto
import
tqdm
from
tqdm.auto
import
tqdm
import
colossalai
import
transformers
from
colossalai.context
import
ParallelMode
from
colossalai.core
import
global_context
as
gpc
from
colossalai.logging
import
disable_existing_loggers
,
get_dist_logger
from
colossalai.nn.optimizer
import
HybridAdam
from
colossalai.nn.parallel
import
ZeroDDP
from
colossalai.tensor
import
ProcessGroup
from
colossalai.utils
import
get_current_device
,
get_dataloader
from
colossalai.utils.model.colo_init_context
import
ColoInitContext
from
colossalai.zero
import
ZeroOptimizer
from
transformers
import
(
from
transformers
import
(
CONFIG_MAPPING
,
CONFIG_MAPPING
,
MODEL_MAPPING
,
MODEL_MAPPING
,
...
@@ -61,6 +50,17 @@ from transformers import (
...
@@ -61,6 +50,17 @@ from transformers import (
)
)
from
transformers.utils.versions
import
require_version
from
transformers.utils.versions
import
require_version
import
colossalai
from
colossalai.context
import
ParallelMode
from
colossalai.core
import
global_context
as
gpc
from
colossalai.logging
import
disable_existing_loggers
,
get_dist_logger
from
colossalai.nn.optimizer
import
HybridAdam
from
colossalai.nn.parallel
import
ZeroDDP
from
colossalai.tensor
import
ProcessGroup
from
colossalai.utils
import
get_current_device
,
get_dataloader
from
colossalai.utils.model.colo_init_context
import
ColoInitContext
from
colossalai.zero
import
ZeroOptimizer
require_version
(
"datasets>=1.8.0"
,
"To fix: pip install -r examples/pytorch/language-modeling/requirements.txt"
)
require_version
(
"datasets>=1.8.0"
,
"To fix: pip install -r examples/pytorch/language-modeling/requirements.txt"
)
MODEL_CONFIG_CLASSES
=
list
(
MODEL_MAPPING
.
keys
())
MODEL_CONFIG_CLASSES
=
list
(
MODEL_MAPPING
.
keys
())
...
@@ -544,7 +544,7 @@ def main():
...
@@ -544,7 +544,7 @@ def main():
model
.
train
()
model
.
train
()
for
step
,
batch
in
enumerate
(
train_dataloader
):
for
step
,
batch
in
enumerate
(
train_dataloader
):
batch
=
{
k
:
v
.
cuda
()
for
k
,
v
in
batch
.
items
()}
batch
=
{
k
:
v
.
cuda
()
for
k
,
v
in
batch
.
items
()}
outputs
=
model
(
**
batch
)
outputs
=
model
(
use_cache
=
False
,
**
batch
)
loss
=
outputs
[
'loss'
]
loss
=
outputs
[
'loss'
]
optimizer
.
backward
(
loss
)
optimizer
.
backward
(
loss
)
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment