Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
wangsen
paddle_dbnet
Commits
a2cdabd0
Commit
a2cdabd0
authored
Apr 29, 2022
by
LDOUBLEV
Browse files
fix det cml + pact + distribute training bug
parent
6dff6d97
Changes
2
Show whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
30 additions
and
15 deletions
+30
-15
deploy/slim/quantization/quant.py
deploy/slim/quantization/quant.py
+6
-6
ppocr/optimizer/optimizer.py
ppocr/optimizer/optimizer.py
+24
-9
No files found.
deploy/slim/quantization/quant.py
View file @
a2cdabd0
...
@@ -161,12 +161,6 @@ def main(config, device, logger, vdl_writer):
...
@@ -161,12 +161,6 @@ def main(config, device, logger, vdl_writer):
if
config
[
"Global"
][
"pretrained_model"
]
is
not
None
:
if
config
[
"Global"
][
"pretrained_model"
]
is
not
None
:
pre_best_model_dict
=
load_model
(
config
,
model
)
pre_best_model_dict
=
load_model
(
config
,
model
)
quanter
=
QAT
(
config
=
quant_config
,
act_preprocess
=
PACT
)
quanter
.
quantize
(
model
)
if
config
[
'Global'
][
'distributed'
]:
model
=
paddle
.
DataParallel
(
model
)
# build loss
# build loss
loss_class
=
build_loss
(
config
[
'Loss'
])
loss_class
=
build_loss
(
config
[
'Loss'
])
...
@@ -181,6 +175,12 @@ def main(config, device, logger, vdl_writer):
...
@@ -181,6 +175,12 @@ def main(config, device, logger, vdl_writer):
if
config
[
"Global"
][
"checkpoints"
]
is
not
None
:
if
config
[
"Global"
][
"checkpoints"
]
is
not
None
:
pre_best_model_dict
=
load_model
(
config
,
model
,
optimizer
)
pre_best_model_dict
=
load_model
(
config
,
model
,
optimizer
)
quanter
=
QAT
(
config
=
quant_config
,
act_preprocess
=
PACT
)
quanter
.
quantize
(
model
)
if
config
[
'Global'
][
'distributed'
]:
model
=
paddle
.
DataParallel
(
model
)
# build metric
# build metric
eval_class
=
build_metric
(
config
[
'Metric'
])
eval_class
=
build_metric
(
config
[
'Metric'
])
...
...
ppocr/optimizer/optimizer.py
View file @
a2cdabd0
...
@@ -43,12 +43,15 @@ class Momentum(object):
...
@@ -43,12 +43,15 @@ class Momentum(object):
self
.
grad_clip
=
grad_clip
self
.
grad_clip
=
grad_clip
def
__call__
(
self
,
model
):
def
__call__
(
self
,
model
):
train_params
=
[
param
for
param
in
model
.
parameters
()
if
param
.
trainable
is
True
]
opt
=
optim
.
Momentum
(
opt
=
optim
.
Momentum
(
learning_rate
=
self
.
learning_rate
,
learning_rate
=
self
.
learning_rate
,
momentum
=
self
.
momentum
,
momentum
=
self
.
momentum
,
weight_decay
=
self
.
weight_decay
,
weight_decay
=
self
.
weight_decay
,
grad_clip
=
self
.
grad_clip
,
grad_clip
=
self
.
grad_clip
,
parameters
=
model
.
parameters
()
)
parameters
=
train_params
)
return
opt
return
opt
...
@@ -76,6 +79,9 @@ class Adam(object):
...
@@ -76,6 +79,9 @@ class Adam(object):
self
.
lazy_mode
=
lazy_mode
self
.
lazy_mode
=
lazy_mode
def
__call__
(
self
,
model
):
def
__call__
(
self
,
model
):
train_params
=
[
param
for
param
in
model
.
parameters
()
if
param
.
trainable
is
True
]
opt
=
optim
.
Adam
(
opt
=
optim
.
Adam
(
learning_rate
=
self
.
learning_rate
,
learning_rate
=
self
.
learning_rate
,
beta1
=
self
.
beta1
,
beta1
=
self
.
beta1
,
...
@@ -85,7 +91,7 @@ class Adam(object):
...
@@ -85,7 +91,7 @@ class Adam(object):
grad_clip
=
self
.
grad_clip
,
grad_clip
=
self
.
grad_clip
,
name
=
self
.
name
,
name
=
self
.
name
,
lazy_mode
=
self
.
lazy_mode
,
lazy_mode
=
self
.
lazy_mode
,
parameters
=
model
.
parameters
()
)
parameters
=
train_params
)
return
opt
return
opt
...
@@ -118,6 +124,9 @@ class RMSProp(object):
...
@@ -118,6 +124,9 @@ class RMSProp(object):
self
.
grad_clip
=
grad_clip
self
.
grad_clip
=
grad_clip
def
__call__
(
self
,
model
):
def
__call__
(
self
,
model
):
train_params
=
[
param
for
param
in
model
.
parameters
()
if
param
.
trainable
is
True
]
opt
=
optim
.
RMSProp
(
opt
=
optim
.
RMSProp
(
learning_rate
=
self
.
learning_rate
,
learning_rate
=
self
.
learning_rate
,
momentum
=
self
.
momentum
,
momentum
=
self
.
momentum
,
...
@@ -125,7 +134,7 @@ class RMSProp(object):
...
@@ -125,7 +134,7 @@ class RMSProp(object):
epsilon
=
self
.
epsilon
,
epsilon
=
self
.
epsilon
,
weight_decay
=
self
.
weight_decay
,
weight_decay
=
self
.
weight_decay
,
grad_clip
=
self
.
grad_clip
,
grad_clip
=
self
.
grad_clip
,
parameters
=
model
.
parameters
()
)
parameters
=
train_params
)
return
opt
return
opt
...
@@ -149,6 +158,9 @@ class Adadelta(object):
...
@@ -149,6 +158,9 @@ class Adadelta(object):
self
.
name
=
name
self
.
name
=
name
def
__call__
(
self
,
model
):
def
__call__
(
self
,
model
):
train_params
=
[
param
for
param
in
model
.
parameters
()
if
param
.
trainable
is
True
]
opt
=
optim
.
Adadelta
(
opt
=
optim
.
Adadelta
(
learning_rate
=
self
.
learning_rate
,
learning_rate
=
self
.
learning_rate
,
epsilon
=
self
.
epsilon
,
epsilon
=
self
.
epsilon
,
...
@@ -156,7 +168,7 @@ class Adadelta(object):
...
@@ -156,7 +168,7 @@ class Adadelta(object):
weight_decay
=
self
.
weight_decay
,
weight_decay
=
self
.
weight_decay
,
grad_clip
=
self
.
grad_clip
,
grad_clip
=
self
.
grad_clip
,
name
=
self
.
name
,
name
=
self
.
name
,
parameters
=
model
.
parameters
()
)
parameters
=
train_params
)
return
opt
return
opt
...
@@ -190,10 +202,13 @@ class AdamW(object):
...
@@ -190,10 +202,13 @@ class AdamW(object):
self
.
one_dim_param_no_weight_decay
=
one_dim_param_no_weight_decay
self
.
one_dim_param_no_weight_decay
=
one_dim_param_no_weight_decay
def
__call__
(
self
,
model
):
def
__call__
(
self
,
model
):
parameters
=
model
.
parameters
()
parameters
=
[
param
for
param
in
model
.
parameters
()
if
param
.
trainable
is
True
]
self
.
no_weight_decay_param_name_list
=
[
self
.
no_weight_decay_param_name_list
=
[
p
.
name
for
n
,
p
in
model
.
named_parameters
()
if
any
(
nd
in
n
for
nd
in
self
.
no_weight_decay_name_list
)
p
.
name
for
n
,
p
in
model
.
named_parameters
()
if
any
(
nd
in
n
for
nd
in
self
.
no_weight_decay_name_list
)
]
]
if
self
.
one_dim_param_no_weight_decay
:
if
self
.
one_dim_param_no_weight_decay
:
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment