Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
ModelZoo
ResNet50_tensorflow
Commits
a55cf4d3
Commit
a55cf4d3
authored
Aug 25, 2022
by
Yeqing Li
Committed by
A. Unique TensorFlower
Aug 25, 2022
Browse files
Internal change
PiperOrigin-RevId: 470172579
parent
462bb852
Changes
2
Show whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
70 additions
and
52 deletions
+70
-52
official/projects/vit/modeling/vit.py
official/projects/vit/modeling/vit.py
+2
-52
official/projects/vit/modeling/vit_specs.py
official/projects/vit/modeling/vit_specs.py
+68
-0
No files found.
official/projects/vit/modeling/vit.py
View file @
a55cf4d3
...
...
@@ -17,68 +17,18 @@ from typing import Optional, Tuple
from
absl
import
logging
import
immutabledict
#
import immutabledict
import
tensorflow
as
tf
from
official.modeling
import
activations
from
official.projects.vit.modeling
import
nn_blocks
from
official.projects.vit.modeling.vit_specs
import
VIT_SPECS
from
official.vision.modeling.backbones
import
factory
from
official.vision.modeling.layers
import
nn_layers
layers
=
tf
.
keras
.
layers
VIT_SPECS
=
immutabledict
.
immutabledict
({
'vit-ti16'
:
dict
(
hidden_size
=
192
,
patch_size
=
16
,
transformer
=
dict
(
mlp_dim
=
768
,
num_heads
=
3
,
num_layers
=
12
),
),
'vit-s16'
:
dict
(
hidden_size
=
384
,
patch_size
=
16
,
transformer
=
dict
(
mlp_dim
=
1536
,
num_heads
=
6
,
num_layers
=
12
),
),
'vit-b16'
:
dict
(
hidden_size
=
768
,
patch_size
=
16
,
transformer
=
dict
(
mlp_dim
=
3072
,
num_heads
=
12
,
num_layers
=
12
),
),
'vit-b32'
:
dict
(
hidden_size
=
768
,
patch_size
=
32
,
transformer
=
dict
(
mlp_dim
=
3072
,
num_heads
=
12
,
num_layers
=
12
),
),
'vit-l16'
:
dict
(
hidden_size
=
1024
,
patch_size
=
16
,
transformer
=
dict
(
mlp_dim
=
4096
,
num_heads
=
16
,
num_layers
=
24
),
),
'vit-l32'
:
dict
(
hidden_size
=
1024
,
patch_size
=
32
,
transformer
=
dict
(
mlp_dim
=
4096
,
num_heads
=
16
,
num_layers
=
24
),
),
'vit-h14'
:
dict
(
hidden_size
=
1280
,
patch_size
=
14
,
transformer
=
dict
(
mlp_dim
=
5120
,
num_heads
=
16
,
num_layers
=
32
),
),
'vit-g14'
:
dict
(
hidden_size
=
1664
,
patch_size
=
14
,
transformer
=
dict
(
mlp_dim
=
8192
,
num_heads
=
16
,
num_layers
=
48
),
),
})
class
AddPositionEmbs
(
tf
.
keras
.
layers
.
Layer
):
"""Adds (optionally learned) positional embeddings to the inputs."""
...
...
official/projects/vit/modeling/vit_specs.py
0 → 100644
View file @
a55cf4d3
# Copyright 2022 The TensorFlow Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""VisionTransformer backbone specs."""
import
immutabledict
VIT_SPECS
=
immutabledict
.
immutabledict
({
'vit-ti16'
:
dict
(
hidden_size
=
192
,
patch_size
=
16
,
transformer
=
dict
(
mlp_dim
=
768
,
num_heads
=
3
,
num_layers
=
12
),
),
'vit-s16'
:
dict
(
hidden_size
=
384
,
patch_size
=
16
,
transformer
=
dict
(
mlp_dim
=
1536
,
num_heads
=
6
,
num_layers
=
12
),
),
'vit-b16'
:
dict
(
hidden_size
=
768
,
patch_size
=
16
,
transformer
=
dict
(
mlp_dim
=
3072
,
num_heads
=
12
,
num_layers
=
12
),
),
'vit-b32'
:
dict
(
hidden_size
=
768
,
patch_size
=
32
,
transformer
=
dict
(
mlp_dim
=
3072
,
num_heads
=
12
,
num_layers
=
12
),
),
'vit-l16'
:
dict
(
hidden_size
=
1024
,
patch_size
=
16
,
transformer
=
dict
(
mlp_dim
=
4096
,
num_heads
=
16
,
num_layers
=
24
),
),
'vit-l32'
:
dict
(
hidden_size
=
1024
,
patch_size
=
32
,
transformer
=
dict
(
mlp_dim
=
4096
,
num_heads
=
16
,
num_layers
=
24
),
),
'vit-h14'
:
dict
(
hidden_size
=
1280
,
patch_size
=
14
,
transformer
=
dict
(
mlp_dim
=
5120
,
num_heads
=
16
,
num_layers
=
32
),
),
'vit-g14'
:
dict
(
hidden_size
=
1664
,
patch_size
=
14
,
transformer
=
dict
(
mlp_dim
=
8192
,
num_heads
=
16
,
num_layers
=
48
),
),
})
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment