Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
ComfyUI
Commits
9cc12c83
"...git@developer.sourcefind.cn:wangsen/paddle_dbnet.git" did not exist on "1751be75b0c3e22d2bdfa9cebdbf01995a83d2c0"
Commit
9cc12c83
authored
Aug 14, 2023
by
comfyanonymous
Browse files
CLIPVisionEncode can now encode multiple images.
parent
0cb6dac9
Changes
3
Hide whitespace changes
Inline
Side-by-side
Showing
3 changed files
with
12 additions
and
12 deletions
+12
-12
comfy/clip_vision.py
comfy/clip_vision.py
+2
-2
comfy/model_base.py
comfy/model_base.py
+9
-9
nodes.py
nodes.py
+1
-1
No files found.
comfy/clip_vision.py
View file @
9cc12c83
...
@@ -24,8 +24,8 @@ class ClipVisionModel():
...
@@ -24,8 +24,8 @@ class ClipVisionModel():
return
self
.
model
.
load_state_dict
(
sd
,
strict
=
False
)
return
self
.
model
.
load_state_dict
(
sd
,
strict
=
False
)
def
encode_image
(
self
,
image
):
def
encode_image
(
self
,
image
):
img
=
torch
.
clip
((
255.
*
image
[
0
]
),
0
,
255
).
round
().
int
()
img
=
torch
.
clip
((
255.
*
image
),
0
,
255
).
round
().
int
()
inputs
=
self
.
processor
(
images
=
[
img
]
,
return_tensors
=
"pt"
)
inputs
=
self
.
processor
(
images
=
img
,
return_tensors
=
"pt"
)
outputs
=
self
.
model
(
**
inputs
)
outputs
=
self
.
model
(
**
inputs
)
return
outputs
return
outputs
...
...
comfy/model_base.py
View file @
9cc12c83
...
@@ -120,15 +120,15 @@ class SD21UNCLIP(BaseModel):
...
@@ -120,15 +120,15 @@ class SD21UNCLIP(BaseModel):
weights
=
[]
weights
=
[]
noise_aug
=
[]
noise_aug
=
[]
for
unclip_cond
in
unclip_conditioning
:
for
unclip_cond
in
unclip_conditioning
:
adm_cond
=
unclip_cond
[
"clip_vision_output"
].
image_embeds
for
adm_cond
in
unclip_cond
[
"clip_vision_output"
].
image_embeds
:
weight
=
unclip_cond
[
"strength"
]
weight
=
unclip_cond
[
"strength"
]
noise_augment
=
unclip_cond
[
"noise_augmentation"
]
noise_augment
=
unclip_cond
[
"noise_augmentation"
]
noise_level
=
round
((
self
.
noise_augmentor
.
max_noise_level
-
1
)
*
noise_augment
)
noise_level
=
round
((
self
.
noise_augmentor
.
max_noise_level
-
1
)
*
noise_augment
)
c_adm
,
noise_level_emb
=
self
.
noise_augmentor
(
adm_cond
.
to
(
device
),
noise_level
=
torch
.
tensor
([
noise_level
],
device
=
device
))
c_adm
,
noise_level_emb
=
self
.
noise_augmentor
(
adm_cond
.
to
(
device
),
noise_level
=
torch
.
tensor
([
noise_level
],
device
=
device
))
adm_out
=
torch
.
cat
((
c_adm
,
noise_level_emb
),
1
)
*
weight
adm_out
=
torch
.
cat
((
c_adm
,
noise_level_emb
),
1
)
*
weight
weights
.
append
(
weight
)
weights
.
append
(
weight
)
noise_aug
.
append
(
noise_augment
)
noise_aug
.
append
(
noise_augment
)
adm_inputs
.
append
(
adm_out
)
adm_inputs
.
append
(
adm_out
)
if
len
(
noise_aug
)
>
1
:
if
len
(
noise_aug
)
>
1
:
adm_out
=
torch
.
stack
(
adm_inputs
).
sum
(
0
)
adm_out
=
torch
.
stack
(
adm_inputs
).
sum
(
0
)
...
...
nodes.py
View file @
9cc12c83
...
@@ -771,7 +771,7 @@ class StyleModelApply:
...
@@ -771,7 +771,7 @@ class StyleModelApply:
CATEGORY
=
"conditioning/style_model"
CATEGORY
=
"conditioning/style_model"
def
apply_stylemodel
(
self
,
clip_vision_output
,
style_model
,
conditioning
):
def
apply_stylemodel
(
self
,
clip_vision_output
,
style_model
,
conditioning
):
cond
=
style_model
.
get_cond
(
clip_vision_output
)
cond
=
style_model
.
get_cond
(
clip_vision_output
)
.
flatten
(
start_dim
=
0
,
end_dim
=
1
).
unsqueeze
(
dim
=
0
)
c
=
[]
c
=
[]
for
t
in
conditioning
:
for
t
in
conditioning
:
n
=
[
torch
.
cat
((
t
[
0
],
cond
),
dim
=
1
),
t
[
1
].
copy
()]
n
=
[
torch
.
cat
((
t
[
0
],
cond
),
dim
=
1
),
t
[
1
].
copy
()]
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment