Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
llama.cpp
Commits
e6262690
Commit
e6262690
authored
Dec 05, 2024
by
xuxzh1
🎱
Browse files
fix
parent
d5e51aa9
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
1 addition
and
1 deletion
+1
-1
ggml/src/ggml-cuda/softmax.cu
ggml/src/ggml-cuda/softmax.cu
+1
-1
No files found.
ggml/src/ggml-cuda/softmax.cu
View file @
e6262690
...
...
@@ -12,7 +12,7 @@ __device__ float __forceinline__ t2f32<half>(half val) {
}
template
<
bool
vals_smem
,
int
ncols_template
,
int
block_size_template
,
typename
T
>
static
__global__
void
soft_max_f32
(
const
float
*
x
,
const
T
*
mask
,
float
*
dst
,
const
int
ncols_par
,
const
int
nrows_y
,
const
float
scale
,
const
float
max_bias
,
const
float
m0
,
const
float
m1
,
uint32_t
n_head_log2
)
{
static
__global__
void
__launch_bounds__
(
1024
)
soft_max_f32
(
const
float
*
x
,
const
T
*
mask
,
float
*
dst
,
const
int
ncols_par
,
const
int
nrows_y
,
const
float
scale
,
const
float
max_bias
,
const
float
m0
,
const
float
m1
,
uint32_t
n_head_log2
)
{
const
int
ncols
=
ncols_template
==
0
?
ncols_par
:
ncols_template
;
const
int
tid
=
threadIdx
.
x
;
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment