Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
gaoqiong
composable_kernel_ROCM
Commits
29dbf01c
Commit
29dbf01c
authored
Jan 13, 2025
by
Po Yen, Chen
Browse files
Move all headers under same directory
parent
aee06365
Changes
12
Hide whitespace changes
Inline
Side-by-side
Showing
12 changed files
with
6 additions
and
5 deletions
+6
-5
example/ck_tile/18_paged_attention/include/attention_dtypes.h
...ple/ck_tile/18_paged_attention/include/attention_dtypes.h
+0
-0
example/ck_tile/18_paged_attention/include/attention_generic.cuh
.../ck_tile/18_paged_attention/include/attention_generic.cuh
+0
-0
example/ck_tile/18_paged_attention/include/dtype_bfloat16.cuh
...ple/ck_tile/18_paged_attention/include/dtype_bfloat16.cuh
+0
-0
example/ck_tile/18_paged_attention/include/dtype_float16.cuh
example/ck_tile/18_paged_attention/include/dtype_float16.cuh
+0
-0
example/ck_tile/18_paged_attention/include/dtype_float32.cuh
example/ck_tile/18_paged_attention/include/dtype_float32.cuh
+0
-0
example/ck_tile/18_paged_attention/include/dtype_fp8.cuh
example/ck_tile/18_paged_attention/include/dtype_fp8.cuh
+0
-0
example/ck_tile/18_paged_attention/include/hip_compat.h
example/ck_tile/18_paged_attention/include/hip_compat.h
+0
-0
example/ck_tile/18_paged_attention/include/hip_float8.h
example/ck_tile/18_paged_attention/include/hip_float8.h
+0
-0
example/ck_tile/18_paged_attention/include/hip_float8_impl.h
example/ck_tile/18_paged_attention/include/hip_float8_impl.h
+0
-0
example/ck_tile/18_paged_attention/include/paged_attention_kernel.hpp
...ile/18_paged_attention/include/paged_attention_kernel.hpp
+4
-3
example/ck_tile/18_paged_attention/include/quant_utils.cuh
example/ck_tile/18_paged_attention/include/quant_utils.cuh
+1
-1
example/ck_tile/18_paged_attention/itfs/paged_attention.cpp
example/ck_tile/18_paged_attention/itfs/paged_attention.cpp
+1
-1
No files found.
example/ck_tile/18_paged_attention/include/attention
/attention
_dtypes.h
→
example/ck_tile/18_paged_attention/include/attention_dtypes.h
View file @
29dbf01c
File moved
example/ck_tile/18_paged_attention/include/attention
/attention
_generic.cuh
→
example/ck_tile/18_paged_attention/include/attention_generic.cuh
View file @
29dbf01c
File moved
example/ck_tile/18_paged_attention/include/
attention/
dtype_bfloat16.cuh
→
example/ck_tile/18_paged_attention/include/dtype_bfloat16.cuh
View file @
29dbf01c
File moved
example/ck_tile/18_paged_attention/include/
attention/
dtype_float16.cuh
→
example/ck_tile/18_paged_attention/include/dtype_float16.cuh
View file @
29dbf01c
File moved
example/ck_tile/18_paged_attention/include/
attention/
dtype_float32.cuh
→
example/ck_tile/18_paged_attention/include/dtype_float32.cuh
View file @
29dbf01c
File moved
example/ck_tile/18_paged_attention/include/
attention/
dtype_fp8.cuh
→
example/ck_tile/18_paged_attention/include/dtype_fp8.cuh
View file @
29dbf01c
File moved
example/ck_tile/18_paged_attention/include/
cuda
_compat.h
→
example/ck_tile/18_paged_attention/include/
hip
_compat.h
View file @
29dbf01c
File moved
example/ck_tile/18_paged_attention/include/
quantization/fp8/amd/
hip_float8.h
→
example/ck_tile/18_paged_attention/include/hip_float8.h
View file @
29dbf01c
File moved
example/ck_tile/18_paged_attention/include/
quantization/fp8/amd/
hip_float8_impl.h
→
example/ck_tile/18_paged_attention/include/hip_float8_impl.h
View file @
29dbf01c
File moved
example/ck_tile/18_paged_attention/include/
kernel/
paged_attention_kernel.hpp
→
example/ck_tile/18_paged_attention/include/paged_attention_kernel.hpp
View file @
29dbf01c
...
...
@@ -17,12 +17,13 @@
#pragma once
#include <hip/hip_bf16.h>
#include "cuda_compat.h"
#include <algorithm>
#include <cfloat>
#include "attention/dtype_fp8.cuh"
#include "quantization/fp8/amd/quant_utils.cuh"
#include "dtype_fp8.cuh"
#include "hip_compat.h"
#include "quant_utils.cuh"
#if defined(__HIPCC__) && \
(defined(__gfx90a__) || defined(__gfx940__) || defined(__gfx941__) || defined(__gfx942__))
...
...
example/ck_tile/18_paged_attention/include/
quantization/fp8/amd/
quant_utils.cuh
→
example/ck_tile/18_paged_attention/include/quant_utils.cuh
View file @
29dbf01c
...
...
@@ -5,7 +5,7 @@
#include <hip/hip_bf16.h>
#include <hip/hip_bfloat16.h>
#include "
../../../attention/
attention_dtypes.h"
#include "attention_dtypes.h"
namespace
vllm
{
#ifdef USE_ROCM
...
...
example/ck_tile/18_paged_attention/itfs/paged_attention.cpp
View file @
29dbf01c
...
...
@@ -19,7 +19,7 @@
#include <hip/hip_runtime.h>
#include "paged_attention.hpp"
#include "
kernel/
paged_attention_kernel.hpp"
#include "paged_attention_kernel.hpp"
#define LAUNCH_CUSTOM_ATTENTION(GQA_RATIO) \
paged_attention_ll4mi_QKV_kernel<T, \
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment