Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
ollama
Commits
1deafd82
Unverified
Commit
1deafd82
authored
Jan 08, 2025
by
Jeffrey Morgan
Committed by
GitHub
Jan 08, 2025
Browse files
llama: update vendored code to commit 46e3556 (#8308)
parent
57f038ec
Changes
305
Show whitespace changes
Inline
Side-by-side
Showing
20 changed files
with
27 additions
and
20 deletions
+27
-20
llama/ggml-cuda/template-instances/mmq-instance-q3_k.cu
llama/ggml-cuda/template-instances/mmq-instance-q3_k.cu
+1
-1
llama/ggml-cuda/template-instances/mmq-instance-q4_0.cu
llama/ggml-cuda/template-instances/mmq-instance-q4_0.cu
+1
-1
llama/ggml-cuda/template-instances/mmq-instance-q4_1.cu
llama/ggml-cuda/template-instances/mmq-instance-q4_1.cu
+1
-1
llama/ggml-cuda/template-instances/mmq-instance-q4_k.cu
llama/ggml-cuda/template-instances/mmq-instance-q4_k.cu
+1
-1
llama/ggml-cuda/template-instances/mmq-instance-q5_0.cu
llama/ggml-cuda/template-instances/mmq-instance-q5_0.cu
+1
-1
llama/ggml-cuda/template-instances/mmq-instance-q5_1.cu
llama/ggml-cuda/template-instances/mmq-instance-q5_1.cu
+1
-1
llama/ggml-cuda/template-instances/mmq-instance-q5_k.cu
llama/ggml-cuda/template-instances/mmq-instance-q5_k.cu
+1
-1
llama/ggml-cuda/template-instances/mmq-instance-q6_k.cu
llama/ggml-cuda/template-instances/mmq-instance-q6_k.cu
+1
-1
llama/ggml-cuda/template-instances/mmq-instance-q8_0.cu
llama/ggml-cuda/template-instances/mmq-instance-q8_0.cu
+1
-1
llama/ggml-cuda/tsembd.cu
llama/ggml-cuda/tsembd.cu
+1
-1
llama/ggml-cuda/tsembd.cuh
llama/ggml-cuda/tsembd.cuh
+1
-1
llama/ggml-cuda/unary.cu
llama/ggml-cuda/unary.cu
+1
-1
llama/ggml-cuda/unary.cuh
llama/ggml-cuda/unary.cuh
+1
-1
llama/ggml-cuda/upscale.cu
llama/ggml-cuda/upscale.cu
+1
-1
llama/ggml-cuda/upscale.cuh
llama/ggml-cuda/upscale.cuh
+1
-1
llama/ggml-cuda/vecdotq.cuh
llama/ggml-cuda/vecdotq.cuh
+1
-1
llama/ggml-cuda/vendors/cuda.h
llama/ggml-cuda/vendors/cuda.h
+2
-1
llama/ggml-cuda/vendors/hip.h
llama/ggml-cuda/vendors/hip.h
+4
-1
llama/ggml-cuda/vendors/musa.h
llama/ggml-cuda/vendors/musa.h
+4
-1
llama/ggml-cuda/wkv6.cu
llama/ggml-cuda/wkv6.cu
+1
-1
No files found.
llama/ggml-cuda/template-instances/mmq-instance-q3_k.cu
View file @
1deafd82
/**
* llama.cpp - commit
ba1cb19cdd0d92e012e0f6e009e0620f854b6afd
- do not edit this file
* llama.cpp - commit
46e3556e01b824e52395fb050b29804b6cff2a7c
- do not edit this file
*
* MIT License
*
...
...
llama/ggml-cuda/template-instances/mmq-instance-q4_0.cu
View file @
1deafd82
/**
* llama.cpp - commit
ba1cb19cdd0d92e012e0f6e009e0620f854b6afd
- do not edit this file
* llama.cpp - commit
46e3556e01b824e52395fb050b29804b6cff2a7c
- do not edit this file
*
* MIT License
*
...
...
llama/ggml-cuda/template-instances/mmq-instance-q4_1.cu
View file @
1deafd82
/**
* llama.cpp - commit
ba1cb19cdd0d92e012e0f6e009e0620f854b6afd
- do not edit this file
* llama.cpp - commit
46e3556e01b824e52395fb050b29804b6cff2a7c
- do not edit this file
*
* MIT License
*
...
...
llama/ggml-cuda/template-instances/mmq-instance-q4_k.cu
View file @
1deafd82
/**
* llama.cpp - commit
ba1cb19cdd0d92e012e0f6e009e0620f854b6afd
- do not edit this file
* llama.cpp - commit
46e3556e01b824e52395fb050b29804b6cff2a7c
- do not edit this file
*
* MIT License
*
...
...
llama/ggml-cuda/template-instances/mmq-instance-q5_0.cu
View file @
1deafd82
/**
* llama.cpp - commit
ba1cb19cdd0d92e012e0f6e009e0620f854b6afd
- do not edit this file
* llama.cpp - commit
46e3556e01b824e52395fb050b29804b6cff2a7c
- do not edit this file
*
* MIT License
*
...
...
llama/ggml-cuda/template-instances/mmq-instance-q5_1.cu
View file @
1deafd82
/**
* llama.cpp - commit
ba1cb19cdd0d92e012e0f6e009e0620f854b6afd
- do not edit this file
* llama.cpp - commit
46e3556e01b824e52395fb050b29804b6cff2a7c
- do not edit this file
*
* MIT License
*
...
...
llama/ggml-cuda/template-instances/mmq-instance-q5_k.cu
View file @
1deafd82
/**
* llama.cpp - commit
ba1cb19cdd0d92e012e0f6e009e0620f854b6afd
- do not edit this file
* llama.cpp - commit
46e3556e01b824e52395fb050b29804b6cff2a7c
- do not edit this file
*
* MIT License
*
...
...
llama/ggml-cuda/template-instances/mmq-instance-q6_k.cu
View file @
1deafd82
/**
* llama.cpp - commit
ba1cb19cdd0d92e012e0f6e009e0620f854b6afd
- do not edit this file
* llama.cpp - commit
46e3556e01b824e52395fb050b29804b6cff2a7c
- do not edit this file
*
* MIT License
*
...
...
llama/ggml-cuda/template-instances/mmq-instance-q8_0.cu
View file @
1deafd82
/**
* llama.cpp - commit
ba1cb19cdd0d92e012e0f6e009e0620f854b6afd
- do not edit this file
* llama.cpp - commit
46e3556e01b824e52395fb050b29804b6cff2a7c
- do not edit this file
*
* MIT License
*
...
...
llama/ggml-cuda/tsembd.cu
View file @
1deafd82
/**
* llama.cpp - commit
ba1cb19cdd0d92e012e0f6e009e0620f854b6afd
- do not edit this file
* llama.cpp - commit
46e3556e01b824e52395fb050b29804b6cff2a7c
- do not edit this file
*
* MIT License
*
...
...
llama/ggml-cuda/tsembd.cuh
View file @
1deafd82
/**
* llama.cpp - commit
ba1cb19cdd0d92e012e0f6e009e0620f854b6afd
- do not edit this file
* llama.cpp - commit
46e3556e01b824e52395fb050b29804b6cff2a7c
- do not edit this file
*
* MIT License
*
...
...
llama/ggml-cuda/unary.cu
View file @
1deafd82
/**
* llama.cpp - commit
ba1cb19cdd0d92e012e0f6e009e0620f854b6afd
- do not edit this file
* llama.cpp - commit
46e3556e01b824e52395fb050b29804b6cff2a7c
- do not edit this file
*
* MIT License
*
...
...
llama/ggml-cuda/unary.cuh
View file @
1deafd82
/**
* llama.cpp - commit
ba1cb19cdd0d92e012e0f6e009e0620f854b6afd
- do not edit this file
* llama.cpp - commit
46e3556e01b824e52395fb050b29804b6cff2a7c
- do not edit this file
*
* MIT License
*
...
...
llama/ggml-cuda/upscale.cu
View file @
1deafd82
/**
* llama.cpp - commit
ba1cb19cdd0d92e012e0f6e009e0620f854b6afd
- do not edit this file
* llama.cpp - commit
46e3556e01b824e52395fb050b29804b6cff2a7c
- do not edit this file
*
* MIT License
*
...
...
llama/ggml-cuda/upscale.cuh
View file @
1deafd82
/**
* llama.cpp - commit
ba1cb19cdd0d92e012e0f6e009e0620f854b6afd
- do not edit this file
* llama.cpp - commit
46e3556e01b824e52395fb050b29804b6cff2a7c
- do not edit this file
*
* MIT License
*
...
...
llama/ggml-cuda/vecdotq.cuh
View file @
1deafd82
/**
* llama.cpp - commit
ba1cb19cdd0d92e012e0f6e009e0620f854b6afd
- do not edit this file
* llama.cpp - commit
46e3556e01b824e52395fb050b29804b6cff2a7c
- do not edit this file
*
* MIT License
*
...
...
llama/ggml-cuda/vendors/cuda.h
View file @
1deafd82
/**
* llama.cpp - commit
ba1cb19cdd0d92e012e0f6e009e0620f854b6afd
- do not edit this file
* llama.cpp - commit
46e3556e01b824e52395fb050b29804b6cff2a7c
- do not edit this file
*
* MIT License
*
...
...
@@ -29,6 +29,7 @@
#include <cuda_runtime.h>
#include <cuda.h>
#include <cublas_v2.h>
#include <cuda_bf16.h>
#include <cuda_fp16.h>
#if CUDART_VERSION < 11020
...
...
llama/ggml-cuda/vendors/hip.h
View file @
1deafd82
/**
* llama.cpp - commit
ba1cb19cdd0d92e012e0f6e009e0620f854b6afd
- do not edit this file
* llama.cpp - commit
46e3556e01b824e52395fb050b29804b6cff2a7c
- do not edit this file
*
* MIT License
*
...
...
@@ -29,6 +29,7 @@
#include <hip/hip_runtime.h>
#include <hipblas/hipblas.h>
#include <hip/hip_fp16.h>
#include <hip/hip_bfloat16.h>
#ifdef __HIP_PLATFORM_AMD__
// for rocblas_initialize()
#include "rocblas/rocblas.h"
...
...
@@ -147,6 +148,8 @@
#define __has_builtin(x) 0
#endif
typedef
hip_bfloat16
nv_bfloat16
;
typedef
int8_t
int8x4_t
__attribute__
((
ext_vector_type
(
4
)));
typedef
uint8_t
uint8x4_t
__attribute__
((
ext_vector_type
(
4
)));
static
__device__
__forceinline__
int
__vsubss4
(
const
int
a
,
const
int
b
)
{
...
...
llama/ggml-cuda/vendors/musa.h
View file @
1deafd82
/**
* llama.cpp - commit
ba1cb19cdd0d92e012e0f6e009e0620f854b6afd
- do not edit this file
* llama.cpp - commit
46e3556e01b824e52395fb050b29804b6cff2a7c
- do not edit this file
*
* MIT License
*
...
...
@@ -29,6 +29,7 @@
#include <musa_runtime.h>
#include <musa.h>
#include <mublas.h>
#include <musa_bf16.h>
#include <musa_fp16.h>
#define CUBLAS_COMPUTE_16F CUDA_R_16F
#define CUBLAS_COMPUTE_32F CUDA_R_32F
...
...
@@ -158,3 +159,5 @@
#define cudaKernelNodeParams musaKernelNodeParams
#define cudaStreamCaptureModeRelaxed musaStreamCaptureModeRelaxed
#define cudaStreamEndCapture musaStreamEndCapture
typedef
mt_bfloat16
nv_bfloat16
;
llama/ggml-cuda/wkv6.cu
View file @
1deafd82
/**
* llama.cpp - commit
ba1cb19cdd0d92e012e0f6e009e0620f854b6afd
- do not edit this file
* llama.cpp - commit
46e3556e01b824e52395fb050b29804b6cff2a7c
- do not edit this file
*
* MIT License
*
...
...
Prev
1
…
7
8
9
10
11
12
13
14
15
16
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment