Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
ModelZoo
Qwen_lmdeploy
Commits
8aa6eb10
Unverified
Commit
8aa6eb10
authored
Jul 01, 2023
by
lvhan028
Committed by
GitHub
Jul 01, 2023
Browse files
change FasterTransformer to TurboMind (#37)
parent
70e6ab26
Changes
3
Hide whitespace changes
Inline
Side-by-side
Showing
3 changed files
with
16 additions
and
16 deletions
+16
-16
CMakeLists.txt
CMakeLists.txt
+8
-8
cmake/TurboMindConfig.cmake.in
cmake/TurboMindConfig.cmake.in
+4
-4
tests/gemm_dequantize/th_gemm_dequantize.py
tests/gemm_dequantize/th_gemm_dequantize.py
+4
-4
No files found.
CMakeLists.txt
View file @
8aa6eb10
...
@@ -13,7 +13,7 @@
...
@@ -13,7 +13,7 @@
# limitations under the License.
# limitations under the License.
cmake_minimum_required
(
VERSION 3.11 FATAL_ERROR
)
# for PyTorch extensions, version should be greater than 3.13
cmake_minimum_required
(
VERSION 3.11 FATAL_ERROR
)
# for PyTorch extensions, version should be greater than 3.13
project
(
FasterTransformer
LANGUAGES CXX CUDA
)
project
(
TurboMind
LANGUAGES CXX CUDA
)
find_package
(
CUDA 10.2 REQUIRED
)
find_package
(
CUDA 10.2 REQUIRED
)
...
@@ -354,18 +354,18 @@ set_target_properties(transformer-shared PROPERTIES LINKER_LANGUAGE CXX)
...
@@ -354,18 +354,18 @@ set_target_properties(transformer-shared PROPERTIES LINKER_LANGUAGE CXX)
target_link_libraries
(
transformer-shared PUBLIC -lcudart -lcublas -lcublasLt -lcurand
)
target_link_libraries
(
transformer-shared PUBLIC -lcudart -lcublas -lcublasLt -lcurand
)
include
(
GNUInstallDirs
)
include
(
GNUInstallDirs
)
set
(
INSTALL_CONFIGDIR
${
CMAKE_INSTALL_LIBDIR
}
/cmake/
FasterTransformer
)
set
(
INSTALL_CONFIGDIR
${
CMAKE_INSTALL_LIBDIR
}
/cmake/
TurboMind
)
include
(
CMakePackageConfigHelpers
)
include
(
CMakePackageConfigHelpers
)
configure_package_config_file
(
configure_package_config_file
(
${
CMAKE_CURRENT_LIST_DIR
}
/cmake/
FasterTransformer
Config.cmake.in
${
CMAKE_CURRENT_LIST_DIR
}
/cmake/
TurboMind
Config.cmake.in
${
CMAKE_CURRENT_BINARY_DIR
}
/
FasterTransformer
Config.cmake
${
CMAKE_CURRENT_BINARY_DIR
}
/
TurboMind
Config.cmake
INSTALL_DESTINATION
${
INSTALL_CONFIGDIR
}
INSTALL_DESTINATION
${
INSTALL_CONFIGDIR
}
)
)
install
(
install
(
FILES
FILES
${
CMAKE_CURRENT_BINARY_DIR
}
/
FasterTransformer
Config.cmake
${
CMAKE_CURRENT_BINARY_DIR
}
/
TurboMind
Config.cmake
DESTINATION
${
INSTALL_CONFIGDIR
}
DESTINATION
${
INSTALL_CONFIGDIR
}
)
)
...
@@ -382,7 +382,7 @@ install(
...
@@ -382,7 +382,7 @@ install(
EXPORT
EXPORT
transformer-shared-targets
transformer-shared-targets
FILE
FILE
FasterTransformer
Targets.cmake
TurboMind
Targets.cmake
DESTINATION
DESTINATION
${
INSTALL_CONFIGDIR
}
${
INSTALL_CONFIGDIR
}
)
)
...
@@ -391,9 +391,9 @@ export(
...
@@ -391,9 +391,9 @@ export(
EXPORT
EXPORT
transformer-shared-targets
transformer-shared-targets
FILE
FILE
${
CMAKE_CURRENT_BINARY_DIR
}
/
FasterTransformer
Targets.cmake
${
CMAKE_CURRENT_BINARY_DIR
}
/
TurboMind
Targets.cmake
NAMESPACE
NAMESPACE
TritonCore::
TritonCore::
)
)
export
(
PACKAGE
FasterTransformer
)
export
(
PACKAGE
TurboMind
)
cmake/
FasterTransformer
Config.cmake.in
→
cmake/
TurboMind
Config.cmake.in
View file @
8aa6eb10
...
@@ -27,13 +27,13 @@
...
@@ -27,13 +27,13 @@
include(CMakeFindDependencyMacro)
include(CMakeFindDependencyMacro)
get_filename_component(
get_filename_component(
FASTERTRANSFORMER
_CMAKE_DIR "${CMAKE_CURRENT_LIST_FILE}" PATH
TURBOMIND
_CMAKE_DIR "${CMAKE_CURRENT_LIST_FILE}" PATH
)
)
list(APPEND CMAKE_MODULE_PATH ${
FASTERTRANSFORMER
_CMAKE_DIR})
list(APPEND CMAKE_MODULE_PATH ${
TURBOMIND
_CMAKE_DIR})
if(NOT TARGET transformer-shared)
if(NOT TARGET transformer-shared)
include("${
FASTERTRANSFORMER_CMAKE_DIR}/FasterTransformer
Targets.cmake")
include("${
TURBOMIND_CMAKE_DIR}/TurboMind
Targets.cmake")
endif()
endif()
set(
FASTERTRANSFORMER
_LIBRARIES transformer-shared)
set(
TURBOMIND
_LIBRARIES transformer-shared)
tests/gemm_dequantize/th_gemm_dequantize.py
View file @
8aa6eb10
...
@@ -13,14 +13,14 @@ class TestGemmDequantize(unittest.TestCase):
...
@@ -13,14 +13,14 @@ class TestGemmDequantize(unittest.TestCase):
def
setUp
(
self
)
->
None
:
def
setUp
(
self
)
->
None
:
torch
.
classes
.
load_library
(
'lib/libth_transformer.so'
)
torch
.
classes
.
load_library
(
'lib/libth_transformer.so'
)
torch
.
classes
.
load_library
(
'lib/libgemm_dq_unit_ops.so'
)
torch
.
classes
.
load_library
(
'lib/libgemm_dq_unit_ops.so'
)
self
.
unpack_packed_int4s
=
torch
.
ops
.
fastertransformer
.
unpack_int4_packed_tensor_to_int8
self
.
unpack_packed_int4s
=
torch
.
ops
.
turbomind
.
unpack_int4_packed_tensor_to_int8
self
.
pack_int4s
=
torch
.
ops
.
fastertransformer
.
pack_int8_tensor_to_packed_int4
self
.
pack_int4s
=
torch
.
ops
.
turbomind
.
pack_int8_tensor_to_packed_int4
self
.
fused_gemm_dq
=
torch
.
ops
.
gemm_dq_unit_ops
.
fused_gemm_dq
self
.
fused_gemm_dq
=
torch
.
ops
.
gemm_dq_unit_ops
.
fused_gemm_dq
self
.
fused_gemm_dq_bias_act
=
torch
.
ops
.
gemm_dq_unit_ops
.
fused_gemm_dq_bias_act
self
.
fused_gemm_dq_bias_act
=
torch
.
ops
.
gemm_dq_unit_ops
.
fused_gemm_dq_bias_act
self
.
bench
=
torch
.
ops
.
gemm_dq_unit_ops
.
benchmark_against_cublas_fp
self
.
bench
=
torch
.
ops
.
gemm_dq_unit_ops
.
benchmark_against_cublas_fp
self
.
preprocess_weights_for_mixed_gemm
=
torch
.
ops
.
fastertransformer
.
preprocess_weights_for_mixed_gemm
self
.
preprocess_weights_for_mixed_gemm
=
torch
.
ops
.
turbomind
.
preprocess_weights_for_mixed_gemm
self
.
symmetric_quantizer
=
torch
.
ops
.
fastertransformer
.
_symmetric_quantize_last_axis_of_batched_matrix
self
.
symmetric_quantizer
=
torch
.
ops
.
turbomind
.
_symmetric_quantize_last_axis_of_batched_matrix
torch
.
manual_seed
(
734876213
)
torch
.
manual_seed
(
734876213
)
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment