Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
Lmdeploy
Commits
8aa6eb10
Unverified
Commit
8aa6eb10
authored
Jul 01, 2023
by
lvhan028
Committed by
GitHub
Jul 01, 2023
Browse files
change FasterTransformer to TurboMind (#37)
parent
70e6ab26
Changes
3
Show whitespace changes
Inline
Side-by-side
Showing
3 changed files
with
16 additions
and
16 deletions
+16
-16
CMakeLists.txt
CMakeLists.txt
+8
-8
cmake/TurboMindConfig.cmake.in
cmake/TurboMindConfig.cmake.in
+4
-4
tests/gemm_dequantize/th_gemm_dequantize.py
tests/gemm_dequantize/th_gemm_dequantize.py
+4
-4
No files found.
CMakeLists.txt
View file @
8aa6eb10
...
@@ -13,7 +13,7 @@
...
@@ -13,7 +13,7 @@
# limitations under the License.
# limitations under the License.
cmake_minimum_required
(
VERSION 3.11 FATAL_ERROR
)
# for PyTorch extensions, version should be greater than 3.13
cmake_minimum_required
(
VERSION 3.11 FATAL_ERROR
)
# for PyTorch extensions, version should be greater than 3.13
project
(
FasterTransformer
LANGUAGES CXX CUDA
)
project
(
TurboMind
LANGUAGES CXX CUDA
)
find_package
(
CUDA 10.2 REQUIRED
)
find_package
(
CUDA 10.2 REQUIRED
)
...
@@ -354,18 +354,18 @@ set_target_properties(transformer-shared PROPERTIES LINKER_LANGUAGE CXX)
...
@@ -354,18 +354,18 @@ set_target_properties(transformer-shared PROPERTIES LINKER_LANGUAGE CXX)
target_link_libraries
(
transformer-shared PUBLIC -lcudart -lcublas -lcublasLt -lcurand
)
target_link_libraries
(
transformer-shared PUBLIC -lcudart -lcublas -lcublasLt -lcurand
)
include
(
GNUInstallDirs
)
include
(
GNUInstallDirs
)
set
(
INSTALL_CONFIGDIR
${
CMAKE_INSTALL_LIBDIR
}
/cmake/
FasterTransformer
)
set
(
INSTALL_CONFIGDIR
${
CMAKE_INSTALL_LIBDIR
}
/cmake/
TurboMind
)
include
(
CMakePackageConfigHelpers
)
include
(
CMakePackageConfigHelpers
)
configure_package_config_file
(
configure_package_config_file
(
${
CMAKE_CURRENT_LIST_DIR
}
/cmake/
FasterTransformer
Config.cmake.in
${
CMAKE_CURRENT_LIST_DIR
}
/cmake/
TurboMind
Config.cmake.in
${
CMAKE_CURRENT_BINARY_DIR
}
/
FasterTransformer
Config.cmake
${
CMAKE_CURRENT_BINARY_DIR
}
/
TurboMind
Config.cmake
INSTALL_DESTINATION
${
INSTALL_CONFIGDIR
}
INSTALL_DESTINATION
${
INSTALL_CONFIGDIR
}
)
)
install
(
install
(
FILES
FILES
${
CMAKE_CURRENT_BINARY_DIR
}
/
FasterTransformer
Config.cmake
${
CMAKE_CURRENT_BINARY_DIR
}
/
TurboMind
Config.cmake
DESTINATION
${
INSTALL_CONFIGDIR
}
DESTINATION
${
INSTALL_CONFIGDIR
}
)
)
...
@@ -382,7 +382,7 @@ install(
...
@@ -382,7 +382,7 @@ install(
EXPORT
EXPORT
transformer-shared-targets
transformer-shared-targets
FILE
FILE
FasterTransformer
Targets.cmake
TurboMind
Targets.cmake
DESTINATION
DESTINATION
${
INSTALL_CONFIGDIR
}
${
INSTALL_CONFIGDIR
}
)
)
...
@@ -391,9 +391,9 @@ export(
...
@@ -391,9 +391,9 @@ export(
EXPORT
EXPORT
transformer-shared-targets
transformer-shared-targets
FILE
FILE
${
CMAKE_CURRENT_BINARY_DIR
}
/
FasterTransformer
Targets.cmake
${
CMAKE_CURRENT_BINARY_DIR
}
/
TurboMind
Targets.cmake
NAMESPACE
NAMESPACE
TritonCore::
TritonCore::
)
)
export
(
PACKAGE
FasterTransformer
)
export
(
PACKAGE
TurboMind
)
cmake/
FasterTransformer
Config.cmake.in
→
cmake/
TurboMind
Config.cmake.in
View file @
8aa6eb10
...
@@ -27,13 +27,13 @@
...
@@ -27,13 +27,13 @@
include(CMakeFindDependencyMacro)
include(CMakeFindDependencyMacro)
get_filename_component(
get_filename_component(
FASTERTRANSFORMER
_CMAKE_DIR "${CMAKE_CURRENT_LIST_FILE}" PATH
TURBOMIND
_CMAKE_DIR "${CMAKE_CURRENT_LIST_FILE}" PATH
)
)
list(APPEND CMAKE_MODULE_PATH ${
FASTERTRANSFORMER
_CMAKE_DIR})
list(APPEND CMAKE_MODULE_PATH ${
TURBOMIND
_CMAKE_DIR})
if(NOT TARGET transformer-shared)
if(NOT TARGET transformer-shared)
include("${
FASTERTRANSFORMER_CMAKE_DIR}/FasterTransformer
Targets.cmake")
include("${
TURBOMIND_CMAKE_DIR}/TurboMind
Targets.cmake")
endif()
endif()
set(
FASTERTRANSFORMER
_LIBRARIES transformer-shared)
set(
TURBOMIND
_LIBRARIES transformer-shared)
tests/gemm_dequantize/th_gemm_dequantize.py
View file @
8aa6eb10
...
@@ -13,14 +13,14 @@ class TestGemmDequantize(unittest.TestCase):
...
@@ -13,14 +13,14 @@ class TestGemmDequantize(unittest.TestCase):
def
setUp
(
self
)
->
None
:
def
setUp
(
self
)
->
None
:
torch
.
classes
.
load_library
(
'lib/libth_transformer.so'
)
torch
.
classes
.
load_library
(
'lib/libth_transformer.so'
)
torch
.
classes
.
load_library
(
'lib/libgemm_dq_unit_ops.so'
)
torch
.
classes
.
load_library
(
'lib/libgemm_dq_unit_ops.so'
)
self
.
unpack_packed_int4s
=
torch
.
ops
.
fastertransformer
.
unpack_int4_packed_tensor_to_int8
self
.
unpack_packed_int4s
=
torch
.
ops
.
turbomind
.
unpack_int4_packed_tensor_to_int8
self
.
pack_int4s
=
torch
.
ops
.
fastertransformer
.
pack_int8_tensor_to_packed_int4
self
.
pack_int4s
=
torch
.
ops
.
turbomind
.
pack_int8_tensor_to_packed_int4
self
.
fused_gemm_dq
=
torch
.
ops
.
gemm_dq_unit_ops
.
fused_gemm_dq
self
.
fused_gemm_dq
=
torch
.
ops
.
gemm_dq_unit_ops
.
fused_gemm_dq
self
.
fused_gemm_dq_bias_act
=
torch
.
ops
.
gemm_dq_unit_ops
.
fused_gemm_dq_bias_act
self
.
fused_gemm_dq_bias_act
=
torch
.
ops
.
gemm_dq_unit_ops
.
fused_gemm_dq_bias_act
self
.
bench
=
torch
.
ops
.
gemm_dq_unit_ops
.
benchmark_against_cublas_fp
self
.
bench
=
torch
.
ops
.
gemm_dq_unit_ops
.
benchmark_against_cublas_fp
self
.
preprocess_weights_for_mixed_gemm
=
torch
.
ops
.
fastertransformer
.
preprocess_weights_for_mixed_gemm
self
.
preprocess_weights_for_mixed_gemm
=
torch
.
ops
.
turbomind
.
preprocess_weights_for_mixed_gemm
self
.
symmetric_quantizer
=
torch
.
ops
.
fastertransformer
.
_symmetric_quantize_last_axis_of_batched_matrix
self
.
symmetric_quantizer
=
torch
.
ops
.
turbomind
.
_symmetric_quantize_last_axis_of_batched_matrix
torch
.
manual_seed
(
734876213
)
torch
.
manual_seed
(
734876213
)
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment