Unverified Commit 8aa6eb10 authored by lvhan028's avatar lvhan028 Committed by GitHub
Browse files

change FasterTransformer to TurboMind (#37)

parent 70e6ab26
...@@ -13,7 +13,7 @@ ...@@ -13,7 +13,7 @@
# limitations under the License. # limitations under the License.
cmake_minimum_required(VERSION 3.11 FATAL_ERROR) # for PyTorch extensions, version should be greater than 3.13 cmake_minimum_required(VERSION 3.11 FATAL_ERROR) # for PyTorch extensions, version should be greater than 3.13
project(FasterTransformer LANGUAGES CXX CUDA) project(TurboMind LANGUAGES CXX CUDA)
find_package(CUDA 10.2 REQUIRED) find_package(CUDA 10.2 REQUIRED)
...@@ -354,18 +354,18 @@ set_target_properties(transformer-shared PROPERTIES LINKER_LANGUAGE CXX) ...@@ -354,18 +354,18 @@ set_target_properties(transformer-shared PROPERTIES LINKER_LANGUAGE CXX)
target_link_libraries(transformer-shared PUBLIC -lcudart -lcublas -lcublasLt -lcurand) target_link_libraries(transformer-shared PUBLIC -lcudart -lcublas -lcublasLt -lcurand)
include(GNUInstallDirs) include(GNUInstallDirs)
set(INSTALL_CONFIGDIR ${CMAKE_INSTALL_LIBDIR}/cmake/FasterTransformer) set(INSTALL_CONFIGDIR ${CMAKE_INSTALL_LIBDIR}/cmake/TurboMind)
include(CMakePackageConfigHelpers) include(CMakePackageConfigHelpers)
configure_package_config_file( configure_package_config_file(
${CMAKE_CURRENT_LIST_DIR}/cmake/FasterTransformerConfig.cmake.in ${CMAKE_CURRENT_LIST_DIR}/cmake/TurboMindConfig.cmake.in
${CMAKE_CURRENT_BINARY_DIR}/FasterTransformerConfig.cmake ${CMAKE_CURRENT_BINARY_DIR}/TurboMindConfig.cmake
INSTALL_DESTINATION ${INSTALL_CONFIGDIR} INSTALL_DESTINATION ${INSTALL_CONFIGDIR}
) )
install( install(
FILES FILES
${CMAKE_CURRENT_BINARY_DIR}/FasterTransformerConfig.cmake ${CMAKE_CURRENT_BINARY_DIR}/TurboMindConfig.cmake
DESTINATION ${INSTALL_CONFIGDIR} DESTINATION ${INSTALL_CONFIGDIR}
) )
...@@ -382,7 +382,7 @@ install( ...@@ -382,7 +382,7 @@ install(
EXPORT EXPORT
transformer-shared-targets transformer-shared-targets
FILE FILE
FasterTransformerTargets.cmake TurboMindTargets.cmake
DESTINATION DESTINATION
${INSTALL_CONFIGDIR} ${INSTALL_CONFIGDIR}
) )
...@@ -391,9 +391,9 @@ export( ...@@ -391,9 +391,9 @@ export(
EXPORT EXPORT
transformer-shared-targets transformer-shared-targets
FILE FILE
${CMAKE_CURRENT_BINARY_DIR}/FasterTransformerTargets.cmake ${CMAKE_CURRENT_BINARY_DIR}/TurboMindTargets.cmake
NAMESPACE NAMESPACE
TritonCore:: TritonCore::
) )
export(PACKAGE FasterTransformer) export(PACKAGE TurboMind)
...@@ -27,13 +27,13 @@ ...@@ -27,13 +27,13 @@
include(CMakeFindDependencyMacro) include(CMakeFindDependencyMacro)
get_filename_component( get_filename_component(
FASTERTRANSFORMER_CMAKE_DIR "${CMAKE_CURRENT_LIST_FILE}" PATH TURBOMIND_CMAKE_DIR "${CMAKE_CURRENT_LIST_FILE}" PATH
) )
list(APPEND CMAKE_MODULE_PATH ${FASTERTRANSFORMER_CMAKE_DIR}) list(APPEND CMAKE_MODULE_PATH ${TURBOMIND_CMAKE_DIR})
if(NOT TARGET transformer-shared) if(NOT TARGET transformer-shared)
include("${FASTERTRANSFORMER_CMAKE_DIR}/FasterTransformerTargets.cmake") include("${TURBOMIND_CMAKE_DIR}/TurboMindTargets.cmake")
endif() endif()
set(FASTERTRANSFORMER_LIBRARIES transformer-shared) set(TURBOMIND_LIBRARIES transformer-shared)
...@@ -13,14 +13,14 @@ class TestGemmDequantize(unittest.TestCase): ...@@ -13,14 +13,14 @@ class TestGemmDequantize(unittest.TestCase):
def setUp(self) -> None: def setUp(self) -> None:
torch.classes.load_library('lib/libth_transformer.so') torch.classes.load_library('lib/libth_transformer.so')
torch.classes.load_library('lib/libgemm_dq_unit_ops.so') torch.classes.load_library('lib/libgemm_dq_unit_ops.so')
self.unpack_packed_int4s = torch.ops.fastertransformer.unpack_int4_packed_tensor_to_int8 self.unpack_packed_int4s = torch.ops.turbomind.unpack_int4_packed_tensor_to_int8
self.pack_int4s = torch.ops.fastertransformer.pack_int8_tensor_to_packed_int4 self.pack_int4s = torch.ops.turbomind.pack_int8_tensor_to_packed_int4
self.fused_gemm_dq = torch.ops.gemm_dq_unit_ops.fused_gemm_dq self.fused_gemm_dq = torch.ops.gemm_dq_unit_ops.fused_gemm_dq
self.fused_gemm_dq_bias_act = torch.ops.gemm_dq_unit_ops.fused_gemm_dq_bias_act self.fused_gemm_dq_bias_act = torch.ops.gemm_dq_unit_ops.fused_gemm_dq_bias_act
self.bench = torch.ops.gemm_dq_unit_ops.benchmark_against_cublas_fp self.bench = torch.ops.gemm_dq_unit_ops.benchmark_against_cublas_fp
self.preprocess_weights_for_mixed_gemm = torch.ops.fastertransformer.preprocess_weights_for_mixed_gemm self.preprocess_weights_for_mixed_gemm = torch.ops.turbomind.preprocess_weights_for_mixed_gemm
self.symmetric_quantizer = torch.ops.fastertransformer._symmetric_quantize_last_axis_of_batched_matrix self.symmetric_quantizer = torch.ops.turbomind._symmetric_quantize_last_axis_of_batched_matrix
torch.manual_seed(734876213) torch.manual_seed(734876213)
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment