1. 14 Jun, 2024 1 commit
  2. 13 Jun, 2024 3 commits
  3. 03 Jun, 2024 2 commits
  4. 28 May, 2024 1 commit
  5. 27 May, 2024 2 commits
  6. 23 May, 2024 1 commit
  7. 30 Apr, 2024 1 commit
  8. 29 Apr, 2024 1 commit
  9. 23 Apr, 2024 3 commits
  10. 20 Apr, 2024 1 commit
  11. 19 Apr, 2024 1 commit
  12. 20 Mar, 2024 1 commit
  13. 19 Mar, 2024 2 commits
  14. 18 Mar, 2024 2 commits
  15. 04 Mar, 2024 1 commit
  16. 21 Feb, 2024 1 commit
  17. 20 Feb, 2024 1 commit
  18. 15 Feb, 2024 1 commit
  19. 10 Feb, 2024 2 commits
  20. 22 Jan, 2024 1 commit
  21. 17 Jan, 2024 1 commit
  22. 09 Jan, 2024 1 commit
  23. 22 Dec, 2023 1 commit
  24. 19 Dec, 2023 2 commits
  25. 07 Dec, 2023 1 commit
  26. 02 Nov, 2023 1 commit
  27. 01 Nov, 2023 1 commit
  28. 25 Oct, 2023 1 commit
  29. 10 Oct, 2023 1 commit
  30. 08 Oct, 2023 1 commit
    • shiyu1994's avatar
      [CUDA] CUDA Quantized Training (fixes #5606) (#5933) · f901f471
      shiyu1994 authored
      * add quantized training (first stage)
      
      * add histogram construction functions for integer gradients
      
      * add stochastic rounding
      
      * update docs
      
      * fix compilation errors by adding template instantiations
      
      * update files for compilation
      
      * fix compilation of gpu version
      
      * initialize gradient discretizer before share states
      
      * add a test case for quantized training
      
      * add quantized training for data distributed training
      
      * Delete origin.pred
      
      * Delete ifelse.pred
      
      * Delete LightGBM_model.txt
      
      * remove useless changes
      
      * fix lint error
      
      * remove debug loggings
      
      * fix mismatch of vector and allocator types
      
      * remove changes in main.cpp
      
      * fix bugs with uninitialized gradient discretizer
      
      * initialize ordered gradients in gradient discretizer
      
      * disable quantized training with gpu and cuda
      
      fix msvc compilation errors and warnings
      
      * fix bug in data parallel tree learner
      
      * make quantized training test deterministic
      
      * make quantized training in test case more accurate
      
      * refactor test_quantized_training
      
      * fix leaf splits initialization with quantized training
      
      * check distributed quantized training result
      
      * add cuda gradient discretizer
      
      * add quantized training for CUDA version in tree learner
      
      * remove cuda computability 6.1 and 6.2
      
      * fix parts of gpu quantized training errors and warnings
      
      * fix build-python.sh to install locally built version
      
      * fix memory access bugs
      
      * fix lint errors
      
      * mark cuda quantized training on cuda with categorical features as unsupported
      
      * rename cuda_utils.h to cuda_utils.hu
      
      * enable quantized training with cuda
      
      * fix cuda quantized training with sparse row data
      
      * allow using global memory buffer in histogram construction with cuda quantized training
      
      * recover build-python.sh
      
      enlarge allowed package size to 100M
      f901f471