- 17 Oct, 2025 1 commit
-
-
Nikita Titov authored
-
- 24 Aug, 2025 1 commit
-
-
James Lamb authored
* [ci] [c++] use 'pre-commit' to run 'cpplint', upgrade to 'cpplint' 2.0.2 * remove bashisms * one more pipefail use * another pipefail
-
- 01 Dec, 2024 1 commit
-
-
Oliver Borchert authored
Co-authored-by:Nikita Titov <nekit94-08@mail.ru>
-
- 14 Jul, 2024 1 commit
-
-
James Lamb authored
[ci] [R-package] add CI jobs covering more CRAN "additional checks", fix R_NO_REMAP warnings (fixes #6369) (#6523) * [ci] [R-package] add CI jobs testing newer compilers * set up vignette-processing dependencies * remove tinytex * set up pandoc * allow NOTEs, one more R_NO_REMAP guard * restore other CI * one more R_NO_REMAP guard * fail builds producing more than 1 NOTE * update approach for running R CMD check * fix filepath * fix paths * fix variable name * echo lines * fix piping * fix file extension * clean up * make R CMD check stricter
-
- 04 May, 2024 1 commit
-
-
James Lamb authored
-
- 19 Mar, 2024 1 commit
-
-
James Lamb authored
-
- 07 Dec, 2023 1 commit
-
-
James Lamb authored
[R-package] [c++] add tighter multithreading control, avoid global OpenMP side effects (fixes #4705, fixes #5102) (#6226)
-
- 10 Oct, 2023 1 commit
-
-
James Lamb authored
-
- 30 Jun, 2023 1 commit
-
-
maskedcoder1337 authored
-
- 14 Feb, 2023 1 commit
-
-
Scott Votaw authored
* Add serialization of reference dataset * lint and missing file * Fixes from reviewers * responded to comments * revert sdk change
-
- 21 Nov, 2022 1 commit
-
-
James Lamb authored
-
- 29 May, 2022 1 commit
-
-
Nikita Titov authored
* Update tree.cpp * Update common.h * Update common.h
-
- 22 May, 2022 1 commit
-
-
James Lamb authored
-
- 16 Nov, 2021 1 commit
-
-
chjinche authored
* add customized parser support * fix typo of parser_config_file description * make delimiter as parameter of JoinedLines
-
- 28 Oct, 2021 1 commit
-
-
Zhiyuan He authored
* mock func for no openmp * use omp_get_max_threads Co-authored-by:hzy46 <email@example.com>
-
- 17 Sep, 2021 1 commit
-
-
david-cortes authored
* fix R memory leaks * attempt at solving linter complaints * fix compilation on windows * move R_API_BEGIN to correct place * make sure exception objects reach out of scope * better way to solve rchk complaints * remove goto statement
-
- 20 Aug, 2021 1 commit
-
-
James Lamb authored
-
- 07 Jul, 2021 1 commit
-
-
shiyu1994 authored
-
- 18 Jun, 2021 3 commits
-
-
-
Chen Yufei authored
* Log warning instead of fatal when parsing float get under/overflow. For texts that resolve to infinity, under or overflow should be accepted. * Remove outdated unit test. * empty commit to trigger ci
-
shiyu1994 authored
-
- 07 May, 2021 1 commit
-
-
Chen Yufei authored
* New build option: USE_PRECISE_TEXT_PARSER. Use fast_double_parser for text file parsing. For each number, fallback to strtod in case of parse failure. * Add benchmark for CSVParser with Atof and AtofPrecise. * Fix lint complaint. * Fix typo in open result error message. * Revert "Fix lint complaint." This reverts commit 92ab0b6bce9f17d7be9eaeb20f19d4a0a36f0387. * Revert "Add benchmark for CSVParser with Atof and AtofPrecise." This reverts commit 4f8639abd06c679d4382eb715a1793afd94df3d2. * Use AtofPrecise in Common::__StringToTHelper. * [option] precise_float_parser: precise float number parsing for text input. * Remove USE_PRECISE_TEXT_PARSER compile option. * test: add test for Common::AtofPrecise. * test: remove ChunkedArrayTest with 0 length. This triggers Log::Fatal which aborts the test program. * fix lint, add copyright. * Revert "test: remove ChunkedArrayTest with 0 length." This reverts commit 346c76affe9e78b6ca2738c4a56dbb9c00f31102. * Use LightGBM::Common::Sign * save precise_float_parser in model file. * Fix error checking in AtofPrecise. Add more test cases. * Remove test case that can't pass under macOS. * Apply suggestions from code review Co-authored-by:
Nikita Titov <nekit94-08@mail.ru> Co-authored-by:
Nikita Titov <nekit94-08@mail.ru>
-
- 04 May, 2021 1 commit
-
-
Andrew Ziem authored
* Correct spelling Most changes were in comments, and there were a few changes to literals for log output. There were no changes to variable names, function names, IDs, or functionality. * Clarify a phrase in a comment Co-authored-by:
James Lamb <jaylamb20@gmail.com> * Clarify a phrase in a comment Co-authored-by:
James Lamb <jaylamb20@gmail.com> * Clarify a phrase in a comment Co-authored-by:
James Lamb <jaylamb20@gmail.com> * Correct spelling Most are code comments, but one case is a literal in a logging message. There are a few grammar fixes too. Co-authored-by:
James Lamb <jaylamb20@gmail.com>
-
- 28 Mar, 2021 1 commit
-
-
Nikita Titov authored
* Update chunked_array.hpp * Update ChunkedArray_API_extensions.i * Update StringArray.i * apply cpplint to cpp tests * Update test_chunked_array to please cpplint (#4121) * Update test_chunked_array to please cpplint * Simplify var name * Add comment Co-authored-by:Alberto Ferreira <AlbertoEAF@users.noreply.github.com>
-
- 21 Mar, 2021 1 commit
-
-
Alberto Ferreira authored
* [feature] Add ChunkedArray to SWIG * Add ChunkedArray * Add ChunkedArray_API_extensions.i * Add SWIG class wrappers * Address some review comments * Fix linting issues * Move test to tests/test_ChunkedArray_manually.cpp * Add test note * Move ChunkedArray to include/LightGBM/utils/ * Declare more explicit types of ChunkedArray in the SWIG API. * Port ChunkedArray tests to googletest * Please C++ linter * Address StrikerRUS' review comments * Update SWIG doc & disable ChunkedArray<int64_t> * Use CHECK_EQ instead of assert * Change include order (linting) * Rename ChunkedArray -> chunked_array files * Change header guards * Address last comments from StrikerRUS
-
- 16 Mar, 2021 1 commit
-
-
mjmckp authored
* Fix index out-of-range exception generated by BaggingHelper on small datasets. Prior to this change, the line "score_t threshold = tmp_gradients[top_k - 1];" would generate an exception, since tmp_gradients would be empty when the cnt input value to the function is zero. * Update goss.hpp * Update goss.hpp * Add API method LGBM_BoosterPredictForMats which runs prediction on a data set given as of array of pointers to rows (as opposed to existing method LGBM_BoosterPredictForMat which requires data given as contiguous array) * Fix incorrect upstream merge * Add link to LightGBM.NET * Fix indenting to 2 spaces * Dummy edit to trigger CI * Dummy edit to trigger CI * remove duplicate functions from merge * Fix parsing of non-finite values. Current implementation silently returns zero when input string is "inf", "-inf", or "nan" when compiled with VS2017, so instead just explicitly check for these values and fail if there is no match. No attempt to optimise string allocations in this implementation since it is usually rarely invoked. * Dummy commit to trigger CI * Also handle -nan in double parsing method * Update include/LightGBM/utils/common.h Remove trailing whitespace to pass linting tests Co-authored-by:
Nikita Titov <nekit94-08@mail.ru> Co-authored-by:
matthew-peacock <matthew.peacock@whiteoakam.com> Co-authored-by:
Guolin Ke <guolin.ke@outlook.com> Co-authored-by:
Nikita Titov <nekit94-08@mail.ru>
-
- 13 Feb, 2021 2 commits
-
-
Benjamin Sergeant authored
* openmp_wrapper.h stubs signature use __GOMP_NOTHROW Fix #3915. OpenMP stubs do not use the noexcept attibute which is present in the gcc version of openmp, and which trigger compilation errors as seen below. dmlc-core uses the same technique and macro. /usr/lib/gcc/x86_64-linux-gnu/9/include/omp.h:114:12: error: declaration of ‘int omp_get_thread_num() noexcept’ has a different exception specifier 114 | extern int omp_get_thread_num (void) __GOMP_NOTHROW; | ^~~~~~~~~~~~~~~~~~ ... xxx/include/LightGBM/utils/openmp_wrapper.h:81:14: note: from previous declaration ‘int omp_get_thread_num()’ 81 | inline int omp_get_thread_num() {return 0;} | ^~~~~~~~~~~~~~~~~~ * move __GOMP_NOTHROW definition in the no open mp stub conditional branch * Update include/LightGBM/utils/openmp_wrapper.h Yes make sense, just changed it Co-authored-by:James Lamb <jaylamb20@gmail.com> * add NOLINT macro to disable cpplint on a safe line of code Co-authored-by:
James Lamb <jaylamb20@gmail.com>
-
mjmckp authored
Fix access violation exception that can occur during invocation of loop lambda function when inner_start >= inner_end in 'For' template (#3936) * Fix index out-of-range exception generated by BaggingHelper on small datasets. Prior to this change, the line "score_t threshold = tmp_gradients[top_k - 1];" would generate an exception, since tmp_gradients would be empty when the cnt input value to the function is zero. * Update goss.hpp * Update goss.hpp * Add API method LGBM_BoosterPredictForMats which runs prediction on a data set given as of array of pointers to rows (as opposed to existing method LGBM_BoosterPredictForMat which requires data given as contiguous array) * Fix incorrect upstream merge * Add link to LightGBM.NET * Fix indenting to 2 spaces * Dummy edit to trigger CI * Dummy edit to trigger CI * remove duplicate functions from merge * Fix access violation exception that can occur during invocation of loop lambda function when inner_start >= inner_end in 'For' template. In particular, this can occur in Tree::AddPredictionToScore on line 291 where the loop lambda function body (created by the PredictionFun macro) dereferences used_data_indices[start]. For reference, the particular case which triggered this exception in my case was: * start = 0 * end = 93,203 * n_block = 56 * min_block_size = 512 for which the BlockInfo method gave: * n_block = 56 * num_inner = 1,696 and so, for the case i=55 (i.e., the last case in the loop), we get * inner_start = start + num_inner * i = 93,280 which is greater than 'end' and hence triggers the exception. * Change formatting of proposed modification Co-authored-by:matthew-peacock <matthew.peacock@whiteoakam.com> Co-authored-by:
Guolin Ke <guolin.ke@outlook.com>
-
- 24 Dec, 2020 1 commit
-
-
Belinda Trotta authored
* Add Eigen library. * Working for simple test. * Apply changes to config params. * Handle nan data. * Update docs. * Add test. * Only load raw data if boosting=gbdt_linear * Remove unneeded code. * Minor updates. * Update to work with sk-learn interface. * Update to work with chunked datasets. * Throw error if we try to create a Booster with an already-constructed dataset having incompatible parameters. * Save raw data in binary dataset file. * Update docs and fix parameter checking. * Fix dataset loading. * Add test for regularization. * Fix bugs when saving and loading tree. * Add test for load/save linear model. * Remove unneeded code. * Fix case where not enough leaf data for linear model. * Simplify code. * Speed up code. * Speed up code. * Simplify code. * Speed up code. * Fix bugs. * Working version. * Store feature data column-wise (not fully working yet). * Fix bugs. * Speed up. * Speed up. * Remove unneeded code. * Small speedup. * Speed up. * Minor updates. * Remove unneeded code. * Fix bug. * Fix bug. * Speed up. * Speed up. * Simplify code. * Remove unneeded code. * Fix bug, add more tests. * Fix bug and add test. * Only store numerical features * Fix bug and speed up using templates. * Speed up prediction. * Fix bug with regularisation * Visual studio files. * Working version * Only check nans if necessary * Store coeff matrix as an array. * Align cache lines * Align cache lines * Preallocation coefficient calculation matrices * Small speedups * Small speedup * Reverse cache alignment changes * Change to dynamic schedule * Update docs. * Refactor so that linear tree learner is not a separate class. * Add refit capability. * Speed up * Small speedups. * Speed up add prediction to score. * Fix bug * Fix bug and speed up. * Speed up dataload. * Speed up dataload * Use vectors instead of pointers * Fix bug * Add OMP exception handling. * Change return type of LGBM_BoosterGetLinear to bool * Change return type of LGBM_BoosterGetLinear back to int, only parameter type needed to change * Remove unused internal_parent_ property of tree * Remove unused parameter to CreateTreeLearner * Remove reference to LinearTreeLearner * Minor style issues * Remove unneeded check * Reverse temporary testing change * Fix Visual Studio project files * Restore LightGBM.vcxproj.filters * Speed up * Speed up * Simplify code * Update docs * Simplify code * Initialise storage space for max num threads * Move Eigen to include directory and delete unused files * Remove old files. * Fix so it compiles with mingw * Fix gpu tree learner * Change AddPredictionToScore back to const * Fix python lint error * Fix C++ lint errors * Change eigen to a submodule * Update comment * Add the eigen folder * Try to fix build issues with eigen * Remove eigen files * Add eigen as submodule * Fix include paths * Exclude eigen files from Python linter * Ignore eigen folders for pydocstyle * Fix C++ linting errors * Fix docs * Fix docs * Exclude eigen directories from doxygen * Update manifest to include eigen * Update build_r to include eigen files * Fix compiler warnings * Store raw feature data as float * Use float for calculating linear coefficients * Remove eigen directory from GLOB * Don't compile linear model code when building R package * Fix doxygen issue * Fix lint issue * Fix lint issue * Remove uneeded code * Restore delected lines * Restore delected lines * Change return type of has_raw to bool * Update docs * Rename some variables and functions for readability * Make tree_learner parameter const in AddScore * Fix style issues * Pass vectors as const reference when setting tree properties * Make temporary storage of serial_tree_learner mutable so we can make the object's methods const * Remove get_raw_size, use num_numeric_features instead * Fix typo * Make contains_nan_ and any_nan_ properties immutable again * Remove data_has_nan_ property of tree * Remove temporary test code * Make linear_tree a dataset param * Fix lint error * Make LinearTreeLearner a separate class * Fix lint errors * Fix lint error * Add linear_tree_learner.o * Simulate omp_get_max_threads if openmp is not available * Update PushOneData to also store raw data. * Cast size to int * Fix bug in ReshapeRaw * Speed up code with multithreading * Use OMP_NUM_THREADS * Speed up with multithreading * Update to use ArrayToString * Fix tests * Fix test * Fix bug introduced in merge * Minor updates * Update docs
-
- 08 Dec, 2020 1 commit
-
-
Alberto Ferreira authored
* Fix LightGBM models locale sensitivity and improve R/W performance. When Java is used, the default C++ locale is broken. This is true for Java providers that use the C API or even Python models that require JEP. This patch solves that issue making the model reads/writes insensitive to such settings. To achieve it, within the model read/write codebase: - C++ streams are imbued with the classic locale - Calls to functions that are dependent on the locale are replaced - The default locale is not changed! This approach means: - The user's locale is never tampered with, avoiding issues such as https://github.com/microsoft/LightGBM/issues/2979 with the previous approach https://github.com/microsoft/LightGBM/pull/2891 - Datasets can still be read according the user's locale - The model file has a single format independent of locale Changes: - Add CommonC namespace which provides faster locale-independent versions of Common's methods - Model code makes conversions through CommonC - Cleanup unused Common methods - Performance improvements. Use fast libraries for locale-agnostic conversion: - value->string: https://github.com/fmtlib/fmt - string->double: https://github.com/lemire/fast_double_parser (10x faster double parsing according to their benchmark) Bugfixes: - https://github.com/microsoft/LightGBM/issues/2500 - https://github.com/microsoft/LightGBM/issues/2890 - https://github.com/ninia/jep/issues/205 (as it is related to LGBM as well) * Align CommonC namespace * Add new external_libs/ to python setup * Try fast_double_parser fix #1 Testing commit e09e5aad828bcb16bea7ed0ed8322e019112fdbe If it works it should fix more LGBM builds * CMake: Attempt to link fmt without explicit PUBLIC tag * Exclude external_libs from linting * Add exernal_libs to MANIFEST.in * Set dynamic linking option for fmt. * linting issues * Try to fix lint includes * Try to pass fPIC with static fmt lib * Try CMake P_I_C option with fmt library * [R-package] Add CMake support for R and CRAN * Cleanup CMakeLists * Try fmt hack to remove stdout * Switch to header-only mode * Add PRIVATE argument to target_link_libraries * use fmt in header-only mode * Remove CMakeLists comment * Change OpenMP to PUBLIC linking in Mac * Update fmt submodule to 7.1.2 * Use fmt in header-only-mode * Remove fmt from CMakeLists.txt * Upgrade fast_double_parser to v0.2.0 * Revert "Add PRIVATE argument to target_link_libraries" This reverts commit 3dd45dde7b92531b2530ab54522bb843c56227a7. * Address James Lamb's comments * Update R-package/.Rbuildignore Co-authored-by:James Lamb <jaylamb20@gmail.com> * Upgrade to fast_double_parser v0.3.0 - Solaris support * Use legacy code only in Solaris * Fix lint issues * Fix comment * Address StrikerRUS's comments (solaris ifdef). * Change header guards Co-authored-by:
James Lamb <jaylamb20@gmail.com>
-
- 23 Nov, 2020 1 commit
-
-
shiyu1994 authored
* remove max_block_size_ in train states (fix #3570) * avoid zero elements per row * add min constraint for min_block_size_
-
- 13 Nov, 2020 1 commit
-
-
shiyu1994 authored
* store without offset in multi_val_dense_bin * fix offset bug * add comment for offset * add comment for bin type selection * faster operations for offset * keep most freq bin in histogram for multi val dense * use original feature iterators * consider 9 cases (3 x 3) for multi val bin construction * fix dense bin setting * fix bin data in multi val group * fix offset of the first feature histogram * use float hist buf * avx in histogram construction * use avx for hist construction without prefetch * vectorize bin extraction * use only 128 vec * use avx2 * use vectorization for sparse row wise * add bit size for multi val dense bin * float with no vectorization * change multithreading strategy to dynamic * remove intrinsic header * fix dense multi val col copy * remove bit size * use large enough block size when the bin number is large * calc min block size by sparsity * rescale gradients * rollback gradients scaling * single precision histogram buffer as an option * add float hist buffer with thread buffer * fix setting zero in hist data * fix hist begin pointer in tree learners * remove debug logs * remove omp simd * update Makevars of R-package * fix feature group binary storing * two row wise for double hist buffer * add subfeature for two row wise * remove useless code and fix two row wise * refactor code * grouping the dense feature groups can get sparse multi val bin * clean format problems * one thread for two blocks in sep row wise * use ordered gradients for sep row wise * fix grad ptr * ordered grad with combined block for sep row wise * fix block threading * use the same min block size * rollback share min block size * remove logs * Update src/io/dataset.cpp Co-authored-by:
Guolin Ke <guolin.ke@outlook.com> * fix parameter description * remove sep_row_wise * remove check codes * add check for empty multi val bin * fix lint error * rollback changes in config.h * Apply suggestions from code review Co-authored-by:
Ubuntu <shiyu@gbdt-04.ren3kv4wanvufliwrpy4k03lsf.xx.internal.cloudapp.net> Co-authored-by:
Guolin Ke <guolin.ke@outlook.com>
-
- 27 Oct, 2020 2 commits
- 18 Oct, 2020 1 commit
-
-
James Lamb authored
* fix int64 write error * attempt * [WIP] [ci] [R-package] Add CI job that runs valgrind tests * update all-successful * install * executable * fix redirect stuff * Apply suggestions from code review Co-authored-by:
Guolin Ke <guolin.ke@outlook.com> * more flags * add mc to msvc proj * fix memory leak in mc * Update monotone_constraints.hpp * Update r_package.yml * remove R_INT64_PTR * disable openmp * Update gbdt_model_text.cpp * Update gbdt_model_text.cpp * Apply suggestions from code review * try to free vector * free more memories. * Update src/boosting/gbdt_model_text.cpp * fix using * try the UNPROTECT(1); * fix a const pointer * fix Common * reduce UNPROTECT * remove UNPROTECT(1); * fix null handle * fix predictor * use NULL after free * fix a leaking in test * try more fixes * test the effect of tests * throw exception in Fatal * add test back * Apply suggestions from code review * commet some tests * Apply suggestions from code review * Apply suggestions from code review * trying to comment out tests * Update openmp_wrapper.h * Apply suggestions from code review * Update configure * Update configure.ac * trying to uncomment * more comments * more uncommenting * more uncommenting * fix comment * more uncommenting * uncomment fully-commented out stuff * try uncommenting more dataset tests * uncommenting more tests * ok getting closer * more uncommenting * free dataset * skipping a test, more uncommenting * more skipping * re-enable OpenMP * allow on OpenMP thing * move valgrind to comment-only job * Apply suggestions from code review Co-authored-by:
Nikita Titov <nekit94-08@mail.ru> * changes from code review * Apply suggestions from code review Co-authored-by:
Nikita Titov <nekit94-08@mail.ru> * linting * issue comments too * remove issue_comment Co-authored-by:
Guolin Ke <guolin.ke@outlook.com> Co-authored-by:
Nikita Titov <nekit94-08@mail.ru>
-
- 30 Sep, 2020 2 commits
-
-
Guolin Ke authored
* Update serial_tree_learner.cpp * Update src/treelearner/serial_tree_learner.cpp * stable multi-threading reduction * Update src/treelearner/serial_tree_learner.cpp * more fixes * Apply suggestions from code review * Apply suggestions from code review * Update src/boosting/gbdt.cpp
-
Guolin Ke authored
* fix dataset binary file alignment * many fixes * fix warnings * fix bug * Update file_io.cpp * Update file_io.cpp * simplify code * Apply suggestions from code review * general * remove unneeded alignment * Update file_io.h * int32 to byte8 alignment * Apply suggestions from code review * Apply suggestions from code review
-
- 29 Jul, 2020 2 commits
-
-
James Lamb authored
-
James Lamb authored
* [R-package] make package installable with CRAN toolchain (fixes #2960) * Apply suggestions from code review Co-authored-by:
Nikita Titov <nekit94-08@mail.ru> * remove GPU stuff * use wildcard to find objects to build * use -lomp * build configure before moving files * using wildcard for objects * Update .github/workflows/main.yml Co-authored-by:
Nikita Titov <nekit94-08@mail.ru> * add explicit objects back * reduce allowed R CMD check NOTEs and catch stderr from build-cran-package on Windows * fixing things * pin autoconf version * show diff * add automake back * run less checks * command was in the wrong place * fix autoconf version * change strategy for handling configure * fix Rbuildignore * fix NOTEs * fix notes about unrecognized files * fixing extra files * remove USE_R35 * add OpenMP check for Mac CRAN build * run all checks * Apply suggestions from code review Co-authored-by:
Nikita Titov <nekit94-08@mail.ru> * suggestions from code review * undo indenting * remove 03 from Makevars.win.in * update language about OpenMP in configure script * checking if configure.ac check works * add autoconf back * remove testing code in configure.ac * more fixes for CI on configure script * print git diff * add VERSION.txt when checking configure * fix relative paths * remove git diff Co-authored-by:
Nikita Titov <nekit94-08@mail.ru>
-
- 19 Jul, 2020 1 commit
-
-
Joan Fontanals authored
* Add capability to get possible max and min values for a model * Change implementation to have return value in tree.cpp, change naming to upper and lower bound, move implementation to gdbt.cpp * Update include/LightGBM/c_api.h Co-Authored-By:
Nikita Titov <nekit94-08@mail.ru> * Change iteration to avoid potential overflow, add bindings to R and Python and a basic test * Adjust test values * Consider const correctness and multithreading protection * Put everything possible as const * Include shared_mutex, for now as unique_lock * Update test values * Put everything possible as const * Include shared_mutex, for now as unique_lock * Make PredictSingleRow const and share the lock with other reading threads * Update test values * Add test to check that model is exactly the same in all platforms * Try to parse the model to get the expected values * Try to parse the model to get the expected values * Fix implementation, num_leaves can be lower than the leaf_value_ size * Do not check for num_leaves to be smaller than actual size and get back to test with hardcoded value * Change test order * Add gpu_use_dp option in test * Remove helper test method * Remove TODO * Add preprocessing option to compile with c++17 * Update python-package/setup.py Co-Authored-By:
Nikita Titov <nekit94-08@mail.ru> * Remove unwanted changes * Move option * Fix problems introduced by conflict fix * Avoid switching to c++17 and use yamc mutex library to access shared lock functionality * Add extra yamc include * Change header order * some lint fix * change include order and remove some extra blank lines * Further fix lint issues * Update c_api.cpp * Further fix lint issues * Move yamc include files to a new yamc folder * Use standard unique_lock * Update windows/LightGBM.vcxproj Co-authored-by:
Guolin Ke <guolin.ke@outlook.com> * Update windows/LightGBM.vcxproj.filters Co-authored-by:
Guolin Ke <guolin.ke@outlook.com> * Update windows/LightGBM.vcxproj.filters Co-authored-by:
Nikita Titov <nekit94-08@mail.ru> * Update windows/LightGBM.vcxproj.filters Co-authored-by:
Nikita Titov <nekit94-08@mail.ru> * Update windows/LightGBM.vcxproj.filters Co-authored-by:
Nikita Titov <nekit94-08@mail.ru> * Fix problems coming from merge conflict resolution Co-authored-by:
Nikita Titov <nekit94-08@mail.ru> Co-authored-by:
joanfontanals <jfontanals@ntent.com> Co-authored-by:
Guolin Ke <guolin.ke@outlook.com>
-