1. 26 Sep, 2019 1 commit
  2. 15 Sep, 2019 1 commit
    • kenmatsu4's avatar
      [python] Bug fix for first_metric_only on earlystopping. (#2209) · 84754399
      kenmatsu4 authored
      * Bug fix for first_metric_only if the first metric is train metric.
      
      * Update bug fix for feval issue.
      
      * Disable feval for first_metric_only.
      
      * Additional test items.
      
      * Fix wrong assertEqual settings & formating.
      
      * Change dataset of test.
      
      * Fix random seed for test.
      
      * Modiry assumed test result due to different sklearn verion between CI and local.
      
      * Remove f-string
      
      * Applying variable  assumed test result for test.
      
      * Fix flake8 error.
      
      * Modifying  in accordance with review comments.
      
      * Modifying for pylint.
      
      * simplified tests
      
      * Deleting error criteria `if eval_metric is None`.
      
      * Delete test items of classification.
      
      * Simplifying if condition.
      
      * Applying first_metric_only for sklearn wrapper.
      
      * Modifying test_sklearn for comforming to python 2.x
      
      * Fix flake8 error.
      
      * Additional fix for sklearn and add tests.
      
      * Bug fix and add test cases.
      
      * some refactor
      
      * fixed lint
      
      * fixed lint
      
      * Fix duplicated metrics scores to pass the test.
      
      * Fix the case first_metric_only not in params.
      
      * Converting metrics aliases.
      
      * Add comment.
      
      * Modify comment for pylint.
      
      * Modify comment for pydocstyle.
      
      * Using split test set for two eval_set.
      
      * added test case for metric aliases and length checks
      
      * minor style fixes
      
      * fixed rmse name and alias position
      
      * Fix the case metric=[]
      
      * Fix using env.model._train_data_name
      
      * Fix wrong test condition.
      
      * Move initial process to _init() func.
      
      * Modify test setting for test_sklearn & training data matching on callback.py
      
      * test_sklearn.py
      -> A test case for training is wrong, so fixed.
      
      * callback.py
      -> A condition of if statement for detecting test dataset is wrong, so fixed.
      
      * Support composite name metrics.
      
      * Remove metric check process & reduce redundant test cases.
      
      For #2273 fixed not only the order of metrics in cpp, removing metric check process at callback.py
      
      * Revised according to the matters pointed out on a review.
      
      * increased code readability
      
      * Fix the issue of order of validation set.
      
      * Changing to OrderdDict from default dict for score result.
      
      * added missed check in cv function for first_metric_only and feval co-occurrence
      
      * keep order only for metrics but not for datasets in best_score
      
      * move OrderedDict initialization to init phase
      
      * fixed minor printing issues
      
      * move first metric detection to init phase and split can be performed without checks
      
      * split only once during callback
      
      * removed excess code
      
      * fixed typo in variable name and squashed ifs
      
      * use setdefault
      
      * hotfix
      
      * fixed failing test
      
      * refined tests
      
      * refined sklearn test
      
      * Making "feval" effective on early stopping.
      
      * allow feval and first_metric_only for cv
      
      * removed unused code
      
      * added tests for feval
      
      * fixed printing
      
      * add note about whitespaces in feval name
      
      * Modifying final iteration process in case valid set is  training data.
      84754399
  3. 12 Sep, 2019 1 commit
  4. 09 Sep, 2019 1 commit
  5. 08 Sep, 2019 1 commit
    • CharlesAuguste's avatar
      [python] Improved python tree plots (#2304) · f52be9be
      CharlesAuguste authored
      * Some basic changes to the plot of the trees to make them readable.
      
      * Squeezed the information in the nodes.
      
      * Added colouring when a dictionnary mapping the features to the constraints is passed.
      
      * Fix spaces.
      
      * Added data percentage as an option in the nodes.
      
      * Squeezed the information in the leaves.
      
      * Important information is now in bold.
      
      * Added a legend for the color of monotone splits.
      
      * Changed "split_gain" to "gain" and "internal_value" to "value".
      
      * Sqeezed leaves a bit more.
      
      * Changed description in the legend.
      
      * Revert "Sqeezed leaves a bit more."
      
      This reverts commit dd8bf14a3ba604b0dfae3b7bb1c64b6784d15e03.
      
      * Increased the readability for the gain.
      
      * Tidied up the legend.
      
      * Added the data percentage in the leaves.
      
      * Added the monotone constraints to the dumped model.
      
      * Monotone constraints are now specified automatically when plotting trees.
      
      * Raise an exception instead of the bug that was here before.
      
      * Removed operators on the branches for a clearer design.
      
      * Small cleaning of the code.
      
      * Setting a monotone constraint on a categorical feature now returns an exception instead of doing nothing.
      
      * Fix bug when monotone constraints are empty.
      
      * Fix another bug when monotone constraints are empty.
      
      * Variable name change.
      
      * Added is / isn't on every edge of the trees.
      
      * Fix test "tree_create_digraph".
      
      * Add new test for plotting trees with monotone constraints.
      
      * Typo.
      
      * Update documentation of categorical features.
      
      * Typo.
      
      * Information in nodes more explicit.
      
      * Used regular strings instead of raw strings.
      
      * Small refactoring.
      
      * Some cleaning.
      
      * Added future statement.
      
      * Changed output for consistency.
      
      * Updated documentation.
      
      * Added comments for colors.
      
      * Changed text on edges for more clarity.
      
      * Small refactoring.
      
      * Modified text in leaves for consistency with nodes.
      
      * Updated default values and documentaton for consistency.
      
      * Replaced CHECK with Log::Fatal for user-friendliness.
      
      * Updated tests.
      
      * Typo.
      
      * Simplify imports.
      
      * Swapped count and weight to improve readibility of the leaves in the plotted trees.
      
      * Thresholds in bold.
      
      * Made information in nodes written in a specific order.
      
      * Added information to clarify legend.
      
      * Code cleaning.
      f52be9be
  6. 03 Sep, 2019 2 commits
  7. 02 Sep, 2019 1 commit
  8. 24 Aug, 2019 1 commit
    • Guolin Ke's avatar
      normalize the lambdas in lambdamart objective (#2331) · 0dfda826
      Guolin Ke authored
      * norm the lambda scores
      
      * change default to false
      
      * update doc
      
      * typo
      
      * Update Parameters.rst
      
      * Update config.h
      
      * Update test_sklearn.py
      
      * Update test_sklearn.py
      
      * Update test_sklearn.py
      
      * Update test_sklearn.py
      
      * Update test_sklearn.py
      
      * Update rank_objective.hpp
      
      * Update Parameters.rst
      
      * Update config.h
      
      * Update test_sklearn.py
      
      * Update test_sklearn.py
      
      * Update test_sklearn.py
      0dfda826
  9. 20 Aug, 2019 1 commit
  10. 17 Aug, 2019 1 commit
    • sbruch's avatar
      sigmoid_ in grad and hess for rank objective (#2322) · aee92f63
      sbruch authored
      * Lambdas and hessians need to factor sigmoid_ into the computation. Additionally, the sigmoid function has an arbitrary factor of 2 in the exponent; it is not just non-standard but the gradients are not computed correctly anyway.
      
      * Update unit test
      
      * Also remove a heuristic that normalizes the gradient by the difference in scores.
      
      * Also fix unit test after removing the heuristic
      aee92f63
  11. 16 Aug, 2019 1 commit
    • Belinda Trotta's avatar
      Bug fix: small values of max_bin cause program to crash (#2299) · c421f898
      Belinda Trotta authored
      * Fix bug where small values of max_bin cause crash.
      
      * Revert "Fix bug where small values of max_bin cause crash."
      
      This reverts commit fe5c8e2547057c1fa5750bcddd359dd7708fab4b.
      
      * Fix bug where small values of max_bin cause crash.
      
      * Reset random seed in test, remove extra blank line.
      
      * Minor bug fix. Remove extra blank line.
      
      * Change old test to account for new binning behavior.
      c421f898
  12. 13 Aug, 2019 1 commit
  13. 24 Jul, 2019 1 commit
  14. 12 Jul, 2019 1 commit
    • Guolin Ke's avatar
      fix init_model with subset (#2252) · 7360cff9
      Guolin Ke authored
      * fix init_model with subset
      
      * Update basic.py
      
      * added test
      
      * fix predictor naming issue
      
      * Update basic.py
      
      * fix bug
      
      * fix pylint
      
      * fix comments
      
      * Update basic.py
      
      * Update basic.py
      
      * updated test
      
      * fixed bug
      
      * fixed lint
      
      * fix warning
      
      * add get_data before initial prediction
      
      * refine the warning in get_data
      
      * refine warning
      
      * Update basic.py
      7360cff9
  15. 09 Jul, 2019 1 commit
  16. 08 Jul, 2019 1 commit
    • Belinda Trotta's avatar
      Max bin by feature (#2190) · 291752de
      Belinda Trotta authored
      * Add parameter max_bin_by_feature.
      
      * Fix minor bug.
      
      * Fix minor bug.
      
      * Fix calculation of header size for writing binary file.
      
      * Fix style issues.
      
      * Fix python style issue.
      
      * Fix test and python style issue.
      291752de
  17. 20 Jun, 2019 1 commit
  18. 04 Jun, 2019 1 commit
  19. 27 May, 2019 1 commit
  20. 26 May, 2019 1 commit
    • Belinda Trotta's avatar
      Top k multi error (#2178) · b3db9e92
      Belinda Trotta authored
      * Implement top-k multiclass error metric. Add new parameter top_k_threshold.
      
      * Add test for multiclass metrics
      
      * Make test less sensitive to avoid floating-point issues.
      
      * Change tabs to spaces.
      
      * Fix problem with test in Python 2. Refactor to use np.testing. Decrease number of training rounds so loss is larger and easier to compare.
      
      * Move multiclass tests into test_engine.py
      
      * Change parameter name from top_k_threshold to multi_error_top_k.
      
      * Fix top-k error metric to handle case where scores are equal. Update tests and docs.
      
      * Change name of top-k metric to multi_error@k.
      
      * Change tabs to spaces.
      
      * Fix formatting.
      
      * Fix minor issues in docs.
      b3db9e92
  21. 15 May, 2019 1 commit
  22. 08 May, 2019 1 commit
  23. 01 May, 2019 1 commit
    • Nikita Titov's avatar
      [python] added plot_split_value_histogram function (#2043) · 611cf5d4
      Nikita Titov authored
      * added plot_split_value_histogram function
      
      * updated init module
      
      * added plot split value histogram example
      
      * added plot_split_value_histogram to notebook
      
      * added test
      
      * fixed pylint
      
      * updated API docs
      
      * fixed grammar
      
      * set y ticks to int value in more sufficient way
      611cf5d4
  24. 22 Apr, 2019 1 commit
  25. 19 Apr, 2019 1 commit
  26. 16 Apr, 2019 1 commit
    • kenmatsu4's avatar
      [python] add flag of displaying train loss for lgb.cv() (#2089) · ca85b679
      kenmatsu4 authored
      * [python] displaying train loss during training with lgb.cv
      
      * modifying only display running type when disp_train_loss==True
      
      * Add test for display train loss
      
      * del .idea files
      
      * Rename disp_train_loss to show_train_loss and revise comment.
      
      * Change aug name show_train_loss -> eval_train_metric , and add a test item.
      
      * Modifying comment of eval_train_metric.
      ca85b679
  27. 13 Apr, 2019 1 commit
  28. 04 Apr, 2019 1 commit
    • remcob-gr's avatar
      Add Cost Effective Gradient Boosting (#2014) · 76102284
      remcob-gr authored
      * Add configuration parameters for CEGB.
      
      * Add skeleton CEGB tree learner
      
      Like the original CEGB version, this inherits from SerialTreeLearner.
      Currently, it changes nothing from the original.
      
      * Track features used in CEGB tree learner.
      
      * Pull CEGB tradeoff and coupled feature penalty from config.
      
      * Implement finding best splits for CEGB
      
      This is heavily based on the serial version, but just adds using the coupled penalties.
      
      * Set proper defaults for cegb parameters.
      
      * Ensure sanity checks don't switch off CEGB.
      
      * Implement per-data-point feature penalties in CEGB.
      
      * Implement split penalty and remove unused parameters.
      
      * Merge changes from CEGB tree learner into serial tree learner
      
      * Represent features_used_in_data by a bitset, to reduce the memory overhead of CEGB, and add sanity checks for the lengths of the penalty vectors.
      
      * Fix bug where CEGB would incorrectly penalise a previously used feature
      
      The tree learner did not update the gains of previously computed leaf splits when splitting a leaf elsewhere in the tree.
      This caused it to prefer new features due to incorrectly penalising splitting on previously used features.
      
      * Document CEGB parameters and add them to the appropriate section.
      
      * Remove leftover reference to cegb tree learner.
      
      * Remove outdated diff.
      
      * Fix warnings
      
      * Fix minor issues identified by @StrikerRUS.
      
      * Add docs section on CEGB, including citation.
      
      * Fix link.
      
      * Fix CI failure.
      
      * Add some unit tests
      
      * Fix pylint issues.
      
      * Fix remaining pylint issue
      76102284
  29. 25 Mar, 2019 1 commit
    • kenmatsu4's avatar
      [python] Use first_metric_only flag for early_stopping function. (#2049) · 011cc90a
      kenmatsu4 authored
      * Use first_metric_only flag for early_stopping function.
      
      In order to apply early stopping with only first metric, applying first_metric_only flag for early_stopping function.
      
      * upcate comment
      
      * Revert "upcate comment"
      
      This reverts commit 1e75a1a415cc16cfbe795181e148ebfe91469be4.
      
      * added test
      
      * fixed docstring
      
      * cut comment and save one line
      
      * document new feature
      011cc90a
  30. 14 Mar, 2019 1 commit
  31. 09 Mar, 2019 1 commit
  32. 07 Mar, 2019 1 commit
  33. 26 Feb, 2019 1 commit
    • remcob-gr's avatar
      Add ability to move features from one data set to another in memory (#2006) · 219c943d
      remcob-gr authored
      * Initial attempt to implement appending features in-memory to another data set
      
      The intent is for this to enable munging files together easily, without needing to round-trip via numpy or write multiple copies to disk.
      In turn, that enables working more efficiently with data sets that were written separately.
      
      * Implement Dataset.dump_text, and fix small bug in appending of group bin boundaries.
      
      Dumping to text enables us to compare results, without having to worry about issues like features being reordered.
      
      * Add basic tests for validation logic for add_features_from.
      
      * Remove various internal mapping items from dataset text dumps
      
      These are too sensitive to the exact feature order chosen, which is not visible to the user.
      Including them in tests appears unnecessary, as the data dumping code should provide enough coverage.
      
      * Add test that add_features_from results in identical data sets according to dump_text.
      
      * Add test that booster behaviour after using add_features_from matches that of training on the full data
      
      This checks:
      - That training after add_features_from works at all
      - That add_features_from does not cause training to misbehave
      
      * Expose feature_penalty and monotone_types/constraints via get_field
      
      These getters allow us to check that add_features_from does the right thing with these vectors.
      
      * Add tests that add_features correctly handles feature_penalty and monotone_constraints.
      
      * Ensure add_features_from properly frees the added dataset and add unit test for this
      
      Since add_features_from moves the feature group pointers from the added dataset to the dataset being added to, the added dataset is invalid after the call.
      We must ensure we do not try and access this handle.
      
      * Remove some obsolete TODOs
      
      * Tidy up DumpTextFile by using a single iterator for each feature
      
      This iterators were also passed around as raw pointers without being freed, which is now fixed.
      
      * Factor out offsetting logic in AddFeaturesFrom
      
      * Remove obsolete TODO
      
      * Remove another TODO
      
      This one is debatable, test code can be a bit messy and duplicate-heavy, factoring it out tends to end badly.
      Leaving this for now, will revisit if adding more tests later on becomes a mess.
      
      * Add documentation for newly-added methods.
      
      * Fix whitespace issues identified by pylint.
      
      * Fix a few more whitespace issues.
      
      * Fix doc comments
      
      * Implement deep copying for feature groups.
      
      * Replace awkward std::move usage by emplace_back, and reduce vector size to num_features rather than num_total_features.
      
      * Copy feature groups in addFeaturesFrom, rather than moving them.
      
      * Fix bugs in FeatureGroup copy constructor and ensure source dataset remains usable
      
      * Add reserve to PushVector and PushOffset
      
      * Move definition of Clone into class body
      
      * Fix PR review issues
      
      * Fix for loop increment style.
      
      * Fix test failure
      
      * Some more docstring fixes.
      
      * Remove blank line
      219c943d
  34. 02 Feb, 2019 1 commit
  35. 30 Jan, 2019 2 commits
  36. 27 Jan, 2019 1 commit
    • Nikita Titov's avatar
      [tests][python] added tests for metrics' behavior and fixed case for... · f9a1465d
      Nikita Titov authored
      [tests][python] added tests for metrics' behavior and fixed case for multiclass task with custom objective (#1954)
      
      * added metrics test for standard interface
      
      * simplified code
      
      * less trees
      
      * less trees
      
      * use dummy custom objective and metric
      
      * added tests for multiclass metrics aliases
      
      * fixed bug in case of custom obj and num_class > 1
      
      * added metric test for sklearn wrapper
      f9a1465d
  37. 20 Dec, 2018 2 commits