Unverified Commit b3c31c40 authored by Nikita Titov's avatar Nikita Titov Committed by GitHub
Browse files

[docs] Python wrapper doesn't support params in form of list of pairs (#2078)

* fixed Python intro

* fixed typos

* scikit-learn added support of https
parent 691b8428
...@@ -69,7 +69,7 @@ Optimization in Network Communication ...@@ -69,7 +69,7 @@ Optimization in Network Communication
------------------------------------- -------------------------------------
It only needs to use some collective communication algorithms, like "All reduce", "All gather" and "Reduce scatter", in parallel learning of LightGBM. It only needs to use some collective communication algorithms, like "All reduce", "All gather" and "Reduce scatter", in parallel learning of LightGBM.
LightGBM implement state-of-art algorithms\ `[9] <#references>`__. LightGBM implements state-of-art algorithms\ `[9] <#references>`__.
These collective communication algorithms can provide much better performance than point-to-point communication. These collective communication algorithms can provide much better performance than point-to-point communication.
Optimization in Parallel Learning Optimization in Parallel Learning
...@@ -147,7 +147,7 @@ Data Parallel in LightGBM ...@@ -147,7 +147,7 @@ Data Parallel in LightGBM
We reduce communication cost of data parallel in LightGBM: We reduce communication cost of data parallel in LightGBM:
1. Instead of "Merge global histograms from all local histograms", LightGBM use "Reduce Scatter" to merge histograms of different (non-overlapping) features for different workers. 1. Instead of "Merge global histograms from all local histograms", LightGBM uses "Reduce Scatter" to merge histograms of different (non-overlapping) features for different workers.
Then workers find the local best split on local merged histograms and sync up the global best split. Then workers find the local best split on local merged histograms and sync up the global best split.
2. As aforementioned, LightGBM uses histogram subtraction to speed up training. 2. As aforementioned, LightGBM uses histogram subtraction to speed up training.
...@@ -244,9 +244,9 @@ Other Features ...@@ -244,9 +244,9 @@ Other Features
- Validation metric output during training - Validation metric output during training
- Multi validation data - Multiple validation data
- Multi metrics - Multiple metrics
- Early stopping (both training and prediction) - Early stopping (both training and prediction)
......
...@@ -134,14 +134,14 @@ If you are concerned about your memory consumption, you can save memory by: ...@@ -134,14 +134,14 @@ If you are concerned about your memory consumption, you can save memory by:
Setting Parameters Setting Parameters
------------------ ------------------
LightGBM can use either a list of pairs or a dictionary to set `Parameters <./Parameters.rst>`__. LightGBM can use a dictionary to set `Parameters <./Parameters.rst>`__.
For instance: For instance:
- Booster parameters: - Booster parameters:
.. code:: python .. code:: python
param = {'num_leaves':31, 'num_trees':100, 'objective':'binary'} param = {'num_leaves': 31, 'num_trees': 100, 'objective': 'binary'}
param['metric'] = 'auc' param['metric'] = 'auc'
- You can also specify multiple eval metrics: - You can also specify multiple eval metrics:
...@@ -176,7 +176,7 @@ A saved model can be loaded: ...@@ -176,7 +176,7 @@ A saved model can be loaded:
.. code:: python .. code:: python
bst = lgb.Booster(model_file='model.txt') #init model bst = lgb.Booster(model_file='model.txt') # init model
CV CV
-- --
......
...@@ -353,7 +353,7 @@ def cv(params, train_set, num_boost_round=100, ...@@ -353,7 +353,7 @@ def cv(params, train_set, num_boost_round=100,
folds : generator or iterator of (train_idx, test_idx) tuples, scikit-learn splitter object or None, optional (default=None) folds : generator or iterator of (train_idx, test_idx) tuples, scikit-learn splitter object or None, optional (default=None)
If generator or iterator, it should yield the train and test indices for each fold. If generator or iterator, it should yield the train and test indices for each fold.
If object, it should be one of the scikit-learn splitter classes If object, it should be one of the scikit-learn splitter classes
(http://scikit-learn.org/stable/modules/classes.html#splitter-classes) (https://scikit-learn.org/stable/modules/classes.html#splitter-classes)
and have ``split`` method. and have ``split`` method.
This argument has highest priority over other data split arguments. This argument has highest priority over other data split arguments.
nfold : int, optional (default=5) nfold : int, optional (default=5)
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment