Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
tianlh
LightGBM-DCU
Commits
819df012
Commit
819df012
authored
Jan 01, 2018
by
Darío Hereñú
Committed by
Nikita Titov
Jan 01, 2018
Browse files
[docs] Typo on #119 (#1166)
parent
11d8d09e
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
3 additions
and
3 deletions
+3
-3
docs/Python-Intro.rst
docs/Python-Intro.rst
+3
-3
No files found.
docs/Python-Intro.rst
View file @
819df012
...
@@ -34,7 +34,7 @@ Data Interface
...
@@ -34,7 +34,7 @@ Data Interface
The LightGBM Python module is able to load data from:
The LightGBM Python module is able to load data from:
- libsvm/tsv/csv
txt format file
- libsvm/tsv/csv
/
txt format file
- Numpy 2D array, pandas object
- Numpy 2D array, pandas object
...
@@ -91,7 +91,7 @@ In LightGBM, the validation data should be aligned with training data.
...
@@ -91,7 +91,7 @@ In LightGBM, the validation data should be aligned with training data.
train_data = lgb.Dataset(data, label=label, feature_name=['c1', 'c2', 'c3'], categorical_feature=['c3'])
train_data = lgb.Dataset(data, label=label, feature_name=['c1', 'c2', 'c3'], categorical_feature=['c3'])
LightGBM can use categorical features as input directly.
LightGBM can use categorical features as input directly.
It doesn't need to covert to one-hot coding, and is much faster than one-hot coding (about 8x speed-up).
It doesn't need to co
n
vert to one-hot coding, and is much faster than one-hot coding (about 8x speed-up).
**Note**: You should convert your categorical features to ``int`` type before you construct ``Dataset``.
**Note**: You should convert your categorical features to ``int`` type before you construct ``Dataset``.
...
@@ -116,7 +116,7 @@ And you can use ``Dataset.set_init_score()`` to set initial score, and ``Dataset
...
@@ -116,7 +116,7 @@ And you can use ``Dataset.set_init_score()`` to set initial score, and ``Dataset
The ``Dataset`` object in LightGBM is very memory-efficient, due to it only need to save discrete bins.
The ``Dataset`` object in LightGBM is very memory-efficient, due to it only need to save discrete bins.
However, Numpy/Array/Pandas object is memory cost.
However, Numpy/Array/Pandas object is memory cost.
If you concern about your memory consumption, you can save memory acc
r
oding to following:
If you concern about your memory consumption, you can save memory acco
r
ding to following:
1. Let ``free_raw_data=True`` (default is ``True``) when constructing the ``Dataset``
1. Let ``free_raw_data=True`` (default is ``True``) when constructing the ``Dataset``
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment