README.md 2.88 KB
Newer Older
Guolin Ke's avatar
Guolin Ke committed
1
LightGBM, Light Gradient Boosting Machine
wxchan's avatar
wxchan committed
2
=========================================
Guolin Ke's avatar
Guolin Ke committed
3
[![Build Status](https://travis-ci.org/Microsoft/LightGBM.svg?branch=master)](https://travis-ci.org/Microsoft/LightGBM)
Guolin Ke's avatar
Guolin Ke committed
4

5
LightGBM is a gradient boosting framework that uses tree based learning algorithms. It is designed to be distributed and efficient with the following advantages:
Guolin Ke's avatar
Guolin Ke committed
6

7
- Faster training speed and higher efficiency
xuehui's avatar
xuehui committed
8
- Lower memory usage
Guolin Ke's avatar
Guolin Ke committed
9
- Better accuracy
Guolin Ke's avatar
Guolin Ke committed
10
- Parallel learning supported
11
- Capable of handling large-scale data
Guolin Ke's avatar
Guolin Ke committed
12

xuehui's avatar
xuehui committed
13
For more details, please refer to [Features](https://github.com/Microsoft/LightGBM/wiki/Features).
Guolin Ke's avatar
Guolin Ke committed
14

15
[Experiments](https://github.com/Microsoft/LightGBM/wiki/Experiments#comparison-experiment) on public datasets show that LightGBM can outperform other existing boosting framework on both efficiency and accuracy, with significant lower memory consumption. What's more, the [experiments](https://github.com/Microsoft/LightGBM/wiki/Experiments#parallel-experiment) show that LightGBM can achieve a linear speed-up by using multiple machines for training in specific settings.
Guolin Ke's avatar
Guolin Ke committed
16

wxchan's avatar
wxchan committed
17
18
News
----
Guolin Ke's avatar
Guolin Ke committed
19
12/05/2016 : **Categorical Features as input directly**(without one-hot coding). Experiment on [Expo data](http://stat-computing.org/dataexpo/2009/) shows about 8x speed-up with same accuracy compared with one-hot coding (refer to [categorical log]( https://github.com/guolinke/boosting_tree_benchmarks/blob/master/lightgbm/lightgbm_dataexpo_speed.log) and [one-hot log]( https://github.com/guolinke/boosting_tree_benchmarks/blob/master/lightgbm/lightgbm_dataexpo_onehot_speed.log)).
Guolin Ke's avatar
Guolin Ke committed
20
For the setting details, please refer to [IO Parameters](./docs/Parameters.md#io-parameters).
wxchan's avatar
wxchan committed
21

Guolin Ke's avatar
Guolin Ke committed
22
12/02/2016 : Release [**python-package**](./python-package) beta version, welcome to have a try and provide issues and feedback.
wxchan's avatar
wxchan committed
23

Guolin Ke's avatar
Guolin Ke committed
24
25
Get Started
------------
26
To get started, please follow the [Installation Guide](https://github.com/Microsoft/LightGBM/wiki/Installation-Guide) and [Quick Start](https://github.com/Microsoft/LightGBM/wiki/Quick-Start).
Guolin Ke's avatar
Guolin Ke committed
27
28
29

Documents
------------
Guolin Ke's avatar
Guolin Ke committed
30
* [**Wiki**](https://github.com/Microsoft/LightGBM/wiki)
31
* [**Installation Guide**](https://github.com/Microsoft/LightGBM/wiki/Installation-Guide)
Guolin Ke's avatar
Guolin Ke committed
32
* [**Quick Start**](https://github.com/Microsoft/LightGBM/wiki/Quick-Start)
Guolin Ke's avatar
Guolin Ke committed
33
* [**Examples**](./examples)
34
35
36
* [**Features**](https://github.com/Microsoft/LightGBM/wiki/Features)
* [**Parallel Learning Guide**](https://github.com/Microsoft/LightGBM/wiki/Parallel-Learning-Guide)
* [**Configuration**](https://github.com/Microsoft/LightGBM/wiki/Configuration)
Guolin Ke's avatar
Guolin Ke committed
37

Guolin Ke's avatar
Guolin Ke committed
38
39
40
Microsoft Open Source Code of Conduct
------------
This project has adopted the [Microsoft Open Source Code of Conduct](https://opensource.microsoft.com/codeofconduct/). For more information see the [Code of Conduct FAQ](https://opensource.microsoft.com/codeofconduct/faq/) or contact [opencode@microsoft.com](mailto:opencode@microsoft.com) with any additional questions or comments.