gbdt.cpp 29.3 KB
Newer Older
Guolin Ke's avatar
Guolin Ke committed
1
2
#include "gbdt.h"

3
#include <LightGBM/utils/openmp_wrapper.h>
4

Guolin Ke's avatar
Guolin Ke committed
5
6
7
8
9
10
11
12
13
14
15
#include <LightGBM/utils/common.h>

#include <LightGBM/objective_function.h>
#include <LightGBM/metric.h>

#include <ctime>

#include <sstream>
#include <chrono>
#include <string>
#include <vector>
16
#include <utility>
Guolin Ke's avatar
Guolin Ke committed
17
18
19

namespace LightGBM {

Guolin Ke's avatar
Guolin Ke committed
20
21
22
#ifdef TIMETAG
std::chrono::duration<double, std::milli> boosting_time;
std::chrono::duration<double, std::milli> train_score_time;
Guolin Ke's avatar
Guolin Ke committed
23
std::chrono::duration<double, std::milli> out_of_bag_score_time;
Guolin Ke's avatar
Guolin Ke committed
24
25
26
27
28
29
30
std::chrono::duration<double, std::milli> valid_score_time;
std::chrono::duration<double, std::milli> metric_time;
std::chrono::duration<double, std::milli> bagging_time;
std::chrono::duration<double, std::milli> sub_gradient_time;
std::chrono::duration<double, std::milli> tree_time;
#endif // TIMETAG

31
GBDT::GBDT()
32
  :iter_(0),
33
34
35
36
37
  train_data_(nullptr),
  object_function_(nullptr),
  early_stopping_round_(0),
  max_feature_idx_(0),
  num_class_(1),
38
  sigmoid_(-1.0f),
39
  num_iteration_for_pred_(0),
40
  shrinkage_rate_(0.1f),
41
42
  num_init_iteration_(0),
  boost_from_average_(false) {
43
44
45
46
47
#pragma omp parallel
#pragma omp master
    {
      num_threads_ = omp_get_num_threads();
    }
Guolin Ke's avatar
Guolin Ke committed
48
49
50
}

GBDT::~GBDT() {
Guolin Ke's avatar
Guolin Ke committed
51
52
53
#ifdef TIMETAG
  Log::Info("GBDT::boosting costs %f", boosting_time * 1e-3);
  Log::Info("GBDT::train_score costs %f", train_score_time * 1e-3);
Guolin Ke's avatar
Guolin Ke committed
54
  Log::Info("GBDT::out_of_bag_score costs %f", out_of_bag_score_time * 1e-3);
Guolin Ke's avatar
Guolin Ke committed
55
56
57
58
59
60
  Log::Info("GBDT::valid_score costs %f", valid_score_time * 1e-3);
  Log::Info("GBDT::metric costs %f", metric_time * 1e-3);
  Log::Info("GBDT::bagging costs %f", bagging_time * 1e-3);
  Log::Info("GBDT::sub_gradient costs %f", sub_gradient_time * 1e-3);
  Log::Info("GBDT::tree costs %f", tree_time * 1e-3);
#endif
Guolin Ke's avatar
Guolin Ke committed
61
62
}

63
void GBDT::Init(const BoostingConfig* config, const Dataset* train_data, const ObjectiveFunction* object_function,
64
                const std::vector<const Metric*>& training_metrics) {
65
  iter_ = 0;
wxchan's avatar
wxchan committed
66
  num_iteration_for_pred_ = 0;
67
  max_feature_idx_ = 0;
wxchan's avatar
wxchan committed
68
69
  num_class_ = config->num_class;
  train_data_ = nullptr;
Guolin Ke's avatar
Guolin Ke committed
70
  gbdt_config_ = nullptr;
71
  tree_learner_ = nullptr;
wxchan's avatar
wxchan committed
72
73
74
75
  ResetTrainingData(config, train_data, object_function, training_metrics);
}

void GBDT::ResetTrainingData(const BoostingConfig* config, const Dataset* train_data, const ObjectiveFunction* object_function,
76
                             const std::vector<const Metric*>& training_metrics) {
Guolin Ke's avatar
Guolin Ke committed
77
  auto new_config = std::unique_ptr<BoostingConfig>(new BoostingConfig(*config));
wxchan's avatar
wxchan committed
78
79
80
  if (train_data_ != nullptr && !train_data_->CheckAlign(*train_data)) {
    Log::Fatal("cannot reset training data, since new training data has different bin mappers");
  }
Guolin Ke's avatar
Guolin Ke committed
81
82
83
  early_stopping_round_ = new_config->early_stopping_round;
  shrinkage_rate_ = new_config->learning_rate;

Guolin Ke's avatar
Guolin Ke committed
84
  object_function_ = object_function;
Guolin Ke's avatar
Guolin Ke committed
85

Guolin Ke's avatar
Guolin Ke committed
86
  sigmoid_ = -1.0f;
wxchan's avatar
wxchan committed
87
  if (object_function_ != nullptr
88
      && std::string(object_function_->GetName()) == std::string("binary")) {
Guolin Ke's avatar
Guolin Ke committed
89
    // only binary classification need sigmoid transform
Guolin Ke's avatar
Guolin Ke committed
90
    sigmoid_ = new_config->sigmoid;
91
  }
Guolin Ke's avatar
Guolin Ke committed
92

Guolin Ke's avatar
Guolin Ke committed
93
  if (train_data_ != train_data && train_data != nullptr) {
94
95
    if (tree_learner_ == nullptr) {
      tree_learner_ = std::unique_ptr<TreeLearner>(TreeLearner::CreateTreeLearner(new_config->tree_learner_type, &new_config->tree_config));
Guolin Ke's avatar
Guolin Ke committed
96
97
    }
    // init tree learner
98
    tree_learner_->Init(train_data);
Guolin Ke's avatar
Guolin Ke committed
99

Guolin Ke's avatar
Guolin Ke committed
100
101
102
103
104
105
    // push training metrics
    training_metrics_.clear();
    for (const auto& metric : training_metrics) {
      training_metrics_.push_back(metric);
    }
    training_metrics_.shrink_to_fit();
wxchan's avatar
wxchan committed
106
107
108
109
110
111
112
113
114
115
116
117
118
    // not same training data, need reset score and others
    // create score tracker
    train_score_updater_.reset(new ScoreUpdater(train_data, num_class_));
    // update score
    for (int i = 0; i < iter_; ++i) {
      for (int curr_class = 0; curr_class < num_class_; ++curr_class) {
        auto curr_tree = (i + num_init_iteration_) * num_class_ + curr_class;
        train_score_updater_->AddScore(models_[curr_tree].get(), curr_class);
      }
    }
    num_data_ = train_data->num_data();
    // create buffer for gradients and hessians
    if (object_function_ != nullptr) {
119
120
121
      size_t total_size = static_cast<size_t>(num_data_) * num_class_;
      gradients_.resize(total_size);
      hessians_.resize(total_size);
wxchan's avatar
wxchan committed
122
123
124
125
126
    }
    // get max feature index
    max_feature_idx_ = train_data->num_total_features() - 1;
    // get label index
    label_idx_ = train_data->label_idx();
127
128
    // get feature names
    feature_names_ = train_data->feature_names();
Guolin Ke's avatar
Guolin Ke committed
129
130

    feature_infos_ = train_data->feature_infos();
Guolin Ke's avatar
Guolin Ke committed
131
132
  }

Guolin Ke's avatar
Guolin Ke committed
133
  if ((train_data_ != train_data && train_data != nullptr)
134
      || (gbdt_config_ != nullptr && gbdt_config_->bagging_fraction != new_config->bagging_fraction)) {
wxchan's avatar
wxchan committed
135
    // if need bagging, create buffer
Guolin Ke's avatar
Guolin Ke committed
136
    if (new_config->bagging_fraction < 1.0 && new_config->bagging_freq > 0) {
137
138
      bag_data_cnt_ =
        static_cast<data_size_t>(new_config->bagging_fraction * num_data_);
139
      bag_data_indices_.resize(num_data_);
140
141
142
143
144
145
      tmp_indices_.resize(num_data_);
      offsets_buf_.resize(num_threads_);
      left_cnts_buf_.resize(num_threads_);
      right_cnts_buf_.resize(num_threads_);
      left_write_pos_buf_.resize(num_threads_);
      right_write_pos_buf_.resize(num_threads_);
Guolin Ke's avatar
Guolin Ke committed
146
147
      double average_bag_rate = new_config->bagging_fraction / new_config->bagging_freq;
      is_use_subset_ = false;
148
      if (average_bag_rate <= 0.5) {
Guolin Ke's avatar
Guolin Ke committed
149
        tmp_subset_.reset(new Dataset(bag_data_cnt_));
150
        tmp_subset_->CopyFeatureMapperFrom(train_data);
Guolin Ke's avatar
Guolin Ke committed
151
152
153
        is_use_subset_ = true;
        Log::Debug("use subset for bagging");
      }
wxchan's avatar
wxchan committed
154
155
156
    } else {
      bag_data_cnt_ = num_data_;
      bag_data_indices_.clear();
157
      tmp_indices_.clear();
Guolin Ke's avatar
Guolin Ke committed
158
      is_use_subset_ = false;
wxchan's avatar
wxchan committed
159
    }
Guolin Ke's avatar
Guolin Ke committed
160
  }
wxchan's avatar
wxchan committed
161
  train_data_ = train_data;
Guolin Ke's avatar
Guolin Ke committed
162
163
  if (train_data_ != nullptr) {
    // reset config for tree learner
164
    tree_learner_->ResetConfig(&new_config->tree_config);
Guolin Ke's avatar
Guolin Ke committed
165
  }
Guolin Ke's avatar
Guolin Ke committed
166
  gbdt_config_.reset(new_config.release());
Guolin Ke's avatar
Guolin Ke committed
167
168
}

wxchan's avatar
wxchan committed
169
void GBDT::AddValidDataset(const Dataset* valid_data,
170
                           const std::vector<const Metric*>& valid_metrics) {
wxchan's avatar
wxchan committed
171
172
  if (!train_data_->CheckAlign(*valid_data)) {
    Log::Fatal("cannot add validation data, since it has different bin mappers with training data");
173
  }
Guolin Ke's avatar
Guolin Ke committed
174
  // for a validation dataset, we need its score and metric
Guolin Ke's avatar
Guolin Ke committed
175
  auto new_score_updater = std::unique_ptr<ScoreUpdater>(new ScoreUpdater(valid_data, num_class_));
wxchan's avatar
wxchan committed
176
177
178
179
180
181
182
  // update score
  for (int i = 0; i < iter_; ++i) {
    for (int curr_class = 0; curr_class < num_class_; ++curr_class) {
      auto curr_tree = (i + num_init_iteration_) * num_class_ + curr_class;
      new_score_updater->AddScore(models_[curr_tree].get(), curr_class);
    }
  }
Guolin Ke's avatar
Guolin Ke committed
183
  valid_score_updater_.push_back(std::move(new_score_updater));
Guolin Ke's avatar
Guolin Ke committed
184
  valid_metrics_.emplace_back();
185
186
187
  if (early_stopping_round_ > 0) {
    best_iter_.emplace_back();
    best_score_.emplace_back();
Guolin Ke's avatar
Guolin Ke committed
188
    best_msg_.emplace_back();
189
  }
Guolin Ke's avatar
Guolin Ke committed
190
191
  for (const auto& metric : valid_metrics) {
    valid_metrics_.back().push_back(metric);
192
193
194
    if (early_stopping_round_ > 0) {
      best_iter_.back().push_back(0);
      best_score_.back().push_back(kMinScore);
Guolin Ke's avatar
Guolin Ke committed
195
      best_msg_.back().emplace_back();
196
    }
Guolin Ke's avatar
Guolin Ke committed
197
  }
Guolin Ke's avatar
Guolin Ke committed
198
  valid_metrics_.back().shrink_to_fit();
Guolin Ke's avatar
Guolin Ke committed
199
200
}

201
data_size_t GBDT::BaggingHelper(Random& cur_rand, data_size_t start, data_size_t cnt, data_size_t* buffer) {
202
203
204
  if (cnt <= 0) {
    return 0;
  }
205
206
207
208
  data_size_t bag_data_cnt =
    static_cast<data_size_t>(gbdt_config_->bagging_fraction * cnt);
  data_size_t cur_left_cnt = 0;
  data_size_t cur_right_cnt = 0;
Guolin Ke's avatar
Guolin Ke committed
209
  auto right_buffer = buffer + bag_data_cnt;
210
211
  // random bagging, minimal unit is one record
  for (data_size_t i = 0; i < cnt; ++i) {
Guolin Ke's avatar
Guolin Ke committed
212
213
214
    float prob =
      (bag_data_cnt - cur_left_cnt) / static_cast<float>(cnt - i);
    if (cur_rand.NextFloat() < prob) {
215
216
      buffer[cur_left_cnt++] = start + i;
    } else {
Guolin Ke's avatar
Guolin Ke committed
217
      right_buffer[cur_right_cnt++] = start + i;
218
219
220
221
222
    }
  }
  CHECK(cur_left_cnt == bag_data_cnt);
  return cur_left_cnt;
}
Guolin Ke's avatar
Guolin Ke committed
223

Guolin Ke's avatar
Guolin Ke committed
224
225


226
void GBDT::Bagging(int iter) {
Guolin Ke's avatar
Guolin Ke committed
227
  // if need bagging
228
  if (bag_data_cnt_ < num_data_ && iter % gbdt_config_->bagging_freq == 0) {
Guolin Ke's avatar
Guolin Ke committed
229
    const data_size_t min_inner_size = 1000;
230
231
232
    data_size_t inner_size = (num_data_ + num_threads_ - 1) / num_threads_;
    if (inner_size < min_inner_size) { inner_size = min_inner_size; }

233
  #pragma omp parallel for schedule(static,1)
234
235
236
237
238
239
240
    for (int i = 0; i < num_threads_; ++i) {
      left_cnts_buf_[i] = 0;
      right_cnts_buf_[i] = 0;
      data_size_t cur_start = i * inner_size;
      if (cur_start > num_data_) { continue; }
      data_size_t cur_cnt = inner_size;
      if (cur_start + cur_cnt > num_data_) { cur_cnt = num_data_ - cur_start; }
Guolin Ke's avatar
Guolin Ke committed
241
242
      Random cur_rand(gbdt_config_->bagging_seed + iter * num_threads_ + i);
      data_size_t cur_left_count = BaggingHelper(cur_rand, cur_start, cur_cnt, tmp_indices_.data() + cur_start);
243
244
245
246
247
248
249
250
251
252
253
254
255
      offsets_buf_[i] = cur_start;
      left_cnts_buf_[i] = cur_left_count;
      right_cnts_buf_[i] = cur_cnt - cur_left_count;
    }
    data_size_t left_cnt = 0;
    left_write_pos_buf_[0] = 0;
    right_write_pos_buf_[0] = 0;
    for (int i = 1; i < num_threads_; ++i) {
      left_write_pos_buf_[i] = left_write_pos_buf_[i - 1] + left_cnts_buf_[i - 1];
      right_write_pos_buf_[i] = right_write_pos_buf_[i - 1] + right_cnts_buf_[i - 1];
    }
    left_cnt = left_write_pos_buf_[num_threads_ - 1] + left_cnts_buf_[num_threads_ - 1];

256
  #pragma omp parallel for schedule(static, 1)
257
258
259
    for (int i = 0; i < num_threads_; ++i) {
      if (left_cnts_buf_[i] > 0) {
        std::memcpy(bag_data_indices_.data() + left_write_pos_buf_[i],
260
                    tmp_indices_.data() + offsets_buf_[i], left_cnts_buf_[i] * sizeof(data_size_t));
Guolin Ke's avatar
Guolin Ke committed
261
      }
262
263
      if (right_cnts_buf_[i] > 0) {
        std::memcpy(bag_data_indices_.data() + left_cnt + right_write_pos_buf_[i],
264
                    tmp_indices_.data() + offsets_buf_[i] + left_cnts_buf_[i], right_cnts_buf_[i] * sizeof(data_size_t));
Guolin Ke's avatar
Guolin Ke committed
265
266
      }
    }
Guolin Ke's avatar
Guolin Ke committed
267
268
    bag_data_cnt_ = left_cnt;
    CHECK(bag_data_indices_[bag_data_cnt_ - 1] > bag_data_indices_[bag_data_cnt_]);
Guolin Ke's avatar
Guolin Ke committed
269
    Log::Debug("Re-bagging, using %d data to train", bag_data_cnt_);
Guolin Ke's avatar
Guolin Ke committed
270
    // set bagging data to tree learner
Guolin Ke's avatar
Guolin Ke committed
271
272
273
274
    if (!is_use_subset_) {
      tree_learner_->SetBaggingData(bag_data_indices_.data(), bag_data_cnt_);
    } else {
      // get subset
Guolin Ke's avatar
Guolin Ke committed
275
276
      tmp_subset_->ReSize(bag_data_cnt_);
      tmp_subset_->CopySubset(train_data_, bag_data_indices_.data(), bag_data_cnt_, false);
Guolin Ke's avatar
Guolin Ke committed
277
278
      tree_learner_->ResetTrainingData(tmp_subset_.get());
    }
Guolin Ke's avatar
Guolin Ke committed
279
280
281
  }
}

282
void GBDT::UpdateScoreOutOfBag(const Tree* tree, const int curr_class) {
Guolin Ke's avatar
Guolin Ke committed
283
284
285
#ifdef TIMETAG
  auto start_time = std::chrono::steady_clock::now();
#endif
286
  // we need to predict out-of-bag scores of data for boosting
Guolin Ke's avatar
Guolin Ke committed
287
  if (num_data_ - bag_data_cnt_ > 0 && !is_use_subset_) {
288
    train_score_updater_->AddScore(tree, bag_data_indices_.data() + bag_data_cnt_, num_data_ - bag_data_cnt_, curr_class);
Guolin Ke's avatar
Guolin Ke committed
289
  }
Guolin Ke's avatar
Guolin Ke committed
290
#ifdef TIMETAG
Guolin Ke's avatar
Guolin Ke committed
291
  out_of_bag_score_time += std::chrono::steady_clock::now() - start_time;
Guolin Ke's avatar
Guolin Ke committed
292
#endif
Guolin Ke's avatar
Guolin Ke committed
293
294
}

295
bool GBDT::TrainOneIter(const score_t* gradient, const score_t* hessian, bool is_eval) {
296
297
298
299
300
301
302
303
  // boosting from average prediction.
  if (models_.empty() && gbdt_config_->boost_from_average && !train_score_updater_->has_init_score()) {
    std::vector<double> sum_per_class(num_class_, 0.0f);
    auto label = train_data_->metadata().label();
    if (num_class_ > 1) {
      for (data_size_t i = 0; i < num_data_; ++i) {
        sum_per_class[static_cast<int>(label[i])] += 1.0f;
      }
304
    } else if(sigmoid_ < 0.0f){
305
306
307
      for (data_size_t i = 0; i < num_data_; ++i) {
        sum_per_class[0] += label[i];
      }
308
309
310
311
    } else {
      for (data_size_t i = 0; i < num_data_; ++i) {
        sum_per_class[0] += label[i] > 0;
      }
312
    }
313
314
315
316
    std::vector<double > init_scores(num_class_);
    for (int i = 0; i < num_class_; ++i) {
      init_scores[i] = sum_per_class[i] / num_data_;
    }
317
318
319
    if (object_function_ != nullptr) {
      init_scores = object_function_->ConvertToRawScore(init_scores);
    }
320
321
    for (int curr_class = 0; curr_class < num_class_; ++curr_class) {
      std::unique_ptr<Tree> new_tree(new Tree(2));
322
323
      new_tree->Split(0, 0, BinType::NumericalBin, 0, 0, 0, init_scores[curr_class], init_scores[curr_class], 0, num_data_, 1);
      train_score_updater_->AddScore(init_scores[curr_class], curr_class);
324
      for (auto& score_updater : valid_score_updater_) {
325
        score_updater->AddScore(init_scores[curr_class], curr_class);
326
327
328
329
330
      }
      models_.push_back(std::move(new_tree));
    }
    boost_from_average_ = true;
  }
Guolin Ke's avatar
Guolin Ke committed
331
332
  // boosting first
  if (gradient == nullptr || hessian == nullptr) {
333
  #ifdef TIMETAG
Guolin Ke's avatar
Guolin Ke committed
334
    auto start_time = std::chrono::steady_clock::now();
335
  #endif
Guolin Ke's avatar
Guolin Ke committed
336
337
338
    Boosting();
    gradient = gradients_.data();
    hessian = hessians_.data();
339
  #ifdef TIMETAG
Guolin Ke's avatar
Guolin Ke committed
340
    boosting_time += std::chrono::steady_clock::now() - start_time;
341
  #endif
Guolin Ke's avatar
Guolin Ke committed
342
  }
Guolin Ke's avatar
Guolin Ke committed
343
344
345
#ifdef TIMETAG
  auto start_time = std::chrono::steady_clock::now();
#endif
346
347
  // bagging logic
  Bagging(iter_);
Guolin Ke's avatar
Guolin Ke committed
348
349
350
#ifdef TIMETAG
  bagging_time += std::chrono::steady_clock::now() - start_time;
#endif
Guolin Ke's avatar
Guolin Ke committed
351
  if (is_use_subset_ && bag_data_cnt_ < num_data_) {
352
  #ifdef TIMETAG
Guolin Ke's avatar
Guolin Ke committed
353
    start_time = std::chrono::steady_clock::now();
354
  #endif
Guolin Ke's avatar
Guolin Ke committed
355
356
357
358
    if (gradients_.empty()) {
      size_t total_size = static_cast<size_t>(num_data_) * num_class_;
      gradients_.resize(total_size);
      hessians_.resize(total_size);
359
    }
Guolin Ke's avatar
Guolin Ke committed
360
361
362
    // get sub gradients
    for (int curr_class = 0; curr_class < num_class_; ++curr_class) {
      auto bias = curr_class * num_data_;
363
      // cannot multi-threading
Guolin Ke's avatar
Guolin Ke committed
364
365
366
367
368
369
370
      for (int i = 0; i < bag_data_cnt_; ++i) {
        gradients_[bias + i] = gradient[bias + bag_data_indices_[i]];
        hessians_[bias + i] = hessian[bias + bag_data_indices_[i]];
      }
    }
    gradient = gradients_.data();
    hessian = hessians_.data();
371
  #ifdef TIMETAG
Guolin Ke's avatar
Guolin Ke committed
372
    sub_gradient_time += std::chrono::steady_clock::now() - start_time;
373
  #endif
Guolin Ke's avatar
Guolin Ke committed
374
  }
Guolin Ke's avatar
Guolin Ke committed
375
  bool should_continue = false;
Guolin Ke's avatar
Guolin Ke committed
376
  for (int curr_class = 0; curr_class < num_class_; ++curr_class) {
377
  #ifdef TIMETAG
Guolin Ke's avatar
Guolin Ke committed
378
    start_time = std::chrono::steady_clock::now();
379
  #endif
Guolin Ke's avatar
Guolin Ke committed
380
    std::unique_ptr<Tree> new_tree(new Tree(2));
381
382
    // train a new tree
    new_tree.reset(tree_learner_->Train(gradient + curr_class * num_data_, hessian + curr_class * num_data_));
383
  #ifdef TIMETAG
Guolin Ke's avatar
Guolin Ke committed
384
    tree_time += std::chrono::steady_clock::now() - start_time;
385
  #endif
Guolin Ke's avatar
Guolin Ke committed
386
387

    if (new_tree->num_leaves() > 1) {
Guolin Ke's avatar
Guolin Ke committed
388
389
390
391
392
393
      should_continue = true;
      // shrinkage by learning rate
      new_tree->Shrinkage(shrinkage_rate_);
      // update score
      UpdateScore(new_tree.get(), curr_class);
      UpdateScoreOutOfBag(new_tree.get(), curr_class);
394
    }
Guolin Ke's avatar
Guolin Ke committed
395
396
397
    // add model
    models_.push_back(std::move(new_tree));
  }
Guolin Ke's avatar
Guolin Ke committed
398
  if (!should_continue) {
Guolin Ke's avatar
Guolin Ke committed
399
400
401
402
403
404
    Log::Warning("Stopped training because there are no more leaves that meet the split requirements.");
    for (int curr_class = 0; curr_class < num_class_; ++curr_class) {
      models_.pop_back();
    }
    return true;
  }
Guolin Ke's avatar
Guolin Ke committed
405
406
407
408
409
410
  ++iter_;
  if (is_eval) {
    return EvalAndCheckEarlyStopping();
  } else {
    return false;
  }
411

Guolin Ke's avatar
Guolin Ke committed
412
}
413

wxchan's avatar
wxchan committed
414
void GBDT::RollbackOneIter() {
415
  if (iter_ <= 0) { return; }
wxchan's avatar
wxchan committed
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
  int cur_iter = iter_ + num_init_iteration_ - 1;
  // reset score
  for (int curr_class = 0; curr_class < num_class_; ++curr_class) {
    auto curr_tree = cur_iter * num_class_ + curr_class;
    models_[curr_tree]->Shrinkage(-1.0);
    train_score_updater_->AddScore(models_[curr_tree].get(), curr_class);
    for (auto& score_updater : valid_score_updater_) {
      score_updater->AddScore(models_[curr_tree].get(), curr_class);
    }
  }
  // remove model
  for (int curr_class = 0; curr_class < num_class_; ++curr_class) {
    models_.pop_back();
  }
  --iter_;
}

Guolin Ke's avatar
Guolin Ke committed
433
bool GBDT::EvalAndCheckEarlyStopping() {
434
  bool is_met_early_stopping = false;
Guolin Ke's avatar
Guolin Ke committed
435
436
437
#ifdef TIMETAG
  auto start_time = std::chrono::steady_clock::now();
#endif
438
  // print message for metric
Guolin Ke's avatar
Guolin Ke committed
439
  auto best_msg = OutputMetric(iter_);
Guolin Ke's avatar
Guolin Ke committed
440
441
442
#ifdef TIMETAG
  metric_time += std::chrono::steady_clock::now() - start_time;
#endif
Guolin Ke's avatar
Guolin Ke committed
443
  is_met_early_stopping = !best_msg.empty();
444
445
  if (is_met_early_stopping) {
    Log::Info("Early stopping at iteration %d, the best iteration round is %d",
446
              iter_, iter_ - early_stopping_round_);
Guolin Ke's avatar
Guolin Ke committed
447
    Log::Info("Output of best iteration round:\n%s", best_msg.c_str());
448
    // pop last early_stopping_round_ models
449
    for (int i = 0; i < early_stopping_round_ * num_class_; ++i) {
450
451
452
453
      models_.pop_back();
    }
  }
  return is_met_early_stopping;
Guolin Ke's avatar
Guolin Ke committed
454
455
}

456
void GBDT::UpdateScore(const Tree* tree, const int curr_class) {
Guolin Ke's avatar
Guolin Ke committed
457
458
459
#ifdef TIMETAG
  auto start_time = std::chrono::steady_clock::now();
#endif
Guolin Ke's avatar
Guolin Ke committed
460
  // update training score
Guolin Ke's avatar
Guolin Ke committed
461
462
463
464
465
  if (!is_use_subset_) {
    train_score_updater_->AddScore(tree_learner_.get(), curr_class);
  } else {
    train_score_updater_->AddScore(tree, curr_class);
  }
Guolin Ke's avatar
Guolin Ke committed
466
467
468
469
470
471
#ifdef TIMETAG
  train_score_time += std::chrono::steady_clock::now() - start_time;
#endif
#ifdef TIMETAG
  start_time = std::chrono::steady_clock::now();
#endif
Guolin Ke's avatar
Guolin Ke committed
472
  // update validation score
Guolin Ke's avatar
Guolin Ke committed
473
474
  for (auto& score_updater : valid_score_updater_) {
    score_updater->AddScore(tree, curr_class);
Guolin Ke's avatar
Guolin Ke committed
475
  }
Guolin Ke's avatar
Guolin Ke committed
476
477
478
#ifdef TIMETAG
  valid_score_time += std::chrono::steady_clock::now() - start_time;
#endif
Guolin Ke's avatar
Guolin Ke committed
479
480
}

Guolin Ke's avatar
Guolin Ke committed
481
482
483
484
std::string GBDT::OutputMetric(int iter) {
  bool need_output = (iter % gbdt_config_->output_freq) == 0;
  std::string ret = "";
  std::stringstream msg_buf;
485
  std::vector<std::pair<size_t, size_t>> meet_early_stopping_pairs;
Guolin Ke's avatar
Guolin Ke committed
486
  // print training metric
Guolin Ke's avatar
Guolin Ke committed
487
  if (need_output) {
488
489
490
    for (auto& sub_metric : training_metrics_) {
      auto name = sub_metric->GetName();
      auto scores = sub_metric->Eval(train_score_updater_->score());
Guolin Ke's avatar
Guolin Ke committed
491
      for (size_t k = 0; k < name.size(); ++k) {
Guolin Ke's avatar
Guolin Ke committed
492
493
494
495
496
497
498
499
        std::stringstream tmp_buf;
        tmp_buf << "Iteration:" << iter
          << ", training " << name[k]
          << " : " << scores[k];
        Log::Info(tmp_buf.str().c_str());
        if (early_stopping_round_ > 0) {
          msg_buf << tmp_buf.str() << std::endl;
        }
500
      }
501
    }
Guolin Ke's avatar
Guolin Ke committed
502
503
  }
  // print validation metric
Guolin Ke's avatar
Guolin Ke committed
504
  if (need_output || early_stopping_round_ > 0) {
505
506
507
    for (size_t i = 0; i < valid_metrics_.size(); ++i) {
      for (size_t j = 0; j < valid_metrics_[i].size(); ++j) {
        auto test_scores = valid_metrics_[i][j]->Eval(valid_score_updater_[i]->score());
Guolin Ke's avatar
Guolin Ke committed
508
509
510
511
512
513
514
515
516
517
518
        auto name = valid_metrics_[i][j]->GetName();
        for (size_t k = 0; k < name.size(); ++k) {
          std::stringstream tmp_buf;
          tmp_buf << "Iteration:" << iter
            << ", valid_" << i + 1 << " " << name[k]
            << " : " << test_scores[k];
          if (need_output) {
            Log::Info(tmp_buf.str().c_str());
          }
          if (early_stopping_round_ > 0) {
            msg_buf << tmp_buf.str() << std::endl;
519
          }
wxchan's avatar
wxchan committed
520
        }
Guolin Ke's avatar
Guolin Ke committed
521
        if (ret.empty() && early_stopping_round_ > 0) {
522
523
524
          auto cur_score = valid_metrics_[i][j]->factor_to_bigger_better() * test_scores.back();
          if (cur_score > best_score_[i][j]) {
            best_score_[i][j] = cur_score;
525
            best_iter_[i][j] = iter;
Guolin Ke's avatar
Guolin Ke committed
526
            meet_early_stopping_pairs.emplace_back(i, j);
527
          } else {
Guolin Ke's avatar
Guolin Ke committed
528
            if (iter - best_iter_[i][j] >= early_stopping_round_) { ret = best_msg_[i][j]; }
529
          }
wxchan's avatar
wxchan committed
530
531
        }
      }
Guolin Ke's avatar
Guolin Ke committed
532
533
    }
  }
Guolin Ke's avatar
Guolin Ke committed
534
535
536
  for (auto& pair : meet_early_stopping_pairs) {
    best_msg_[pair.first][pair.second] = msg_buf.str();
  }
wxchan's avatar
wxchan committed
537
  return ret;
Guolin Ke's avatar
Guolin Ke committed
538
539
}

540
/*! \brief Get eval result */
541
std::vector<double> GBDT::GetEvalAt(int data_idx) const {
Guolin Ke's avatar
Guolin Ke committed
542
  CHECK(data_idx >= 0 && data_idx <= static_cast<int>(valid_score_updater_.size()));
543
544
  std::vector<double> ret;
  if (data_idx == 0) {
545
546
    for (auto& sub_metric : training_metrics_) {
      auto scores = sub_metric->Eval(train_score_updater_->score());
547
548
549
      for (auto score : scores) {
        ret.push_back(score);
      }
550
    }
551
  } else {
552
553
554
555
556
557
    auto used_idx = data_idx - 1;
    for (size_t j = 0; j < valid_metrics_[used_idx].size(); ++j) {
      auto test_scores = valid_metrics_[used_idx][j]->Eval(valid_score_updater_[used_idx]->score());
      for (auto score : test_scores) {
        ret.push_back(score);
      }
558
559
560
561
562
    }
  }
  return ret;
}

Guolin Ke's avatar
Guolin Ke committed
563
/*! \brief Get training scores result */
564
const double* GBDT::GetTrainingScore(int64_t* out_len) {
565
  *out_len = static_cast<int64_t>(train_score_updater_->num_data()) * num_class_;
Guolin Ke's avatar
Guolin Ke committed
566
  return train_score_updater_->score();
567
568
}

Guolin Ke's avatar
Guolin Ke committed
569
570
void GBDT::GetPredictAt(int data_idx, double* out_result, int64_t* out_len) {
  CHECK(data_idx >= 0 && data_idx <= static_cast<int>(valid_score_updater_.size()));
Guolin Ke's avatar
Guolin Ke committed
571

572
  const double* raw_scores = nullptr;
Guolin Ke's avatar
Guolin Ke committed
573
574
  data_size_t num_data = 0;
  if (data_idx == 0) {
wxchan's avatar
wxchan committed
575
    raw_scores = GetTrainingScore(out_len);
Guolin Ke's avatar
Guolin Ke committed
576
577
578
579
580
    num_data = train_score_updater_->num_data();
  } else {
    auto used_idx = data_idx - 1;
    raw_scores = valid_score_updater_[used_idx]->score();
    num_data = valid_score_updater_[used_idx]->num_data();
581
    *out_len = static_cast<int64_t>(num_data) * num_class_;
Guolin Ke's avatar
Guolin Ke committed
582
583
  }
  if (num_class_ > 1) {
584
  #pragma omp parallel for schedule(static)
Guolin Ke's avatar
Guolin Ke committed
585
    for (data_size_t i = 0; i < num_data; ++i) {
586
      std::vector<double> tmp_result(num_class_);
Guolin Ke's avatar
Guolin Ke committed
587
      for (int j = 0; j < num_class_; ++j) {
588
        tmp_result[j] = raw_scores[j * num_data + i];
Guolin Ke's avatar
Guolin Ke committed
589
590
591
      }
      Common::Softmax(&tmp_result);
      for (int j = 0; j < num_class_; ++j) {
Guolin Ke's avatar
Guolin Ke committed
592
        out_result[j * num_data + i] = static_cast<double>(tmp_result[j]);
Guolin Ke's avatar
Guolin Ke committed
593
594
      }
    }
595
596
  } else if (sigmoid_ > 0.0f) {
  #pragma omp parallel for schedule(static)
Guolin Ke's avatar
Guolin Ke committed
597
    for (data_size_t i = 0; i < num_data; ++i) {
598
      out_result[i] = static_cast<double>(1.0f / (1.0f + std::exp(-sigmoid_ * raw_scores[i])));
Guolin Ke's avatar
Guolin Ke committed
599
600
    }
  } else {
601
  #pragma omp parallel for schedule(static)
Guolin Ke's avatar
Guolin Ke committed
602
    for (data_size_t i = 0; i < num_data; ++i) {
Guolin Ke's avatar
Guolin Ke committed
603
      out_result[i] = static_cast<double>(raw_scores[i]);
Guolin Ke's avatar
Guolin Ke committed
604
605
606
607
608
    }
  }

}

Guolin Ke's avatar
Guolin Ke committed
609
void GBDT::Boosting() {
610
611
612
  if (object_function_ == nullptr) {
    Log::Fatal("No object function provided");
  }
Hui Xue's avatar
Hui Xue committed
613
  // objective function will calculate gradients and hessians
614
  int64_t num_score = 0;
Guolin Ke's avatar
Guolin Ke committed
615
  object_function_->
Guolin Ke's avatar
Guolin Ke committed
616
    GetGradients(GetTrainingScore(&num_score), gradients_.data(), hessians_.data());
Guolin Ke's avatar
Guolin Ke committed
617
618
}

619
std::string GBDT::DumpModel(int num_iteration) const {
Guolin Ke's avatar
Guolin Ke committed
620
  std::stringstream str_buf;
wxchan's avatar
wxchan committed
621

Guolin Ke's avatar
Guolin Ke committed
622
  str_buf << "{";
Guolin Ke's avatar
Guolin Ke committed
623
  str_buf << "\"name\":\"" << SubModelName() << "\"," << std::endl;
Guolin Ke's avatar
Guolin Ke committed
624
625
626
627
  str_buf << "\"num_class\":" << num_class_ << "," << std::endl;
  str_buf << "\"label_index\":" << label_idx_ << "," << std::endl;
  str_buf << "\"max_feature_idx\":" << max_feature_idx_ << "," << std::endl;
  str_buf << "\"sigmoid\":" << sigmoid_ << "," << std::endl;
wxchan's avatar
wxchan committed
628

629
630
631
  str_buf << "\"feature_names\":[\""
    << Common::Join(feature_names_, "\",\"") << "\"],"
    << std::endl;
Guolin Ke's avatar
Guolin Ke committed
632

Guolin Ke's avatar
Guolin Ke committed
633
  str_buf << "\"tree_info\":[";
634
635
  int num_used_model = static_cast<int>(models_.size());
  if (num_iteration > 0) {
Guolin Ke's avatar
Guolin Ke committed
636
    num_iteration += boost_from_average_ ? 1 : 0;
637
    num_used_model = std::min(num_iteration * num_class_, num_used_model);
638
  }
639
  for (int i = 0; i < num_used_model; ++i) {
wxchan's avatar
wxchan committed
640
    if (i > 0) {
Guolin Ke's avatar
Guolin Ke committed
641
      str_buf << ",";
wxchan's avatar
wxchan committed
642
    }
Guolin Ke's avatar
Guolin Ke committed
643
644
645
646
    str_buf << "{";
    str_buf << "\"tree_index\":" << i << ",";
    str_buf << models_[i]->ToJSON();
    str_buf << "}";
wxchan's avatar
wxchan committed
647
  }
Guolin Ke's avatar
Guolin Ke committed
648
  str_buf << "]" << std::endl;
wxchan's avatar
wxchan committed
649

Guolin Ke's avatar
Guolin Ke committed
650
  str_buf << "}" << std::endl;
wxchan's avatar
wxchan committed
651

Guolin Ke's avatar
Guolin Ke committed
652
  return str_buf.str();
wxchan's avatar
wxchan committed
653
654
}

Guolin Ke's avatar
Guolin Ke committed
655
std::string GBDT::SaveModelToString(int num_iteration) const {
656
  std::stringstream ss;
657

658
659
660
661
662
663
664
665
666
667
668
669
670
671
  // output model type
  ss << SubModelName() << std::endl;
  // output number of class
  ss << "num_class=" << num_class_ << std::endl;
  // output label index
  ss << "label_index=" << label_idx_ << std::endl;
  // output max_feature_idx
  ss << "max_feature_idx=" << max_feature_idx_ << std::endl;
  // output objective name
  if (object_function_ != nullptr) {
    ss << "objective=" << object_function_->GetName() << std::endl;
  }
  // output sigmoid parameter
  ss << "sigmoid=" << sigmoid_ << std::endl;
672

673
674
675
  if (boost_from_average_) {
    ss << "boost_from_average" << std::endl;
  }
Guolin Ke's avatar
Guolin Ke committed
676

677
  ss << "feature_names=" << Common::Join(feature_names_, " ") << std::endl;
678

679
  ss << "feature_infos=" << Common::Join(feature_infos_, " ") << std::endl;
680

681
682
  ss << std::endl;
  int num_used_model = static_cast<int>(models_.size());
Guolin Ke's avatar
Guolin Ke committed
683
684
685
  if (num_iteration > 0) {
    num_iteration += boost_from_average_ ? 1 : 0;
    num_used_model = std::min(num_iteration * num_class_, num_used_model);
686
687
688
689
690
691
692
693
694
695
696
697
698
699
  }
  // output tree models
  for (int i = 0; i < num_used_model; ++i) {
    ss << "Tree=" << i << std::endl;
    ss << models_[i]->ToString() << std::endl;
  }

  std::vector<std::pair<size_t, std::string>> pairs = FeatureImportance();
  ss << std::endl << "feature importances:" << std::endl;
  for (size_t i = 0; i < pairs.size(); ++i) {
    ss << pairs[i].second << "=" << std::to_string(pairs[i].first) << std::endl;
  }

  return ss.str();
700
701
}

702
bool GBDT::SaveModelToFile(int num_iteration, const char* filename) const {
wxchan's avatar
wxchan committed
703
704
705
  /*! \brief File to write models */
  std::ofstream output_file;
  output_file.open(filename);
706

707
  output_file << SaveModelToString(num_iteration);
708

wxchan's avatar
wxchan committed
709
  output_file.close();
710
711

  return (bool)output_file;
Guolin Ke's avatar
Guolin Ke committed
712
713
}

714
bool GBDT::LoadModelFromString(const std::string& model_str) {
Guolin Ke's avatar
Guolin Ke committed
715
716
717
  // use serialized string to restore this object
  models_.clear();
  std::vector<std::string> lines = Common::Split(model_str.c_str(), '\n');
718
719

  // get number of classes
720
721
722
723
  auto line = Common::FindFromLines(lines, "num_class=");
  if (line.size() > 0) {
    Common::Atoi(Common::Split(line.c_str(), '=')[1].c_str(), &num_class_);
  } else {
724
    Log::Fatal("Model file doesn't specify the number of classes");
725
    return false;
726
  }
Guolin Ke's avatar
Guolin Ke committed
727
  // get index of label
728
729
730
731
  line = Common::FindFromLines(lines, "label_index=");
  if (line.size() > 0) {
    Common::Atoi(Common::Split(line.c_str(), '=')[1].c_str(), &label_idx_);
  } else {
732
    Log::Fatal("Model file doesn't specify the label index");
733
    return false;
Guolin Ke's avatar
Guolin Ke committed
734
  }
Guolin Ke's avatar
Guolin Ke committed
735
  // get max_feature_idx first
736
737
738
739
  line = Common::FindFromLines(lines, "max_feature_idx=");
  if (line.size() > 0) {
    Common::Atoi(Common::Split(line.c_str(), '=')[1].c_str(), &max_feature_idx_);
  } else {
740
    Log::Fatal("Model file doesn't specify max_feature_idx");
741
    return false;
Guolin Ke's avatar
Guolin Ke committed
742
743
  }
  // get sigmoid parameter
744
745
746
747
  line = Common::FindFromLines(lines, "sigmoid=");
  if (line.size() > 0) {
    Common::Atof(Common::Split(line.c_str(), '=')[1].c_str(), &sigmoid_);
  } else {
748
    sigmoid_ = -1.0f;
Guolin Ke's avatar
Guolin Ke committed
749
  }
750
751
752
753
754
  // get boost_from_average_
  line = Common::FindFromLines(lines, "boost_from_average");
  if (line.size() > 0) {
    boost_from_average_ = true;
  }
Guolin Ke's avatar
Guolin Ke committed
755
756
757
  // get feature names
  line = Common::FindFromLines(lines, "feature_names=");
  if (line.size() > 0) {
Guolin Ke's avatar
Guolin Ke committed
758
    feature_names_ = Common::Split(line.substr(std::strlen("feature_names=")).c_str(), " ");
Guolin Ke's avatar
Guolin Ke committed
759
760
    if (feature_names_.size() != static_cast<size_t>(max_feature_idx_ + 1)) {
      Log::Fatal("Wrong size of feature_names");
761
      return false;
Guolin Ke's avatar
Guolin Ke committed
762
    }
763
  } else {
Guolin Ke's avatar
Guolin Ke committed
764
    Log::Fatal("Model file doesn't contain feature names");
765
    return false;
Guolin Ke's avatar
Guolin Ke committed
766
767
  }

Guolin Ke's avatar
Guolin Ke committed
768
769
770
771
772
773
774
775
776
777
778
779
  line = Common::FindFromLines(lines, "feature_infos=");
  if (line.size() > 0) {
    feature_infos_ = Common::Split(line.substr(std::strlen("feature_infos=")).c_str(), " ");
    if (feature_infos_.size() != static_cast<size_t>(max_feature_idx_ + 1)) {
      Log::Fatal("Wrong size of feature_infos");
      return false;
    }
  } else {
    Log::Fatal("Model file doesn't contain feature infos");
    return false;
  }

Guolin Ke's avatar
Guolin Ke committed
780
  // get tree models
781
  size_t i = 0;
Guolin Ke's avatar
Guolin Ke committed
782
783
784
785
786
787
788
  while (i < lines.size()) {
    size_t find_pos = lines[i].find("Tree=");
    if (find_pos != std::string::npos) {
      ++i;
      int start = static_cast<int>(i);
      while (i < lines.size() && lines[i].find("Tree=") == std::string::npos) { ++i; }
      int end = static_cast<int>(i);
Guolin Ke's avatar
Guolin Ke committed
789
      std::string tree_str = Common::Join<std::string>(lines, start, end, "\n");
Guolin Ke's avatar
Guolin Ke committed
790
791
      auto new_tree = std::unique_ptr<Tree>(new Tree(tree_str));
      models_.push_back(std::move(new_tree));
Guolin Ke's avatar
Guolin Ke committed
792
793
794
795
    } else {
      ++i;
    }
  }
796
  Log::Info("Finished loading %d models", models_.size());
wxchan's avatar
wxchan committed
797
798
  num_iteration_for_pred_ = static_cast<int>(models_.size()) / num_class_;
  num_init_iteration_ = num_iteration_for_pred_;
799
  iter_ = 0;
800
801

  return true;
Guolin Ke's avatar
Guolin Ke committed
802
803
}

wxchan's avatar
wxchan committed
804
std::vector<std::pair<size_t, std::string>> GBDT::FeatureImportance() const {
805

806
  std::vector<size_t> feature_importances(max_feature_idx_ + 1, 0);
807
808
809
  for (size_t iter = 0; iter < models_.size(); ++iter) {
    for (int split_idx = 0; split_idx < models_[iter]->num_leaves() - 1; ++split_idx) {
      ++feature_importances[models_[iter]->split_feature(split_idx)];
wxchan's avatar
wxchan committed
810
    }
811
812
813
814
815
816
  }
  // store the importance first
  std::vector<std::pair<size_t, std::string>> pairs;
  for (size_t i = 0; i < feature_importances.size(); ++i) {
    if (feature_importances[i] > 0) {
      pairs.emplace_back(feature_importances[i], feature_names_[i]);
817
    }
818
819
820
821
822
823
824
825
  }
  // sort the importance
  std::sort(pairs.begin(), pairs.end(),
            [] (const std::pair<size_t, std::string>& lhs,
                const std::pair<size_t, std::string>& rhs) {
    return lhs.first > rhs.first;
  });
  return pairs;
wxchan's avatar
wxchan committed
826
827
}

828
829
std::vector<double> GBDT::PredictRaw(const double* value) const {
  std::vector<double> ret(num_class_, 0.0f);
wxchan's avatar
wxchan committed
830
  for (int i = 0; i < num_iteration_for_pred_; ++i) {
831
832
833
    for (int j = 0; j < num_class_; ++j) {
      ret[j] += models_[i * num_class_ + j]->Predict(value);
    }
Guolin Ke's avatar
Guolin Ke committed
834
835
836
837
  }
  return ret;
}

838
std::vector<double> GBDT::Predict(const double* value) const {
839
  std::vector<double> ret(num_class_, 0.0f);
wxchan's avatar
wxchan committed
840
  for (int i = 0; i < num_iteration_for_pred_; ++i) {
841
842
    for (int j = 0; j < num_class_; ++j) {
      ret[j] += models_[i * num_class_ + j]->Predict(value);
843
844
    }
  }
845
846
  // if need sigmoid transform
  if (sigmoid_ > 0 && num_class_ == 1) {
847
    ret[0] = 1.0f / (1.0f + std::exp(-sigmoid_ * ret[0]));
848
849
850
  } else if (num_class_ > 1) {
    Common::Softmax(&ret);
  }
851
852
853
  return ret;
}

854
std::vector<int> GBDT::PredictLeafIndex(const double* value) const {
wxchan's avatar
wxchan committed
855
  std::vector<int> ret;
wxchan's avatar
wxchan committed
856
  for (int i = 0; i < num_iteration_for_pred_; ++i) {
857
858
859
    for (int j = 0; j < num_class_; ++j) {
      ret.push_back(models_[i * num_class_ + j]->PredictLeafIndex(value));
    }
wxchan's avatar
wxchan committed
860
861
862
863
  }
  return ret;
}

Guolin Ke's avatar
Guolin Ke committed
864
}  // namespace LightGBM