"python-package/vscode:/vscode.git/clone" did not exist on "2323cb3befa0515dd266d4a1393a37c833c11b83"
dataset.cpp 28.8 KB
Newer Older
Guolin Ke's avatar
Guolin Ke committed
1
#include <LightGBM/dataset.h>
Guolin Ke's avatar
Guolin Ke committed
2
#include <LightGBM/feature_group.h>
3
#include <LightGBM/utils/openmp_wrapper.h>
Guolin Ke's avatar
Guolin Ke committed
4
5
#include <LightGBM/utils/threading.h>
#include <LightGBM/utils/array_args.h>
Guolin Ke's avatar
Guolin Ke committed
6

zhangyafeikimi's avatar
zhangyafeikimi committed
7
#include <chrono>
Guolin Ke's avatar
Guolin Ke committed
8
9
10
11
12
13
#include <cstdio>
#include <unordered_map>
#include <limits>
#include <vector>
#include <utility>
#include <string>
Guolin Ke's avatar
Guolin Ke committed
14
#include <sstream>
Guolin Ke's avatar
Guolin Ke committed
15
16
17

namespace LightGBM {

18
const char* Dataset::binary_file_token = "______LightGBM_Binary_File_Token______\n";
Guolin Ke's avatar
Guolin Ke committed
19

Guolin Ke's avatar
Guolin Ke committed
20
Dataset::Dataset() {
21
  data_filename_ = "noname";
Guolin Ke's avatar
Guolin Ke committed
22
  num_data_ = 0;
Guolin Ke's avatar
Guolin Ke committed
23
  is_finish_load_ = false;
Guolin Ke's avatar
Guolin Ke committed
24
25
}

26
Dataset::Dataset(data_size_t num_data) {
Guolin Ke's avatar
Guolin Ke committed
27
  CHECK(num_data > 0);
Guolin Ke's avatar
Guolin Ke committed
28
  data_filename_ = "noname";
Guolin Ke's avatar
Guolin Ke committed
29
  num_data_ = num_data;
Guolin Ke's avatar
Guolin Ke committed
30
  metadata_.Init(num_data_, NO_SPECIFIC, NO_SPECIFIC);
Guolin Ke's avatar
Guolin Ke committed
31
  is_finish_load_ = false;
Guolin Ke's avatar
Guolin Ke committed
32
  group_bin_boundaries_.push_back(0);
Guolin Ke's avatar
Guolin Ke committed
33
34
}

Guolin Ke's avatar
Guolin Ke committed
35
Dataset::~Dataset() {
Guolin Ke's avatar
Guolin Ke committed
36
}
Guolin Ke's avatar
Guolin Ke committed
37

Guolin Ke's avatar
Guolin Ke committed
38
39
40
41
42
43
44
45
46
47
std::vector<std::vector<int>> NoGroup(
  const std::vector<int>& used_features) {
  std::vector<std::vector<int>> features_in_group;
  features_in_group.resize(used_features.size());
  for (size_t i = 0; i < used_features.size(); ++i) {
    features_in_group[i].emplace_back(used_features[i]);
  }
  return features_in_group;
}

Guolin Ke's avatar
Guolin Ke committed
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
int GetConfilctCount(const std::vector<bool>& mark, const int* indices, int num_indices, int max_cnt) {
  int ret = 0;
  for (int i = 0; i < num_indices; ++i) {
    if (mark[indices[i]]) {
      ++ret;
      if (ret > max_cnt) {
        return -1;
      }
    }
  }
  return ret;
}
void MarkUsed(std::vector<bool>& mark, const int* indices, int num_indices) {
  for (int i = 0; i < num_indices; ++i) {
    mark[indices[i]] = true;
  }
}

std::vector<std::vector<int>> FindGroups(const std::vector<std::unique_ptr<BinMapper>>& bin_mappers,
                                         const std::vector<int>& find_order,
                                         int** sample_indices,
                                         const int* num_per_col,
                                         size_t total_sample_cnt,
                                         data_size_t max_error_cnt,
                                         data_size_t filter_cnt,
Guolin Ke's avatar
Guolin Ke committed
73
74
                                         data_size_t num_data,
                                         bool is_use_gpu) {
Guolin Ke's avatar
Guolin Ke committed
75
  const int max_search_group = 100;
Guolin Ke's avatar
Guolin Ke committed
76
  const int gpu_max_bin_per_group = 256;
Guolin Ke's avatar
Guolin Ke committed
77
78
79
80
81
82
83
84
85
86
87
88
  Random rand(num_data);
  std::vector<std::vector<int>> features_in_group;
  std::vector<std::vector<bool>> conflict_marks;
  std::vector<int> group_conflict_cnt;
  std::vector<size_t> group_non_zero_cnt;
  std::vector<int> group_num_bin;

  for (auto fidx : find_order) {
    const size_t cur_non_zero_cnt = num_per_col[fidx];
    bool need_new_group = true;
    std::vector<int> available_groups;
    for (int gid = 0; gid < static_cast<int>(features_in_group.size()); ++gid) {
Guolin Ke's avatar
Guolin Ke committed
89
90
91
92
93
      if (group_non_zero_cnt[gid] + cur_non_zero_cnt <= total_sample_cnt + max_error_cnt){
        if (!is_use_gpu || group_num_bin[gid] + bin_mappers[fidx]->num_bin() + (bin_mappers[fidx]->GetDefaultBin() == 0 ? -1 : 0)
            <= gpu_max_bin_per_group) {
          available_groups.push_back(gid);
        }
Guolin Ke's avatar
Guolin Ke committed
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
      }
    }
    std::vector<int> search_groups;
    if (!available_groups.empty()) {
      int last = static_cast<int>(available_groups.size()) - 1;
      auto indices = rand.Sample(last, std::min(last, max_search_group - 1));
      search_groups.push_back(available_groups.back());
      for (auto idx : indices) {
        search_groups.push_back(available_groups[idx]);
      }
    }
    for (auto gid : search_groups) {
      const int rest_max_cnt = max_error_cnt - group_conflict_cnt[gid];
      int cnt = GetConfilctCount(conflict_marks[gid], sample_indices[fidx], num_per_col[fidx], rest_max_cnt);
      if (cnt >= 0 && cnt <= rest_max_cnt) {
        data_size_t rest_non_zero_data = static_cast<data_size_t>(
          static_cast<double>(cur_non_zero_cnt - cnt) * num_data / total_sample_cnt);
        if (rest_non_zero_data < filter_cnt) { continue; }
        need_new_group = false;
        features_in_group[gid].push_back(fidx);
        group_conflict_cnt[gid] += cnt;
        group_non_zero_cnt[gid] += cur_non_zero_cnt - cnt;
        MarkUsed(conflict_marks[gid], sample_indices[fidx], num_per_col[fidx]);
Guolin Ke's avatar
Guolin Ke committed
117
118
119
        if (is_use_gpu) {
          group_num_bin[gid] += bin_mappers[fidx]->num_bin() + (bin_mappers[fidx]->GetDefaultBin() == 0 ? -1 : 0);
        }
Guolin Ke's avatar
Guolin Ke committed
120
121
122
123
124
125
126
127
128
129
        break;
      }
    }
    if (need_new_group) {
      features_in_group.emplace_back();
      features_in_group.back().push_back(fidx);
      group_conflict_cnt.push_back(0);
      conflict_marks.emplace_back(total_sample_cnt, false);
      MarkUsed(conflict_marks.back(), sample_indices[fidx], num_per_col[fidx]);
      group_non_zero_cnt.emplace_back(cur_non_zero_cnt);
Guolin Ke's avatar
Guolin Ke committed
130
131
132
      if (is_use_gpu) {
        group_num_bin.push_back(1 + bin_mappers[fidx]->num_bin() + (bin_mappers[fidx]->GetDefaultBin() == 0 ? -1 : 0));
      }
Guolin Ke's avatar
Guolin Ke committed
133
134
135
136
137
138
139
140
141
142
143
144
145
146
    }
  }
  return features_in_group;
}

std::vector<std::vector<int>> FastFeatureBundling(std::vector<std::unique_ptr<BinMapper>>& bin_mappers,
                                                  int** sample_indices,
                                                  const int* num_per_col,
                                                  size_t total_sample_cnt,
                                                  const std::vector<int>& used_features,
                                                  double max_conflict_rate,
                                                  data_size_t num_data,
                                                  data_size_t min_data,
                                                  double sparse_threshold,
Guolin Ke's avatar
Guolin Ke committed
147
148
                                                  bool is_enable_sparse,
                                                  bool is_use_gpu) {
Guolin Ke's avatar
Guolin Ke committed
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
  // filter is based on sampling data, so decrease its range
  const data_size_t filter_cnt = static_cast<data_size_t>(static_cast<double>(0.95 * min_data) / num_data * total_sample_cnt);
  const data_size_t max_error_cnt = static_cast<data_size_t>(total_sample_cnt * max_conflict_rate);
  int cur_used_feature_cnt = 0;
  std::vector<size_t> feature_non_zero_cnt;
  // put dense feature first
  for (auto fidx : used_features) {
    feature_non_zero_cnt.emplace_back(num_per_col[fidx]);
    ++cur_used_feature_cnt;
  }
  // sort by non zero cnt
  std::vector<int> sorted_idx;
  for (int i = 0; i < cur_used_feature_cnt; ++i) {
    sorted_idx.emplace_back(i);
  }
  // sort by non zero cnt, bigger first
  std::sort(sorted_idx.begin(), sorted_idx.end(),
            [&feature_non_zero_cnt](int a, int b) {
    return feature_non_zero_cnt[a] > feature_non_zero_cnt[b];
  });

  std::vector<int> feature_order_by_cnt;
  for (auto sidx : sorted_idx) {
    feature_order_by_cnt.push_back(used_features[sidx]);
  }
Guolin Ke's avatar
Guolin Ke committed
174
175
  auto features_in_group = FindGroups(bin_mappers, used_features, sample_indices, num_per_col, total_sample_cnt, max_error_cnt, filter_cnt, num_data, is_use_gpu);
  auto group2 = FindGroups(bin_mappers, feature_order_by_cnt, sample_indices, num_per_col, total_sample_cnt, max_error_cnt, filter_cnt, num_data, is_use_gpu);
Guolin Ke's avatar
Guolin Ke committed
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
  if (features_in_group.size() > group2.size()) {
    features_in_group = group2;
  }
  std::vector<std::vector<int>> ret;
  for (size_t i = 0; i < features_in_group.size(); ++i) {
    if (features_in_group[i].size() <= 1 || features_in_group[i].size() >= 5) {
      ret.push_back(features_in_group[i]);
    } else {
      int cnt_non_zero = 0;
      for (size_t j = 0; j < features_in_group[i].size(); ++j) {
        const int fidx = features_in_group[i][j];
        cnt_non_zero += static_cast<int>(num_data * (1.0f - bin_mappers[fidx]->sparse_rate()));
      }
      double sparse_rate = 1.0f - static_cast<double>(cnt_non_zero) / (num_data);
      // take apart small sparse group, due it will not gain on speed 
      if (sparse_rate >= sparse_threshold && is_enable_sparse) {
        for (size_t j = 0; j < features_in_group[i].size(); ++j) {
          const int fidx = features_in_group[i][j];
          ret.emplace_back();
          ret.back().push_back(fidx);
        }
      } else {
        ret.push_back(features_in_group[i]);
      }
    }
  }
  // shuffle groups
  int num_group = static_cast<int>(ret.size());
  Random tmp_rand(12);
  for (int i = 0; i < num_group - 1; ++i) {
    int j = tmp_rand.NextShort(i + 1, num_group);
    std::swap(ret[i], ret[j]);
  }
  return ret;
}

Guolin Ke's avatar
Guolin Ke committed
212
213
void Dataset::Construct(
  std::vector<std::unique_ptr<BinMapper>>& bin_mappers,
Guolin Ke's avatar
Guolin Ke committed
214
215
216
  int** sample_non_zero_indices,
  const int* num_per_col,
  size_t total_sample_cnt,
Guolin Ke's avatar
Guolin Ke committed
217
  const IOConfig& io_config) {
Guolin Ke's avatar
Guolin Ke committed
218

Guolin Ke's avatar
Guolin Ke committed
219
  num_total_features_ = static_cast<int>(bin_mappers.size());
220
  sparse_threshold_ = io_config.sparse_threshold;
Guolin Ke's avatar
Guolin Ke committed
221
222
223
224
225
  // get num_features
  std::vector<int> used_features;
  for (int i = 0; i < static_cast<int>(bin_mappers.size()); ++i) {
    if (bin_mappers[i] != nullptr && !bin_mappers[i]->is_trival()) {
      used_features.emplace_back(i);
Guolin Ke's avatar
Guolin Ke committed
226
    }
Guolin Ke's avatar
Guolin Ke committed
227
  }
Guolin Ke's avatar
Guolin Ke committed
228
  if (used_features.empty()) {
229
230
231
232
    Log::Fatal("Cannot construct Dataset since there are not useful features. \
                It should be at least two unique rows. \
                If the num_row (num_data) is small, you can set min_data=1 and min_data_in_bin=1 to fix this. \
                Otherwise please make sure you are using the right dataset.");
Guolin Ke's avatar
Guolin Ke committed
233
  }
Guolin Ke's avatar
Guolin Ke committed
234
235
  auto features_in_group = NoGroup(used_features);

Guolin Ke's avatar
Guolin Ke committed
236
237
238
239
240
  if (io_config.enable_bundle) {
    features_in_group = FastFeatureBundling(bin_mappers,
                                            sample_non_zero_indices, num_per_col, total_sample_cnt,
                                            used_features, io_config.max_conflict_rate,
                                            num_data_, io_config.min_data_in_leaf,
Guolin Ke's avatar
Guolin Ke committed
241
                                            sparse_threshold_, io_config.is_enable_sparse, io_config.device_type == std::string("gpu"));
Guolin Ke's avatar
Guolin Ke committed
242
243
  }

Guolin Ke's avatar
Guolin Ke committed
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
  num_features_ = 0;
  for (const auto& fs : features_in_group) {
    num_features_ += static_cast<int>(fs.size());
  }
  int cur_fidx = 0;
  used_feature_map_ = std::vector<int>(num_total_features_, -1);
  num_groups_ = static_cast<int>(features_in_group.size());
  real_feature_idx_.resize(num_features_);
  feature2group_.resize(num_features_);
  feature2subfeature_.resize(num_features_);
  for (int i = 0; i < num_groups_; ++i) {
    auto cur_features = features_in_group[i];
    int cur_cnt_features = static_cast<int>(cur_features.size());
    // get bin_mappers
    std::vector<std::unique_ptr<BinMapper>> cur_bin_mappers;
    for (int j = 0; j < cur_cnt_features; ++j) {
      int real_fidx = cur_features[j];
      used_feature_map_[real_fidx] = cur_fidx;
      real_feature_idx_[cur_fidx] = real_fidx;
      feature2group_[cur_fidx] = i;
      feature2subfeature_[cur_fidx] = j;
      cur_bin_mappers.emplace_back(bin_mappers[real_fidx].release());
      ++cur_fidx;
    }
    feature_groups_.emplace_back(std::unique_ptr<FeatureGroup>(
Guolin Ke's avatar
Guolin Ke committed
269
270
      new FeatureGroup(cur_cnt_features, cur_bin_mappers, num_data_, sparse_threshold_,
                       io_config.is_enable_sparse)));
Guolin Ke's avatar
Guolin Ke committed
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
  }
  feature_groups_.shrink_to_fit();
  group_bin_boundaries_.clear();
  uint64_t num_total_bin = 0;
  group_bin_boundaries_.push_back(num_total_bin);
  for (int i = 0; i < num_groups_; ++i) {
    num_total_bin += feature_groups_[i]->num_total_bin_;
    group_bin_boundaries_.push_back(num_total_bin);
  }
  int last_group = 0;
  group_feature_start_.reserve(num_groups_);
  group_feature_cnt_.reserve(num_groups_);
  group_feature_start_.push_back(0);
  group_feature_cnt_.push_back(1);
  for (int i = 1; i < num_features_; ++i) {
    const int group = feature2group_[i];
    if (group == last_group) {
      group_feature_cnt_.back() = group_feature_cnt_.back() + 1;
    } else {
      group_feature_start_.push_back(i);
      group_feature_cnt_.push_back(1);
      last_group = group;
    }
  }
Guolin Ke's avatar
Guolin Ke committed
295
296
}

Guolin Ke's avatar
Guolin Ke committed
297
void Dataset::FinishLoad() {
Guolin Ke's avatar
Guolin Ke committed
298
  if (is_finish_load_) { return; }
299
  OMP_INIT_EX();
Guolin Ke's avatar
Guolin Ke committed
300
  #pragma omp parallel for schedule(guided)
Guolin Ke's avatar
Guolin Ke committed
301
  for (int i = 0; i < num_groups_; ++i) {
302
    OMP_LOOP_EX_BEGIN();
Guolin Ke's avatar
Guolin Ke committed
303
    feature_groups_[i]->bin_data_->FinishLoad();
304
    OMP_LOOP_EX_END();
Guolin Ke's avatar
Guolin Ke committed
305
  }
306
  OMP_THROW_EX();
Guolin Ke's avatar
Guolin Ke committed
307
  is_finish_load_ = true;
Guolin Ke's avatar
Guolin Ke committed
308
}
Guolin Ke's avatar
Guolin Ke committed
309

310
void Dataset::CopyFeatureMapperFrom(const Dataset* dataset) {
Guolin Ke's avatar
Guolin Ke committed
311
  feature_groups_.clear();
Guolin Ke's avatar
Guolin Ke committed
312
  num_features_ = dataset->num_features_;
Guolin Ke's avatar
Guolin Ke committed
313
  num_groups_ = dataset->num_groups_;
314
  sparse_threshold_ = dataset->sparse_threshold_;
Guolin Ke's avatar
Guolin Ke committed
315
  // copy feature bin mapper data
Guolin Ke's avatar
Guolin Ke committed
316
317
318
319
320
321
322
323
  for (int i = 0; i < num_groups_; ++i) {
    std::vector<std::unique_ptr<BinMapper>> bin_mappers;
    for (int j = 0; j < dataset->feature_groups_[i]->num_feature_; ++j) {
      bin_mappers.emplace_back(new BinMapper(*(dataset->feature_groups_[i]->bin_mappers_[j])));
    }
    feature_groups_.emplace_back(new FeatureGroup(
      dataset->feature_groups_[i]->num_feature_,
      bin_mappers,
Guolin Ke's avatar
Guolin Ke committed
324
      num_data_,
Guolin Ke's avatar
Guolin Ke committed
325
      dataset->feature_groups_[i]->is_sparse_));
Guolin Ke's avatar
Guolin Ke committed
326
  }
Guolin Ke's avatar
Guolin Ke committed
327
  feature_groups_.shrink_to_fit();
Guolin Ke's avatar
Guolin Ke committed
328
329
330
  used_feature_map_ = dataset->used_feature_map_;
  num_total_features_ = dataset->num_total_features_;
  feature_names_ = dataset->feature_names_;
Guolin Ke's avatar
Guolin Ke committed
331
  label_idx_ = dataset->label_idx_;
Guolin Ke's avatar
Guolin Ke committed
332
333
334
335
336
337
338
339
340
341
342
343
  real_feature_idx_ = dataset->real_feature_idx_;
  feature2group_ = dataset->feature2group_;
  feature2subfeature_ = dataset->feature2subfeature_;
  group_bin_boundaries_ = dataset->group_bin_boundaries_;
  group_feature_start_ = dataset->group_feature_start_;
  group_feature_cnt_ = dataset->group_feature_cnt_;
}

void Dataset::CreateValid(const Dataset* dataset) {
  feature_groups_.clear();
  num_features_ = dataset->num_features_;
  num_groups_ = num_features_;
344
  sparse_threshold_ = dataset->sparse_threshold_;
Guolin Ke's avatar
Guolin Ke committed
345
346
347
348
349
350
351
352
353
354
355
  bool is_enable_sparse = true;
  feature2group_.clear();
  feature2subfeature_.clear();
  // copy feature bin mapper data
  for (int i = 0; i < num_features_; ++i) {
    std::vector<std::unique_ptr<BinMapper>> bin_mappers;
    bin_mappers.emplace_back(new BinMapper(*(dataset->FeatureBinMapper(i))));
    feature_groups_.emplace_back(new FeatureGroup(
      1,
      bin_mappers,
      num_data_,
356
      dataset->sparse_threshold_,
Guolin Ke's avatar
Guolin Ke committed
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
      is_enable_sparse));
    feature2group_.push_back(i);
    feature2subfeature_.push_back(0);
  }

  feature_groups_.shrink_to_fit();
  used_feature_map_ = dataset->used_feature_map_;
  num_total_features_ = dataset->num_total_features_;
  feature_names_ = dataset->feature_names_;
  label_idx_ = dataset->label_idx_;
  real_feature_idx_ = dataset->real_feature_idx_;
  group_bin_boundaries_.clear();
  uint64_t num_total_bin = 0;
  group_bin_boundaries_.push_back(num_total_bin);
  for (int i = 0; i < num_groups_; ++i) {
    num_total_bin += feature_groups_[i]->num_total_bin_;
    group_bin_boundaries_.push_back(num_total_bin);
  }
  int last_group = 0;
  group_feature_start_.reserve(num_groups_);
  group_feature_cnt_.reserve(num_groups_);
  group_feature_start_.push_back(0);
  group_feature_cnt_.push_back(1);
  for (int i = 1; i < num_features_; ++i) {
    const int group = feature2group_[i];
    if (group == last_group) {
      group_feature_cnt_.back() = group_feature_cnt_.back() + 1;
    } else {
      group_feature_start_.push_back(i);
      group_feature_cnt_.push_back(1);
      last_group = group;
    }
  }
Guolin Ke's avatar
Guolin Ke committed
390
391
}

Guolin Ke's avatar
Guolin Ke committed
392
393
394
void Dataset::ReSize(data_size_t num_data) {
  if (num_data_ != num_data) {
    num_data_ = num_data;
395
    OMP_INIT_EX();
Guolin Ke's avatar
Guolin Ke committed
396
    #pragma omp parallel for schedule(static)
Guolin Ke's avatar
Guolin Ke committed
397
    for (int group = 0; group < num_groups_; ++group) {
398
      OMP_LOOP_EX_BEGIN();
Guolin Ke's avatar
Guolin Ke committed
399
      feature_groups_[group]->bin_data_->ReSize(num_data_);
400
      OMP_LOOP_EX_END();
Guolin Ke's avatar
Guolin Ke committed
401
    }
402
    OMP_THROW_EX();
Guolin Ke's avatar
Guolin Ke committed
403
404
405
406
407
  }
}

void Dataset::CopySubset(const Dataset* fullset, const data_size_t* used_indices, data_size_t num_used_indices, bool need_meta_data) {
  CHECK(num_used_indices == num_data_);
408
  OMP_INIT_EX();
Guolin Ke's avatar
Guolin Ke committed
409
  #pragma omp parallel for schedule(static)
Guolin Ke's avatar
Guolin Ke committed
410
  for (int group = 0; group < num_groups_; ++group) {
411
    OMP_LOOP_EX_BEGIN();
Guolin Ke's avatar
Guolin Ke committed
412
    feature_groups_[group]->CopySubset(fullset->feature_groups_[group].get(), used_indices, num_used_indices);
413
    OMP_LOOP_EX_END();
Guolin Ke's avatar
Guolin Ke committed
414
  }
415
  OMP_THROW_EX();
Guolin Ke's avatar
Guolin Ke committed
416
  if (need_meta_data) {
Guolin Ke's avatar
Guolin Ke committed
417
    metadata_.Init(fullset->metadata_, used_indices, num_used_indices);
Guolin Ke's avatar
Guolin Ke committed
418
  }
Guolin Ke's avatar
Guolin Ke committed
419
  is_finish_load_ = true;
Guolin Ke's avatar
Guolin Ke committed
420
421
}

422
bool Dataset::SetFloatField(const char* field_name, const float* field_data, data_size_t num_element) {
Guolin Ke's avatar
Guolin Ke committed
423
424
425
  std::string name(field_name);
  name = Common::Trim(name);
  if (name == std::string("label") || name == std::string("target")) {
426
427
428
    #ifdef LABEL_T_USE_DOUBLE
    Log::Fatal("Don't Support LABEL_T_USE_DOUBLE.");
    #else
429
    metadata_.SetLabel(field_data, num_element);
430
    #endif
Guolin Ke's avatar
Guolin Ke committed
431
  } else if (name == std::string("weight") || name == std::string("weights")) {
432
433
434
    #ifdef LABEL_T_USE_DOUBLE
    Log::Fatal("Don't Support LABEL_T_USE_DOUBLE.");
    #else
435
    metadata_.SetWeights(field_data, num_element);
436
    #endif
Guolin Ke's avatar
Guolin Ke committed
437
438
439
440
441
442
443
444
445
446
  } else {
    return false;
  }
  return true;
}

bool Dataset::SetDoubleField(const char* field_name, const double* field_data, data_size_t num_element) {
  std::string name(field_name);
  name = Common::Trim(name);
  if (name == std::string("init_score")) {
447
    metadata_.SetInitScore(field_data, num_element);
Guolin Ke's avatar
Guolin Ke committed
448
  } else {
449
    return false;
Guolin Ke's avatar
Guolin Ke committed
450
  }
451
  return true;
Guolin Ke's avatar
Guolin Ke committed
452
453
}

454
455
456
457
bool Dataset::SetIntField(const char* field_name, const int* field_data, data_size_t num_element) {
  std::string name(field_name);
  name = Common::Trim(name);
  if (name == std::string("query") || name == std::string("group")) {
Guolin Ke's avatar
Guolin Ke committed
458
    metadata_.SetQuery(field_data, num_element);
459
460
461
462
463
464
  } else {
    return false;
  }
  return true;
}

Guolin Ke's avatar
Guolin Ke committed
465
bool Dataset::GetFloatField(const char* field_name, data_size_t* out_len, const float** out_ptr) {
466
467
468
  std::string name(field_name);
  name = Common::Trim(name);
  if (name == std::string("label") || name == std::string("target")) {
469
470
471
    #ifdef LABEL_T_USE_DOUBLE
    Log::Fatal("Don't Support LABEL_T_USE_DOUBLE.");
    #else
472
473
    *out_ptr = metadata_.label();
    *out_len = num_data_;
474
    #endif
475
  } else if (name == std::string("weight") || name == std::string("weights")) {
476
477
478
    #ifdef LABEL_T_USE_DOUBLE
    Log::Fatal("Don't Support LABEL_T_USE_DOUBLE.");
    #else
479
480
    *out_ptr = metadata_.weights();
    *out_len = num_data_;
481
    #endif
Guolin Ke's avatar
Guolin Ke committed
482
483
484
485
486
487
488
489
490
491
  } else {
    return false;
  }
  return true;
}

bool Dataset::GetDoubleField(const char* field_name, data_size_t* out_len, const double** out_ptr) {
  std::string name(field_name);
  name = Common::Trim(name);
  if (name == std::string("init_score")) {
492
    *out_ptr = metadata_.init_score();
Guolin Ke's avatar
Guolin Ke committed
493
    *out_len = static_cast<data_size_t>(metadata_.num_init_score());
494
495
496
  } else {
    return false;
  }
497
  return true;
498
499
}

Guolin Ke's avatar
Guolin Ke committed
500
bool Dataset::GetIntField(const char* field_name, data_size_t* out_len, const int** out_ptr) {
501
502
503
  std::string name(field_name);
  name = Common::Trim(name);
  if (name == std::string("query") || name == std::string("group")) {
504
    *out_ptr = metadata_.query_boundaries();
Guolin Ke's avatar
Guolin Ke committed
505
    *out_len = metadata_.num_queries() + 1;
Guolin Ke's avatar
Guolin Ke committed
506
507
508
  } else {
    return false;
  }
509
  return true;
510
511
}

Guolin Ke's avatar
Guolin Ke committed
512
void Dataset::SaveBinaryFile(const char* bin_filename) {
Guolin Ke's avatar
Guolin Ke committed
513
  if (bin_filename != nullptr
Guolin Ke's avatar
Guolin Ke committed
514
      && std::string(bin_filename) == data_filename_) {
Guolin Ke's avatar
Guolin Ke committed
515
516
517
    Log::Warning("Bianry file %s already existed", bin_filename);
    return;
  }
Guolin Ke's avatar
Guolin Ke committed
518
  // if not pass a filename, just append ".bin" of original file
Guolin Ke's avatar
Guolin Ke committed
519
  std::string bin_filename_str(data_filename_);
Guolin Ke's avatar
Guolin Ke committed
520
521
522
523
  if (bin_filename == nullptr || bin_filename[0] == '\0') {
    bin_filename_str.append(".bin");
    bin_filename = bin_filename_str.c_str();
  }
Guolin Ke's avatar
Guolin Ke committed
524
  bool is_file_existed = false;
525
526

  if (VirtualFileWriter::Exists(bin_filename)) {
Guolin Ke's avatar
Guolin Ke committed
527
528
529
    is_file_existed = true;
    Log::Warning("File %s existed, cannot save binary to it", bin_filename);
  }
Guolin Ke's avatar
Guolin Ke committed
530

Guolin Ke's avatar
Guolin Ke committed
531
  if (!is_file_existed) {
532
533
    auto writer = VirtualFileWriter::Make(bin_filename);
    if (!writer->Init()) {
Guolin Ke's avatar
Guolin Ke committed
534
      Log::Fatal("Cannot write binary data to %s ", bin_filename);
Guolin Ke's avatar
Guolin Ke committed
535
    }
536
    Log::Info("Saving data to binary file %s", bin_filename);
537
    size_t size_of_token = std::strlen(binary_file_token);
538
    writer->Write(binary_file_token, size_of_token);
Guolin Ke's avatar
Guolin Ke committed
539
    // get size of header
Guolin Ke's avatar
Guolin Ke committed
540
    size_t size_of_header = sizeof(num_data_) + sizeof(num_features_) + sizeof(num_total_features_)
Guolin Ke's avatar
Guolin Ke committed
541
      + sizeof(int) * num_total_features_ + sizeof(label_idx_) + sizeof(num_groups_)
Guolin Ke's avatar
Guolin Ke committed
542
      + 3 * sizeof(int) * num_features_ + sizeof(uint64_t) * (num_groups_ + 1) + 2 * sizeof(int) * num_groups_;
543
544
545
546
    // size of feature names
    for (int i = 0; i < num_total_features_; ++i) {
      size_of_header += feature_names_[i].size() + sizeof(int);
    }
547
    writer->Write(&size_of_header, sizeof(size_of_header));
Guolin Ke's avatar
Guolin Ke committed
548
    // write header
549
550
551
552
553
554
555
556
557
558
559
560
    writer->Write(&num_data_, sizeof(num_data_));
    writer->Write(&num_features_, sizeof(num_features_));
    writer->Write(&num_total_features_, sizeof(num_total_features_));
    writer->Write(&label_idx_, sizeof(label_idx_));
    writer->Write(used_feature_map_.data(), sizeof(int) * num_total_features_);
    writer->Write(&num_groups_, sizeof(num_groups_));
    writer->Write(real_feature_idx_.data(), sizeof(int) * num_features_);
    writer->Write(feature2group_.data(), sizeof(int) * num_features_);
    writer->Write(feature2subfeature_.data(), sizeof(int) * num_features_);
    writer->Write(group_bin_boundaries_.data(), sizeof(uint64_t) * (num_groups_ + 1));
    writer->Write(group_feature_start_.data(), sizeof(int) * num_groups_);
    writer->Write(group_feature_cnt_.data(), sizeof(int) * num_groups_);
Guolin Ke's avatar
Guolin Ke committed
561

562
563
564
    // write feature names
    for (int i = 0; i < num_total_features_; ++i) {
      int str_len = static_cast<int>(feature_names_[i].size());
565
      writer->Write(&str_len, sizeof(int));
566
      const char* c_str = feature_names_[i].c_str();
567
      writer->Write(c_str, sizeof(char) * str_len);
568
569
    }

Guolin Ke's avatar
Guolin Ke committed
570
571
    // get size of meta data
    size_t size_of_metadata = metadata_.SizesInByte();
572
    writer->Write(&size_of_metadata, sizeof(size_of_metadata));
Guolin Ke's avatar
Guolin Ke committed
573
    // write meta data
574
    metadata_.SaveBinaryToFile(writer.get());
Guolin Ke's avatar
Guolin Ke committed
575
576

    // write feature data
Guolin Ke's avatar
Guolin Ke committed
577
    for (int i = 0; i < num_groups_; ++i) {
Guolin Ke's avatar
Guolin Ke committed
578
      // get size of feature
Guolin Ke's avatar
Guolin Ke committed
579
      size_t size_of_feature = feature_groups_[i]->SizesInByte();
580
      writer->Write(&size_of_feature, sizeof(size_of_feature));
Guolin Ke's avatar
Guolin Ke committed
581
      // write feature
582
      feature_groups_[i]->SaveBinaryToFile(writer.get());
Guolin Ke's avatar
Guolin Ke committed
583
584
585
586
    }
  }
}

587
588
589
590
void Dataset::ConstructHistograms(const std::vector<int8_t>& is_feature_used,
                                  const data_size_t* data_indices, data_size_t num_data,
                                  int leaf_idx,
                                  std::vector<std::unique_ptr<OrderedBin>>& ordered_bins,
591
592
                                  const score_t* gradients, const score_t* hessians,
                                  score_t* ordered_gradients, score_t* ordered_hessians,
593
594
                                  bool is_constant_hessian,
                                  HistogramBinEntry* hist_data) const {
Guolin Ke's avatar
Guolin Ke committed
595

zhangjin's avatar
zhangjin committed
596
  if (leaf_idx < 0 || num_data < 0 || hist_data == nullptr) {
Guolin Ke's avatar
Guolin Ke committed
597
598
    return;
  }
Guolin Ke's avatar
Guolin Ke committed
599
600
601
602
603
604
605
606
607
608
609
610
611
612

  std::vector<int> used_group;
  used_group.reserve(num_groups_);
  for (int group = 0; group < num_groups_; ++group) {
    const int f_cnt = group_feature_cnt_[group];
    for (int j = 0; j < f_cnt; ++j) {
      const int fidx = group_feature_start_[group] + j;
      if (is_feature_used[fidx]) {
        break;
      }
    }
    used_group.push_back(group);
  }
  int num_used_group = static_cast<int>(used_group.size());
Guolin Ke's avatar
Guolin Ke committed
613
614
615
  auto ptr_ordered_grad = gradients;
  auto ptr_ordered_hess = hessians;
  if (data_indices != nullptr && num_data < num_data_) {
616
617
618
619
620
621
622
623
624
625
626
    if (!is_constant_hessian) {
      #pragma omp parallel for schedule(static)
      for (data_size_t i = 0; i < num_data; ++i) {
        ordered_gradients[i] = gradients[data_indices[i]];
        ordered_hessians[i] = hessians[data_indices[i]];
      }
    } else {
      #pragma omp parallel for schedule(static)
      for (data_size_t i = 0; i < num_data; ++i) {
        ordered_gradients[i] = gradients[data_indices[i]];
      }
Guolin Ke's avatar
Guolin Ke committed
627
628
629
    }
    ptr_ordered_grad = ordered_gradients;
    ptr_ordered_hess = ordered_hessians;
630
631
632
    if (!is_constant_hessian) {
      OMP_INIT_EX();
      #pragma omp parallel for schedule(static)
Guolin Ke's avatar
Guolin Ke committed
633
      for (int gi = 0; gi < num_used_group; ++gi) {
634
        OMP_LOOP_EX_BEGIN();
Guolin Ke's avatar
Guolin Ke committed
635
        int group = used_group[gi];
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
        // feature is not used
        auto data_ptr = hist_data + group_bin_boundaries_[group];
        const int num_bin = feature_groups_[group]->num_total_bin_;
        std::memset(data_ptr + 1, 0, (num_bin - 1) * sizeof(HistogramBinEntry));
        // construct histograms for smaller leaf
        if (ordered_bins[group] == nullptr) {
          // if not use ordered bin
          feature_groups_[group]->bin_data_->ConstructHistogram(
            data_indices,
            num_data,
            ptr_ordered_grad,
            ptr_ordered_hess,
            data_ptr);
        } else {
          // used ordered bin
          ordered_bins[group]->ConstructHistogram(leaf_idx,
                                                  gradients,
                                                  hessians,
                                                  data_ptr);
655
        }
656
        OMP_LOOP_EX_END();
657
      }
658
659
660
661
      OMP_THROW_EX();
    } else {
      OMP_INIT_EX();
      #pragma omp parallel for schedule(static)
Guolin Ke's avatar
Guolin Ke committed
662
      for (int gi = 0; gi < num_used_group; ++gi) {
663
        OMP_LOOP_EX_BEGIN();
Guolin Ke's avatar
Guolin Ke committed
664
        int group = used_group[gi];
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
        // feature is not used
        auto data_ptr = hist_data + group_bin_boundaries_[group];
        const int num_bin = feature_groups_[group]->num_total_bin_;
        std::memset(data_ptr + 1, 0, (num_bin - 1) * sizeof(HistogramBinEntry));
        // construct histograms for smaller leaf
        if (ordered_bins[group] == nullptr) {
          // if not use ordered bin
          feature_groups_[group]->bin_data_->ConstructHistogram(
            data_indices,
            num_data,
            ptr_ordered_grad,
            data_ptr);
        } else {
          // used ordered bin
          ordered_bins[group]->ConstructHistogram(leaf_idx,
                                                  gradients,
                                                  data_ptr);
        }
        // fixed hessian.
        for (int i = 0; i < num_bin; ++i) {
          data_ptr[i].sum_hessians = data_ptr[i].cnt * hessians[0];
        }
        OMP_LOOP_EX_END();
Guolin Ke's avatar
Guolin Ke committed
688
      }
689
      OMP_THROW_EX();
Guolin Ke's avatar
Guolin Ke committed
690
    }
691
  } else {
692
693
694
    if (!is_constant_hessian) {
      OMP_INIT_EX();
      #pragma omp parallel for schedule(static)
Guolin Ke's avatar
Guolin Ke committed
695
      for (int gi = 0; gi < num_used_group; ++gi) {
696
        OMP_LOOP_EX_BEGIN();
Guolin Ke's avatar
Guolin Ke committed
697
        int group = used_group[gi];
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
        // feature is not used
        auto data_ptr = hist_data + group_bin_boundaries_[group];
        const int num_bin = feature_groups_[group]->num_total_bin_;
        std::memset(data_ptr + 1, 0, (num_bin - 1) * sizeof(HistogramBinEntry));
        // construct histograms for smaller leaf
        if (ordered_bins[group] == nullptr) {
          // if not use ordered bin
          feature_groups_[group]->bin_data_->ConstructHistogram(
            num_data,
            ptr_ordered_grad,
            ptr_ordered_hess,
            data_ptr);
        } else {
          // used ordered bin
          ordered_bins[group]->ConstructHistogram(leaf_idx,
                                                  gradients,
                                                  hessians,
                                                  data_ptr);
        }
        OMP_LOOP_EX_END();
718
      }
719
720
721
722
      OMP_THROW_EX();
    } else {
      OMP_INIT_EX();
      #pragma omp parallel for schedule(static)
Guolin Ke's avatar
Guolin Ke committed
723
      for (int gi = 0; gi < num_used_group; ++gi) {
724
        OMP_LOOP_EX_BEGIN();
Guolin Ke's avatar
Guolin Ke committed
725
        int group = used_group[gi];
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
        // feature is not used
        auto data_ptr = hist_data + group_bin_boundaries_[group];
        const int num_bin = feature_groups_[group]->num_total_bin_;
        std::memset(data_ptr + 1, 0, (num_bin - 1) * sizeof(HistogramBinEntry));
        // construct histograms for smaller leaf
        if (ordered_bins[group] == nullptr) {
          // if not use ordered bin
          feature_groups_[group]->bin_data_->ConstructHistogram(
            num_data,
            ptr_ordered_grad,
            data_ptr);
        } else {
          // used ordered bin
          ordered_bins[group]->ConstructHistogram(leaf_idx,
                                                  gradients,
                                                  data_ptr);
        }
        // fixed hessian.
        for (int i = 0; i < num_bin; ++i) {
          data_ptr[i].sum_hessians = data_ptr[i].cnt * hessians[0];
        }
        OMP_LOOP_EX_END();
748
      }
749
      OMP_THROW_EX();
Guolin Ke's avatar
Guolin Ke committed
750
751
752
753
754
    }
  }
}

void Dataset::FixHistogram(int feature_idx, double sum_gradient, double sum_hessian, data_size_t num_data,
755
                           HistogramBinEntry* data) const {
Guolin Ke's avatar
Guolin Ke committed
756
757
758
759
760
761
  const int group = feature2group_[feature_idx];
  const int sub_feature = feature2subfeature_[feature_idx];
  const BinMapper* bin_mapper = feature_groups_[group]->bin_mappers_[sub_feature].get();
  const int default_bin = bin_mapper->GetDefaultBin();
  if (default_bin > 0) {
    const int num_bin = bin_mapper->num_bin();
762
763
764
    data[default_bin].sum_gradients = sum_gradient;
    data[default_bin].sum_hessians = sum_hessian;
    data[default_bin].cnt = num_data;
Guolin Ke's avatar
Guolin Ke committed
765
766
    for (int i = 0; i < num_bin; ++i) {
      if (i != default_bin) {
767
768
769
        data[default_bin].sum_gradients -= data[i].sum_gradients;
        data[default_bin].sum_hessians -= data[i].sum_hessians;
        data[default_bin].cnt -= data[i].cnt;
Guolin Ke's avatar
Guolin Ke committed
770
771
772
773
774
      }
    }
  }
}

Guolin Ke's avatar
Guolin Ke committed
775
}  // namespace LightGBM