dataset_loader.cpp 34.2 KB
Newer Older
Guolin Ke's avatar
Guolin Ke committed
1
2
3
4
5
6
7
8
9
10
11
#include <omp.h>

#include <LightGBM/utils/log.h>
#include <LightGBM/dataset_loader.h>
#include <LightGBM/feature.h>
#include <LightGBM/network.h>


namespace LightGBM {

DatasetLoader::DatasetLoader(const IOConfig& io_config, const PredictFunction& predict_fun)
Guolin Ke's avatar
Guolin Ke committed
12
  :io_config_(io_config), random_(io_config_.data_random_seed), predict_fun_(predict_fun) {
Guolin Ke's avatar
Guolin Ke committed
13
14
15
16
17
18
19

}

DatasetLoader::~DatasetLoader() {

}

Guolin Ke's avatar
Guolin Ke committed
20
void DatasetLoader::SetHeader(const char* filename) {
Guolin Ke's avatar
Guolin Ke committed
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
  TextReader<data_size_t> text_reader(filename, io_config_.has_header);
  std::unordered_map<std::string, int> name2idx;

  // get column names
  if (io_config_.has_header) {
    std::string first_line = text_reader.first_line();
    feature_names_ = Common::Split(first_line.c_str(), "\t ,");
    for (size_t i = 0; i < feature_names_.size(); ++i) {
      name2idx[feature_names_[i]] = static_cast<int>(i);
    }
  }
  std::string name_prefix("name:");

  // load label idx
  if (io_config_.label_column.size() > 0) {
    if (Common::StartsWith(io_config_.label_column, name_prefix)) {
      std::string name = io_config_.label_column.substr(name_prefix.size());
      if (name2idx.count(name) > 0) {
        label_idx_ = name2idx[name];
        Log::Info("Using column %s as label", name.c_str());
      } else {
        Log::Fatal("Could not find label column %s in data file", name.c_str());
      }
    } else {
      if (!Common::AtoiAndCheck(io_config_.label_column.c_str(), &label_idx_)) {
        Log::Fatal("label_column is not a number, \
                      if you want to use a column name, \
                      please add the prefix \"name:\" to the column name");
      }
      Log::Info("Using column number %d as label", label_idx_);
    }
  }
  if (feature_names_.size() > 0) {
    // erase label column name
    feature_names_.erase(feature_names_.begin() + label_idx_);
  }
  // load ignore columns
  if (io_config_.ignore_column.size() > 0) {
    if (Common::StartsWith(io_config_.ignore_column, name_prefix)) {
      std::string names = io_config_.ignore_column.substr(name_prefix.size());
      for (auto name : Common::Split(names.c_str(), ',')) {
        if (name2idx.count(name) > 0) {
          int tmp = name2idx[name];
          // skip for label column
          if (tmp > label_idx_) { tmp -= 1; }
          ignore_features_.emplace(tmp);
        } else {
          Log::Fatal("Could not find ignore column %s in data file", name.c_str());
        }
      }
    } else {
      for (auto token : Common::Split(io_config_.ignore_column.c_str(), ',')) {
        int tmp = 0;
        if (!Common::AtoiAndCheck(token.c_str(), &tmp)) {
          Log::Fatal("ignore_column is not a number, \
                        if you want to use a column name, \
                        please add the prefix \"name:\" to the column name");
        }
        // skip for label column
        if (tmp > label_idx_) { tmp -= 1; }
        ignore_features_.emplace(tmp);
      }
    }

  }

  // load weight idx
  if (io_config_.weight_column.size() > 0) {
    if (Common::StartsWith(io_config_.weight_column, name_prefix)) {
      std::string name = io_config_.weight_column.substr(name_prefix.size());
      if (name2idx.count(name) > 0) {
        weight_idx_ = name2idx[name];
        Log::Info("Using column %s as weight", name.c_str());
      } else {
        Log::Fatal("Could not find weight column %s in data file", name.c_str());
      }
    } else {
      if (!Common::AtoiAndCheck(io_config_.weight_column.c_str(), &weight_idx_)) {
        Log::Fatal("weight_column is not a number, \
                      if you want to use a column name, \
                      please add the prefix \"name:\" to the column name");
      }
      Log::Info("Using column number %d as weight", weight_idx_);
    }
    // skip for label column
    if (weight_idx_ > label_idx_) {
      weight_idx_ -= 1;
    }
    ignore_features_.emplace(weight_idx_);
  }

  if (io_config_.group_column.size() > 0) {
    if (Common::StartsWith(io_config_.group_column, name_prefix)) {
      std::string name = io_config_.group_column.substr(name_prefix.size());
      if (name2idx.count(name) > 0) {
        group_idx_ = name2idx[name];
        Log::Info("Using column %s as group/query id", name.c_str());
      } else {
        Log::Fatal("Could not find group/query column %s in data file", name.c_str());
      }
    } else {
      if (!Common::AtoiAndCheck(io_config_.group_column.c_str(), &group_idx_)) {
        Log::Fatal("group_column is not a number, \
                      if you want to use a column name, \
                      please add the prefix \"name:\" to the column name");
      }
      Log::Info("Using column number %d as group/query id", group_idx_);
    }
    // skip for label column
    if (group_idx_ > label_idx_) {
      group_idx_ -= 1;
    }
    ignore_features_.emplace(group_idx_);
  }
}

Dataset* DatasetLoader::LoadFromFile(const char* filename, int rank, int num_machines) {
  // don't support query id in data file when training in parallel
  if (num_machines > 1 && !io_config_.is_pre_partition) {
    if (group_idx_ > 0) {
      Log::Fatal("Using a query id without pre-partitioning the data file is not supported for parallel training. \
                  Please use an additional query file or pre-partition the data");
    }
  }
Guolin Ke's avatar
Guolin Ke committed
145
  auto parser = std::unique_ptr<Parser>(Parser::CreateParser(filename, io_config_.has_header, 0, label_idx_));
Guolin Ke's avatar
Guolin Ke committed
146
147
148
149
150
  if (parser == nullptr) {
    Log::Fatal("Could not recognize data format of %s", filename);
  }
  data_size_t num_global_data = 0;
  std::vector<data_size_t> used_data_indices;
Guolin Ke's avatar
Guolin Ke committed
151
  auto dataset = std::unique_ptr<Dataset>(new Dataset());
Guolin Ke's avatar
Guolin Ke committed
152
153
154
155
156
157
158
159
160
161
162
163
  dataset->data_filename_ = filename;
  dataset->num_class_ = io_config_.num_class;
  dataset->metadata_.Init(filename, dataset->num_class_);
  bool is_loading_from_binfile = CheckCanLoadFromBin(filename);
  if (!is_loading_from_binfile) {
    if (!io_config_.use_two_round_loading) {
      // read data to memory
      auto text_data = LoadTextDataToMemory(filename, dataset->metadata_, rank, num_machines,&num_global_data, &used_data_indices);
      dataset->num_data_ = static_cast<data_size_t>(text_data.size());
      // sample data
      auto sample_data = SampleTextDataFromMemory(text_data);
      // construct feature bin mappers
Guolin Ke's avatar
Guolin Ke committed
164
      ConstructBinMappersFromTextData(rank, num_machines, sample_data, parser.get(), dataset.get());
Guolin Ke's avatar
Guolin Ke committed
165
166
167
      // initialize label
      dataset->metadata_.Init(dataset->num_data_, io_config_.num_class, weight_idx_, group_idx_);
      // extract features
Guolin Ke's avatar
Guolin Ke committed
168
      ExtractFeaturesFromMemory(text_data, parser.get(), dataset.get());
Guolin Ke's avatar
Guolin Ke committed
169
170
171
172
173
174
175
176
177
178
      text_data.clear();
    } else {
      // sample data from file
      auto sample_data = SampleTextDataFromFile(filename, dataset->metadata_, rank, num_machines, &num_global_data, &used_data_indices);
      if (used_data_indices.size() > 0) {
        dataset->num_data_ = static_cast<data_size_t>(used_data_indices.size());
      } else {
        dataset->num_data_ = num_global_data;
      }
      // construct feature bin mappers
Guolin Ke's avatar
Guolin Ke committed
179
      ConstructBinMappersFromTextData(rank, num_machines, sample_data, parser.get(), dataset.get());
Guolin Ke's avatar
Guolin Ke committed
180
181
182
183
      // initialize label
      dataset->metadata_.Init(dataset->num_data_, dataset->num_class_, weight_idx_, group_idx_);

      // extract features
Guolin Ke's avatar
Guolin Ke committed
184
      ExtractFeaturesFromFile(filename, parser.get(), used_data_indices, dataset.get());
Guolin Ke's avatar
Guolin Ke committed
185
186
187
188
189
    }
  } else {
    // load data from binary file
    std::string bin_filename(filename);
    bin_filename.append(".bin");
Guolin Ke's avatar
Guolin Ke committed
190
    dataset.reset(LoadFromBinFile(bin_filename.c_str(), rank, num_machines));
Guolin Ke's avatar
Guolin Ke committed
191
192
193
194
  }
  // check meta data
  dataset->metadata_.CheckOrPartition(num_global_data, used_data_indices);
  // need to check training data
Guolin Ke's avatar
Guolin Ke committed
195
196
  CheckDataset(dataset.get());
  return dataset.release();
Guolin Ke's avatar
Guolin Ke committed
197
198
199
200
}



Guolin Ke's avatar
Guolin Ke committed
201
Dataset* DatasetLoader::LoadFromFileAlignWithOtherDataset(const char* filename, const Dataset* train_data) {
Guolin Ke's avatar
Guolin Ke committed
202
  auto parser = std::unique_ptr<Parser>(Parser::CreateParser(filename, io_config_.has_header, 0, label_idx_));
Guolin Ke's avatar
Guolin Ke committed
203
204
205
206
207
  if (parser == nullptr) {
    Log::Fatal("Could not recognize data format of %s", filename);
  }
  data_size_t num_global_data = 0;
  std::vector<data_size_t> used_data_indices;
Guolin Ke's avatar
Guolin Ke committed
208
  auto dataset = std::unique_ptr<Dataset>(new Dataset());
Guolin Ke's avatar
Guolin Ke committed
209
210
211
212
213
214
215
216
217
218
219
  dataset->data_filename_ = filename;
  dataset->num_class_ = io_config_.num_class;
  dataset->metadata_.Init(filename, dataset->num_class_);
  bool is_loading_from_binfile = CheckCanLoadFromBin(filename);
  if (!is_loading_from_binfile) {
    if (!io_config_.use_two_round_loading) {
      // read data in memory
      auto text_data = LoadTextDataToMemory(filename, dataset->metadata_, 0, 1, &num_global_data, &used_data_indices);
      dataset->num_data_ = static_cast<data_size_t>(text_data.size());
      // initialize label
      dataset->metadata_.Init(dataset->num_data_, dataset->num_class_, weight_idx_, group_idx_);
Guolin Ke's avatar
Guolin Ke committed
220
      dataset->CopyFeatureMapperFrom(train_data, io_config_.is_enable_sparse);
Guolin Ke's avatar
Guolin Ke committed
221
      // extract features
Guolin Ke's avatar
Guolin Ke committed
222
      ExtractFeaturesFromMemory(text_data, parser.get(), dataset.get());
Guolin Ke's avatar
Guolin Ke committed
223
224
225
226
227
228
229
230
      text_data.clear();
    } else {
      TextReader<data_size_t> text_reader(filename, io_config_.has_header);
      // Get number of lines of data file
      dataset->num_data_ = static_cast<data_size_t>(text_reader.CountLine());
      num_global_data = dataset->num_data_;
      // initialize label
      dataset->metadata_.Init(dataset->num_data_, dataset->num_class_, weight_idx_, group_idx_);
Guolin Ke's avatar
Guolin Ke committed
231
      dataset->CopyFeatureMapperFrom(train_data, io_config_.is_enable_sparse);
Guolin Ke's avatar
Guolin Ke committed
232
      // extract features
Guolin Ke's avatar
Guolin Ke committed
233
      ExtractFeaturesFromFile(filename, parser.get(), used_data_indices, dataset.get());
Guolin Ke's avatar
Guolin Ke committed
234
235
236
237
238
    }
  } else {
    // load data from binary file
    std::string bin_filename(filename);
    bin_filename.append(".bin");
Guolin Ke's avatar
Guolin Ke committed
239
    dataset.reset(LoadFromBinFile(bin_filename.c_str(), 0, 1));
Guolin Ke's avatar
Guolin Ke committed
240
241
242
243
  }
  // not need to check validation data
  // check meta data
  dataset->metadata_.CheckOrPartition(num_global_data, used_data_indices);
Guolin Ke's avatar
Guolin Ke committed
244
  return dataset.release();
Guolin Ke's avatar
Guolin Ke committed
245
246
247
}

Dataset* DatasetLoader::LoadFromBinFile(const char* bin_filename, int rank, int num_machines) {
Guolin Ke's avatar
Guolin Ke committed
248
  auto dataset = std::unique_ptr<Dataset>(new Dataset());
Guolin Ke's avatar
Guolin Ke committed
249
250
251
252
253
254
255
256
257
258
259
260
261
  FILE* file;
#ifdef _MSC_VER
  fopen_s(&file, bin_filename, "rb");
#else
  file = fopen(bin_filename, "rb");
#endif

  if (file == NULL) {
    Log::Fatal("Could not read binary data from %s", bin_filename);
  }

  // buffer to read binary file
  size_t buffer_size = 16 * 1024 * 1024;
Guolin Ke's avatar
Guolin Ke committed
262
  auto buffer = std::vector<char>(buffer_size);
Guolin Ke's avatar
Guolin Ke committed
263
264

  // read size of header
Guolin Ke's avatar
Guolin Ke committed
265
  size_t read_cnt = fread(buffer.data(), sizeof(size_t), 1, file);
Guolin Ke's avatar
Guolin Ke committed
266
267
268
269
270

  if (read_cnt != 1) {
    Log::Fatal("Binary file error: header has the wrong size");
  }

Guolin Ke's avatar
Guolin Ke committed
271
  size_t size_of_head = *(reinterpret_cast<size_t*>(buffer.data()));
Guolin Ke's avatar
Guolin Ke committed
272
273
274
275

  // re-allocmate space if not enough
  if (size_of_head > buffer_size) {
    buffer_size = size_of_head;
Guolin Ke's avatar
Guolin Ke committed
276
    buffer.resize(buffer_size);
Guolin Ke's avatar
Guolin Ke committed
277
278
  }
  // read header
Guolin Ke's avatar
Guolin Ke committed
279
  read_cnt = fread(buffer.data(), 1, size_of_head, file);
Guolin Ke's avatar
Guolin Ke committed
280
281
282
283
284

  if (read_cnt != size_of_head) {
    Log::Fatal("Binary file error: header is incorrect");
  }
  // get header
Guolin Ke's avatar
Guolin Ke committed
285
  const char* mem_ptr = buffer.data();
Guolin Ke's avatar
Guolin Ke committed
286
287
  dataset->num_data_ = *(reinterpret_cast<const data_size_t*>(mem_ptr));
  mem_ptr += sizeof(dataset->num_data_);
Guolin Ke's avatar
Guolin Ke committed
288
289
  dataset->num_class_ = *(reinterpret_cast<const int*>(mem_ptr));
  mem_ptr += sizeof(dataset->num_class_);
Guolin Ke's avatar
Guolin Ke committed
290
291
292
293
294
295
296
297
298
299
300
301
302
  dataset->num_features_ = *(reinterpret_cast<const int*>(mem_ptr));
  mem_ptr += sizeof(dataset->num_features_);
  dataset->num_total_features_ = *(reinterpret_cast<const int*>(mem_ptr));
  mem_ptr += sizeof(dataset->num_total_features_);
  size_t num_used_feature_map = *(reinterpret_cast<const size_t*>(mem_ptr));
  mem_ptr += sizeof(num_used_feature_map);
  const int* tmp_feature_map = reinterpret_cast<const int*>(mem_ptr);
  dataset->used_feature_map_.clear();
  for (size_t i = 0; i < num_used_feature_map; ++i) {
    dataset->used_feature_map_.push_back(tmp_feature_map[i]);
  }
  mem_ptr += sizeof(int) * num_used_feature_map;
  // get feature names
Guolin Ke's avatar
Guolin Ke committed
303
  dataset->feature_names_.clear();
Guolin Ke's avatar
Guolin Ke committed
304
305
306
307
308
309
310
311
312
313
  // write feature names
  for (int i = 0; i < dataset->num_total_features_; ++i) {
    int str_len = *(reinterpret_cast<const int*>(mem_ptr));
    mem_ptr += sizeof(int);
    std::stringstream str_buf;
    for (int j = 0; j < str_len; ++j) {
      char tmp_char = *(reinterpret_cast<const char*>(mem_ptr));
      mem_ptr += sizeof(char);
      str_buf << tmp_char;
    }
Guolin Ke's avatar
Guolin Ke committed
314
    dataset->feature_names_.emplace_back(str_buf.str());
Guolin Ke's avatar
Guolin Ke committed
315
316
317
  }

  // read size of meta data
Guolin Ke's avatar
Guolin Ke committed
318
  read_cnt = fread(buffer.data(), sizeof(size_t), 1, file);
Guolin Ke's avatar
Guolin Ke committed
319
320
321
322
323

  if (read_cnt != 1) {
    Log::Fatal("Binary file error: meta data has the wrong size");
  }

Guolin Ke's avatar
Guolin Ke committed
324
  size_t size_of_metadata = *(reinterpret_cast<size_t*>(buffer.data()));
Guolin Ke's avatar
Guolin Ke committed
325
326
327
328

  // re-allocate space if not enough
  if (size_of_metadata > buffer_size) {
    buffer_size = size_of_metadata;
Guolin Ke's avatar
Guolin Ke committed
329
    buffer.resize(buffer_size);
Guolin Ke's avatar
Guolin Ke committed
330
331
  }
  //  read meta data
Guolin Ke's avatar
Guolin Ke committed
332
  read_cnt = fread(buffer.data(), 1, size_of_metadata, file);
Guolin Ke's avatar
Guolin Ke committed
333
334
335
336
337

  if (read_cnt != size_of_metadata) {
    Log::Fatal("Binary file error: meta data is incorrect");
  }
  // load meta data
Guolin Ke's avatar
Guolin Ke committed
338
  dataset->metadata_.LoadFromMemory(buffer.data());
Guolin Ke's avatar
Guolin Ke committed
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379

  std::vector<data_size_t> used_data_indices;
  data_size_t num_global_data = dataset->num_data_;
  // sample local used data if need to partition
  if (num_machines > 1 && !io_config_.is_pre_partition) {
    const data_size_t* query_boundaries = dataset->metadata_.query_boundaries();
    if (query_boundaries == nullptr) {
      // if not contain query file, minimal sample unit is one record
      for (data_size_t i = 0; i < dataset->num_data_; ++i) {
        if (random_.NextInt(0, num_machines) == rank) {
          used_data_indices.push_back(i);
        }
      }
    } else {
      // if contain query file, minimal sample unit is one query
      data_size_t num_queries = dataset->metadata_.num_queries();
      data_size_t qid = -1;
      bool is_query_used = false;
      for (data_size_t i = 0; i < dataset->num_data_; ++i) {
        if (qid >= num_queries) {
          Log::Fatal("Current query exceeds the range of the query file, please ensure the query file is correct");
        }
        if (i >= query_boundaries[qid + 1]) {
          // if is new query
          is_query_used = false;
          if (random_.NextInt(0, num_machines) == rank) {
            is_query_used = true;
          }
          ++qid;
        }
        if (is_query_used) {
          used_data_indices.push_back(i);
        }
      }
    }
    dataset->num_data_ = static_cast<data_size_t>(used_data_indices.size());
  }
  dataset->metadata_.PartitionLabel(used_data_indices);
  // read feature data
  for (int i = 0; i < dataset->num_features_; ++i) {
    // read feature size
Guolin Ke's avatar
Guolin Ke committed
380
    read_cnt = fread(buffer.data(), sizeof(size_t), 1, file);
Guolin Ke's avatar
Guolin Ke committed
381
382
383
    if (read_cnt != 1) {
      Log::Fatal("Binary file error: feature %d has the wrong size", i);
    }
Guolin Ke's avatar
Guolin Ke committed
384
    size_t size_of_feature = *(reinterpret_cast<size_t*>(buffer.data()));
Guolin Ke's avatar
Guolin Ke committed
385
386
387
    // re-allocate space if not enough
    if (size_of_feature > buffer_size) {
      buffer_size = size_of_feature;
Guolin Ke's avatar
Guolin Ke committed
388
      buffer.resize(buffer_size);
Guolin Ke's avatar
Guolin Ke committed
389
390
    }

Guolin Ke's avatar
Guolin Ke committed
391
    read_cnt = fread(buffer.data(), 1, size_of_feature, file);
Guolin Ke's avatar
Guolin Ke committed
392
393
394
395

    if (read_cnt != size_of_feature) {
      Log::Fatal("Binary file error: feature %d is incorrect, read count: %d", i, read_cnt);
    }
Guolin Ke's avatar
Guolin Ke committed
396
397
398
399
400
    dataset->features_.emplace_back(std::unique_ptr<Feature>(
      new Feature(buffer.data(), 
        num_global_data, 
        used_data_indices)
    ));
Guolin Ke's avatar
Guolin Ke committed
401
  }
Guolin Ke's avatar
Guolin Ke committed
402
  dataset->features_.shrink_to_fit();
Guolin Ke's avatar
Guolin Ke committed
403
404
  fclose(file);
  dataset->is_loading_from_binfile_ = true;
Guolin Ke's avatar
Guolin Ke committed
405
  return dataset.release();
Guolin Ke's avatar
Guolin Ke committed
406
407
}

Guolin Ke's avatar
Guolin Ke committed
408
Dataset* DatasetLoader::CostructFromSampleData(std::vector<std::vector<double>>& sample_values, size_t total_sample_size, data_size_t num_data) {
Guolin Ke's avatar
Guolin Ke committed
409
  std::vector<std::unique_ptr<BinMapper>> bin_mappers(sample_values.size());
Guolin Ke's avatar
Guolin Ke committed
410
411
#pragma omp parallel for schedule(guided)
  for (int i = 0; i < static_cast<int>(sample_values.size()); ++i) {
Guolin Ke's avatar
Guolin Ke committed
412
    bin_mappers[i].reset(new BinMapper());
Guolin Ke's avatar
Guolin Ke committed
413
    bin_mappers[i]->FindBin(&sample_values[i], total_sample_size, io_config_.max_bin);
Guolin Ke's avatar
Guolin Ke committed
414
415
  }

Guolin Ke's avatar
Guolin Ke committed
416
  auto dataset = std::unique_ptr<Dataset>(new Dataset());
Guolin Ke's avatar
Guolin Ke committed
417
  dataset->num_class_ = io_config_.num_class;
Guolin Ke's avatar
Guolin Ke committed
418
419
420
421
422
423
424
425
426
427
428
  dataset->features_.clear();
  dataset->num_data_ = num_data;
  // -1 means doesn't use this feature
  dataset->used_feature_map_ = std::vector<int>(bin_mappers.size(), -1);
  dataset->num_total_features_ = static_cast<int>(bin_mappers.size());

  for (size_t i = 0; i < bin_mappers.size(); ++i) {
    if (!bin_mappers[i]->is_trival()) {
      // map real feature index to used feature index
      dataset->used_feature_map_[i] = static_cast<int>(dataset->features_.size());
      // push new feature
Guolin Ke's avatar
Guolin Ke committed
429
430
431
432
433
434
      dataset->features_.emplace_back(std::unique_ptr<Feature>(
        new Feature(static_cast<int>(i),
          bin_mappers[i].release(),
          dataset->num_data_, 
          io_config_.is_enable_sparse)
        ));
Guolin Ke's avatar
Guolin Ke committed
435
436
437
438
439
    } else {
      // if feature is trival(only 1 bin), free spaces
      Log::Warning("Ignoring Column_%d , only has one value", i);
    }
  }
Guolin Ke's avatar
Guolin Ke committed
440
  dataset->features_.shrink_to_fit();
Guolin Ke's avatar
Guolin Ke committed
441
442
443
444
445
446
447
448
449
  // fill feature_names_ if not header
  if (feature_names_.size() <= 0) {
    for (int i = 0; i < dataset->num_total_features_; ++i) {
      std::stringstream str_buf;
      str_buf << "Column_" << i;
      feature_names_.push_back(str_buf.str());
    }
  }
  dataset->feature_names_ = feature_names_;
450
  dataset->num_features_ = static_cast<int>(dataset->features_.size());
Guolin Ke's avatar
Guolin Ke committed
451
  dataset->metadata_.Init(dataset->num_data_, dataset->num_class_, NO_SPECIFIC, NO_SPECIFIC);
Guolin Ke's avatar
Guolin Ke committed
452
  return dataset.release();
Guolin Ke's avatar
Guolin Ke committed
453
}
Guolin Ke's avatar
Guolin Ke committed
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514


// ---- private functions ----

void DatasetLoader::CheckDataset(const Dataset* dataset) {
  if (dataset->num_data_ <= 0) {
    Log::Fatal("Data file %s is empty", dataset->data_filename_);
  }
  if (dataset->features_.size() <= 0) {
    Log::Fatal("No usable features in data file %s", dataset->data_filename_);
  }
}

std::vector<std::string> DatasetLoader::LoadTextDataToMemory(const char* filename, const Metadata& metadata,
  int rank, int num_machines, int* num_global_data, 
  std::vector<data_size_t>* used_data_indices) {
  TextReader<data_size_t> text_reader(filename, io_config_.has_header);
  used_data_indices->clear();
  if (num_machines == 1 || io_config_.is_pre_partition) {
    // read all lines
    *num_global_data = text_reader.ReadAllLines();
  } else {  // need partition data
            // get query data
    const data_size_t* query_boundaries = metadata.query_boundaries();

    if (query_boundaries == nullptr) {
      // if not contain query data, minimal sample unit is one record
      *num_global_data = text_reader.ReadAndFilterLines([this, rank, num_machines](data_size_t) {
        if (random_.NextInt(0, num_machines) == rank) {
          return true;
        } else {
          return false;
        }
      }, used_data_indices);
    } else {
      // if contain query data, minimal sample unit is one query
      data_size_t num_queries = metadata.num_queries();
      data_size_t qid = -1;
      bool is_query_used = false;
      *num_global_data = text_reader.ReadAndFilterLines(
        [this, rank, num_machines, &qid, &query_boundaries, &is_query_used, num_queries]
      (data_size_t line_idx) {
        if (qid >= num_queries) {
          Log::Fatal("Current query exceeds the range of the query file, please ensure the query file is correct");
        }
        if (line_idx >= query_boundaries[qid + 1]) {
          // if is new query
          is_query_used = false;
          if (random_.NextInt(0, num_machines) == rank) {
            is_query_used = true;
          }
          ++qid;
        }
        return is_query_used;
      }, used_data_indices);
    }
  }
  return std::move(text_reader.Lines());
}

std::vector<std::string> DatasetLoader::SampleTextDataFromMemory(const std::vector<std::string>& data) {
515
516
517
  int sample_cnt = io_config_.bin_construct_sample_cnt;
  if (static_cast<size_t>(sample_cnt) > data.size()) {
    sample_cnt = static_cast<int>(data.size());
518
  }
519
  auto sample_indices = random_.Sample(static_cast<int>(data.size()), sample_cnt);
Guolin Ke's avatar
Guolin Ke committed
520
  std::vector<std::string> out(sample_indices.size());
Guolin Ke's avatar
Guolin Ke committed
521
522
  for (size_t i = 0; i < sample_indices.size(); ++i) {
    const size_t idx = sample_indices[i];
Guolin Ke's avatar
Guolin Ke committed
523
    out[i] = data[idx];
Guolin Ke's avatar
Guolin Ke committed
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
  }
  return out;
}

std::vector<std::string> DatasetLoader::SampleTextDataFromFile(const char* filename, const Metadata& metadata, int rank, int num_machines, int* num_global_data, std::vector<data_size_t>* used_data_indices) {
  const data_size_t sample_cnt = static_cast<data_size_t>(io_config_.bin_construct_sample_cnt);
  TextReader<data_size_t> text_reader(filename, io_config_.has_header);
  std::vector<std::string> out_data;
  if (num_machines == 1 || io_config_.is_pre_partition) {
    *num_global_data = static_cast<data_size_t>(text_reader.SampleFromFile(random_, sample_cnt, &out_data));
  } else {  // need partition data
            // get query data
    const data_size_t* query_boundaries = metadata.query_boundaries();
    if (query_boundaries == nullptr) {
      // if not contain query file, minimal sample unit is one record
      *num_global_data = text_reader.SampleAndFilterFromFile([this, rank, num_machines]
      (data_size_t) {
        if (random_.NextInt(0, num_machines) == rank) {
          return true;
        } else {
          return false;
        }
      }, used_data_indices, random_, sample_cnt, &out_data);
    } else {
      // if contain query file, minimal sample unit is one query
      data_size_t num_queries = metadata.num_queries();
      data_size_t qid = -1;
      bool is_query_used = false;
      *num_global_data = text_reader.SampleAndFilterFromFile(
        [this, rank, num_machines, &qid, &query_boundaries, &is_query_used, num_queries]
      (data_size_t line_idx) {
        if (qid >= num_queries) {
          Log::Fatal("Query id exceeds the range of the query file, \
                      please ensure the query file is correct");
        }
        if (line_idx >= query_boundaries[qid + 1]) {
          // if is new query
          is_query_used = false;
          if (random_.NextInt(0, num_machines) == rank) {
            is_query_used = true;
          }
          ++qid;
        }
        return is_query_used;
      }, used_data_indices, random_, sample_cnt, &out_data);
    }
  }
  return out_data;
}

void DatasetLoader::ConstructBinMappersFromTextData(int rank, int num_machines, const std::vector<std::string>& sample_data, const Parser* parser, Dataset* dataset) {
  // sample_values[i][j], means the value of j-th sample on i-th feature
  std::vector<std::vector<double>> sample_values;
  // temp buffer for one line features and label
  std::vector<std::pair<int, double>> oneline_features;
  double label;
  for (size_t i = 0; i < sample_data.size(); ++i) {
    oneline_features.clear();
    // parse features
    parser->ParseOneLine(sample_data[i].c_str(), &oneline_features, &label);
    for (std::pair<int, double>& inner_data : oneline_features) {
Guolin Ke's avatar
Guolin Ke committed
585
586
587
588
589
590
591
      if (std::fabs(inner_data.second) > 1e-15) {
        if (static_cast<size_t>(inner_data.first) >= sample_values.size()) {
          // if need expand feature set
          size_t need_size = inner_data.first - sample_values.size() + 1;
          for (size_t j = 0; j < need_size; ++j) {
            sample_values.emplace_back();
          }
Guolin Ke's avatar
Guolin Ke committed
592
        }
Guolin Ke's avatar
Guolin Ke committed
593
        sample_values[inner_data.first].push_back(inner_data.second);
Guolin Ke's avatar
Guolin Ke committed
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
      }
    }
  }

  dataset->features_.clear();

  // -1 means doesn't use this feature
  dataset->used_feature_map_ = std::vector<int>(sample_values.size(), -1);
  dataset->num_total_features_ = static_cast<int>(sample_values.size());

  // check the range of label_idx, weight_idx and group_idx
  CHECK(label_idx_ >= 0 && label_idx_ <= dataset->num_total_features_);
  CHECK(weight_idx_ < 0 || weight_idx_ < dataset->num_total_features_);
  CHECK(group_idx_ < 0 || group_idx_ < dataset->num_total_features_);

  // fill feature_names_ if not header
  if (feature_names_.size() <= 0) {
    for (int i = 0; i < dataset->num_total_features_; ++i) {
      std::stringstream str_buf;
      str_buf << "Column_" << i;
      feature_names_.push_back(str_buf.str());
    }
  }
  dataset->feature_names_ = feature_names_;
  // start find bins
  if (num_machines == 1) {
Guolin Ke's avatar
Guolin Ke committed
620
    std::vector<std::unique_ptr<BinMapper>> bin_mappers(sample_values.size());
Guolin Ke's avatar
Guolin Ke committed
621
622
623
624
    // if only one machine, find bin locally
#pragma omp parallel for schedule(guided)
    for (int i = 0; i < static_cast<int>(sample_values.size()); ++i) {
      if (ignore_features_.count(i) > 0) {
Guolin Ke's avatar
Guolin Ke committed
625
        bin_mappers[i].reset(nullptr);
Guolin Ke's avatar
Guolin Ke committed
626
627
        continue;
      }
Guolin Ke's avatar
Guolin Ke committed
628
      bin_mappers[i].reset(new BinMapper());
Guolin Ke's avatar
Guolin Ke committed
629
      bin_mappers[i]->FindBin(&sample_values[i], sample_data.size(), io_config_.max_bin);
Guolin Ke's avatar
Guolin Ke committed
630
631
632
633
634
635
636
637
638
    }

    for (size_t i = 0; i < sample_values.size(); ++i) {
      if (bin_mappers[i] == nullptr) {
        Log::Warning("Ignoring feature %s", feature_names_[i].c_str());
      } else if (!bin_mappers[i]->is_trival()) {
        // map real feature index to used feature index
        dataset->used_feature_map_[i] = static_cast<int>(dataset->features_.size());
        // push new feature
Guolin Ke's avatar
Guolin Ke committed
639
640
641
642
643
644
        dataset->features_.emplace_back(std::unique_ptr<Feature>(
          new Feature(static_cast<int>(i), 
            bin_mappers[i].release(),
            dataset->num_data_,
            io_config_.is_enable_sparse)
          ));
Guolin Ke's avatar
Guolin Ke committed
645
646
647
648
649
650
651
652
653
654
655
      } else {
        // if feature is trival(only 1 bin), free spaces
        Log::Warning("Ignoring feature %s, only has one value", feature_names_[i].c_str());
      }
    }
  } else {
    // if have multi-machines, need find bin distributed
    // different machines will find bin for different features

    // start and len will store the process feature indices for different machines
    // machine i will find bins for features in [ strat[i], start[i] + len[i] )
Guolin Ke's avatar
Guolin Ke committed
656
657
    std::vector<int> start(num_machines);
    std::vector<int> len(num_machines);
Guolin Ke's avatar
Guolin Ke committed
658
659
660
661
662
663
    int total_num_feature = static_cast<int>(sample_values.size());
    int step = (total_num_feature + num_machines - 1) / num_machines;
    if (step < 1) { step = 1; }

    start[0] = 0;
    for (int i = 0; i < num_machines - 1; ++i) {
Guolin Ke's avatar
Guolin Ke committed
664
      len[i] = std::min(step, total_num_feature - start[i]);
Guolin Ke's avatar
Guolin Ke committed
665
666
667
668
669
670
671
      start[i + 1] = start[i] + len[i];
    }
    len[num_machines - 1] = total_num_feature - start[num_machines - 1];
    // get size of bin mapper with max_bin_ size
    int type_size = BinMapper::SizeForSpecificBin(io_config_.max_bin);
    // since sizes of different feature may not be same, we expand all bin mapper to type_size
    int buffer_size = type_size * total_num_feature;
Guolin Ke's avatar
Guolin Ke committed
672
673
    auto input_buffer = std::vector<char>(buffer_size);
    auto output_buffer = std::vector<char>(buffer_size);
Guolin Ke's avatar
Guolin Ke committed
674
675
676
677

    // find local feature bins and copy to buffer
#pragma omp parallel for schedule(guided)
    for (int i = 0; i < len[rank]; ++i) {
Guolin Ke's avatar
Guolin Ke committed
678
679
680
      BinMapper bin_mapper;
      bin_mapper.FindBin(&sample_values[start[rank] + i], sample_data.size(), io_config_.max_bin);
      bin_mapper.CopyTo(input_buffer.data() + i * type_size);
Guolin Ke's avatar
Guolin Ke committed
681
682
683
684
685
686
687
    }
    // convert to binary size
    for (int i = 0; i < num_machines; ++i) {
      start[i] *= type_size;
      len[i] *= type_size;
    }
    // gather global feature bin mappers
Guolin Ke's avatar
Guolin Ke committed
688
    Network::Allgather(input_buffer.data(), buffer_size, start.data(), len.data(), output_buffer.data());
Guolin Ke's avatar
Guolin Ke committed
689
690
691
692
693
694
    // restore features bins from buffer
    for (int i = 0; i < total_num_feature; ++i) {
      if (ignore_features_.count(i) > 0) {
        Log::Warning("Ignoring feature %s", feature_names_[i].c_str());
        continue;
      }
Guolin Ke's avatar
Guolin Ke committed
695
696
      auto bin_mapper = std::unique_ptr<BinMapper>(new BinMapper());
      bin_mapper->CopyFrom(output_buffer.data() + i * type_size);
Guolin Ke's avatar
Guolin Ke committed
697
698
      if (!bin_mapper->is_trival()) {
        dataset->used_feature_map_[i] = static_cast<int>(dataset->features_.size());
Guolin Ke's avatar
Guolin Ke committed
699
700
701
702
703
704
        dataset->features_.emplace_back(std::unique_ptr<Feature>(
          new Feature(static_cast<int>(i),
            bin_mapper.release(),
            dataset->num_data_,
            io_config_.is_enable_sparse)
          ));
Guolin Ke's avatar
Guolin Ke committed
705
706
707
708
709
      } else {
        Log::Warning("Ignoring feature %s, only has one value", feature_names_[i].c_str());
      }
    }
  }
Guolin Ke's avatar
Guolin Ke committed
710
  dataset->features_.shrink_to_fit();
Guolin Ke's avatar
Guolin Ke committed
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
  dataset->num_features_ = static_cast<int>(dataset->features_.size());
}

/*! \brief Extract local features from memory */
void DatasetLoader::ExtractFeaturesFromMemory(std::vector<std::string>& text_data, const Parser* parser, Dataset* dataset) {
  std::vector<std::pair<int, double>> oneline_features;
  double tmp_label = 0.0f;
  if (predict_fun_ == nullptr) {
    // if doesn't need to prediction with initial model
#pragma omp parallel for schedule(guided) private(oneline_features) firstprivate(tmp_label)
    for (data_size_t i = 0; i < dataset->num_data_; ++i) {
      const int tid = omp_get_thread_num();
      oneline_features.clear();
      // parser
      parser->ParseOneLine(text_data[i].c_str(), &oneline_features, &tmp_label);
      // set label
      dataset->metadata_.SetLabelAt(i, static_cast<float>(tmp_label));
      // free processed line:
      text_data[i].clear();
      // shrink_to_fit will be very slow in linux, and seems not free memory, disable for now
      // text_reader_->Lines()[i].shrink_to_fit();
      // push data
      for (auto& inner_data : oneline_features) {
734
        if (inner_data.first >= dataset->num_total_features_) { continue; }
Guolin Ke's avatar
Guolin Ke committed
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
        int feature_idx = dataset->used_feature_map_[inner_data.first];
        if (feature_idx >= 0) {
          // if is used feature
          dataset->features_[feature_idx]->PushData(tid, i, inner_data.second);
        } else {
          if (inner_data.first == weight_idx_) {
            dataset->metadata_.SetWeightAt(i, static_cast<float>(inner_data.second));
          } else if (inner_data.first == group_idx_) {
            dataset->metadata_.SetQueryAt(i, static_cast<data_size_t>(inner_data.second));
          }
        }
      }
    }
  } else {
    // if need to prediction with initial model
Guolin Ke's avatar
Guolin Ke committed
750
    std::vector<score_t> init_score(dataset->num_data_ * dataset->num_class_);
Guolin Ke's avatar
Guolin Ke committed
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
#pragma omp parallel for schedule(guided) private(oneline_features) firstprivate(tmp_label)
    for (data_size_t i = 0; i < dataset->num_data_; ++i) {
      const int tid = omp_get_thread_num();
      oneline_features.clear();
      // parser
      parser->ParseOneLine(text_data[i].c_str(), &oneline_features, &tmp_label);
      // set initial score
      std::vector<double> oneline_init_score = predict_fun_(oneline_features);
      for (int k = 0; k < dataset->num_class_; ++k) {
        init_score[k * dataset->num_data_ + i] = static_cast<float>(oneline_init_score[k]);
      }
      // set label
      dataset->metadata_.SetLabelAt(i, static_cast<float>(tmp_label));
      // free processed line:
      text_data[i].clear();
      // shrink_to_fit will be very slow in linux, and seems not free memory, disable for now
      // text_reader_->Lines()[i].shrink_to_fit();
      // push data
      for (auto& inner_data : oneline_features) {
770
        if (inner_data.first >= dataset->num_total_features_) { continue; }
Guolin Ke's avatar
Guolin Ke committed
771
772
773
774
775
776
777
778
779
780
781
782
783
784
        int feature_idx = dataset->used_feature_map_[inner_data.first];
        if (feature_idx >= 0) {
          // if is used feature
          dataset->features_[feature_idx]->PushData(tid, i, inner_data.second);
        } else {
          if (inner_data.first == weight_idx_) {
            dataset->metadata_.SetWeightAt(i, static_cast<float>(inner_data.second));
          } else if (inner_data.first == group_idx_) {
            dataset->metadata_.SetQueryAt(i, static_cast<data_size_t>(inner_data.second));
          }
        }
      }
    }
    // metadata_ will manage space of init_score
Guolin Ke's avatar
Guolin Ke committed
785
    dataset->metadata_.SetInitScore(init_score.data(), dataset->num_data_ * dataset->num_class_);
Guolin Ke's avatar
Guolin Ke committed
786
  }
Guolin Ke's avatar
Guolin Ke committed
787
  dataset->FinishLoad();
Guolin Ke's avatar
Guolin Ke committed
788
789
790
791
792
793
  // text data can be free after loaded feature values
  text_data.clear();
}

/*! \brief Extract local features from file */
void DatasetLoader::ExtractFeaturesFromFile(const char* filename, const Parser* parser, const std::vector<data_size_t>& used_data_indices, Dataset* dataset) {
Guolin Ke's avatar
Guolin Ke committed
794
  std::vector<score_t> init_score;
Guolin Ke's avatar
Guolin Ke committed
795
  if (predict_fun_ != nullptr) {
Guolin Ke's avatar
Guolin Ke committed
796
    init_score = std::vector<score_t>(dataset->num_data_ * dataset->num_class_);
Guolin Ke's avatar
Guolin Ke committed
797
798
799
800
801
802
803
804
805
806
807
808
809
  }
  std::function<void(data_size_t, const std::vector<std::string>&)> process_fun =
    [this, &init_score, &parser, &dataset]
  (data_size_t start_idx, const std::vector<std::string>& lines) {
    std::vector<std::pair<int, double>> oneline_features;
    double tmp_label = 0.0f;
#pragma omp parallel for schedule(static) private(oneline_features) firstprivate(tmp_label)
    for (data_size_t i = 0; i < static_cast<data_size_t>(lines.size()); ++i) {
      const int tid = omp_get_thread_num();
      oneline_features.clear();
      // parser
      parser->ParseOneLine(lines[i].c_str(), &oneline_features, &tmp_label);
      // set initial score
Guolin Ke's avatar
Guolin Ke committed
810
      if (init_score.size() > 0) {
Guolin Ke's avatar
Guolin Ke committed
811
812
813
814
815
816
817
818
819
        std::vector<double> oneline_init_score = predict_fun_(oneline_features);
        for (int k = 0; k < dataset->num_class_; ++k) {
          init_score[k * dataset->num_data_ + start_idx + i] = static_cast<float>(oneline_init_score[k]);
        }
      }
      // set label
      dataset->metadata_.SetLabelAt(start_idx + i, static_cast<float>(tmp_label));
      // push data
      for (auto& inner_data : oneline_features) {
820
        if (inner_data.first >= dataset->num_total_features_) { continue; }
Guolin Ke's avatar
Guolin Ke committed
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
        int feature_idx = dataset->used_feature_map_[inner_data.first];
        if (feature_idx >= 0) {
          // if is used feature
          dataset->features_[feature_idx]->PushData(tid, start_idx + i, inner_data.second);
        } else {
          if (inner_data.first == weight_idx_) {
            dataset->metadata_.SetWeightAt(start_idx + i, static_cast<float>(inner_data.second));
          } else if (inner_data.first == group_idx_) {
            dataset->metadata_.SetQueryAt(start_idx + i, static_cast<data_size_t>(inner_data.second));
          }
        }
      }
    }
  };
  TextReader<data_size_t> text_reader(filename, io_config_.has_header);
  if (used_data_indices.size() > 0) {
    // only need part of data
    text_reader.ReadPartAndProcessParallel(used_data_indices, process_fun);
  } else {
    // need full data
    text_reader.ReadAllAndProcessParallel(process_fun);
  }

  // metadata_ will manage space of init_score
Guolin Ke's avatar
Guolin Ke committed
845
846
  if (init_score.size() > 0) {
    dataset->metadata_.SetInitScore(init_score.data(), dataset->num_data_ * dataset->num_class_);
Guolin Ke's avatar
Guolin Ke committed
847
  }
Guolin Ke's avatar
Guolin Ke committed
848
  dataset->FinishLoad();
Guolin Ke's avatar
Guolin Ke committed
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
}

/*! \brief Check can load from binary file */
bool DatasetLoader::CheckCanLoadFromBin(const char* filename) {
  std::string bin_filename(filename);
  bin_filename.append(".bin");

  FILE* file;

#ifdef _MSC_VER
  fopen_s(&file, bin_filename.c_str(), "rb");
#else
  file = fopen(bin_filename.c_str(), "rb");
#endif
  if (file == NULL) {
    return false;
  } else {
    fclose(file);
    return true;
  }
}

}