#include #include #include #include #include #include #include #include #include #include #include #include #include #include #include namespace LightGBM { class Booster { public: explicit Booster(const char* filename): boosting_(Boosting::CreateBoosting(filename)) { } Booster(const Dataset* train_data, std::vector valid_data, std::vector valid_names, const char* parameters) :train_data_(train_data), valid_datas_(valid_data) { config_.LoadFromString(parameters); // create boosting if (config_.io_config.input_model.size() > 0) { Log::Warning("continued train from model is not support for c_api, \ please use continued train with input score"); } boosting_ = Boosting::CreateBoosting(config_.boosting_type, ""); // create objective function objective_fun_ = ObjectiveFunction::CreateObjectiveFunction(config_.objective_type, config_.objective_config); // create training metric if (config_.boosting_config->is_provide_training_metric) { for (auto metric_type : config_.metric_types) { Metric* metric = Metric::CreateMetric(metric_type, config_.metric_config); if (metric == nullptr) { continue; } metric->Init("training", train_data_->metadata(), train_data_->num_data()); train_metric_.push_back(metric); } } // add metric for validation data for (size_t i = 0; i < valid_datas_.size(); ++i) { valid_metrics_.emplace_back(); for (auto metric_type : config_.metric_types) { Metric* metric = Metric::CreateMetric(metric_type, config_.metric_config); if (metric == nullptr) { continue; } metric->Init(valid_names[i].c_str(), valid_datas_[i]->metadata(), valid_datas_[i]->num_data()); valid_metrics_.back().push_back(metric); } } // initialize the objective function objective_fun_->Init(train_data_->metadata(), train_data_->num_data()); // initialize the boosting boosting_->Init(config_.boosting_config, train_data_, objective_fun_, ConstPtrInVectorWarpper(train_metric_)); // add validation data into boosting for (size_t i = 0; i < valid_datas_.size(); ++i) { boosting_->AddDataset(valid_datas_[i], ConstPtrInVectorWarpper(valid_metrics_[i])); } } ~Booster() { for (auto& metric : train_metric_) { if (metric != nullptr) { delete metric; } } for (auto& metric : valid_metrics_) { for (auto& sub_metric : metric) { if (sub_metric != nullptr) { delete sub_metric; } } } valid_metrics_.clear(); if (boosting_ != nullptr) { delete boosting_; } if (objective_fun_ != nullptr) { delete objective_fun_; } } private: Boosting* boosting_; /*! \brief All configs */ OverallConfig config_; /*! \brief Training data */ const Dataset* train_data_; /*! \brief Validation data */ std::vector valid_datas_; /*! \brief Metric for training data */ std::vector train_metric_; /*! \brief Metrics for validation data */ std::vector> valid_metrics_; /*! \brief Training objective function */ ObjectiveFunction* objective_fun_; }; } using namespace LightGBM; DllExport const char* LGBM_GetLastError() { return "Not error msg now, will support soon"; } DllExport int LGBM_CreateDatasetFromFile(const char* filename, const char* parameters, const DatesetHandle* reference, DatesetHandle* out) { OverallConfig config; config.LoadFromString(parameters); DatasetLoader loader(config.io_config, nullptr); if (reference == nullptr) { *out = loader.LoadFromFile(filename); } else { *out = loader.LoadFromFileLikeOthers(filename, reinterpret_cast(*reference)); } return 0; } DllExport int LGBM_CreateDatasetFromBinaryFile(const char* filename, DatesetHandle* out) { OverallConfig config; DatasetLoader loader(config.io_config, nullptr); *out = loader.LoadFromBinFile(filename, 0, 1); return 0; } DllExport int LGBM_CreateDatasetFromMat(const void* data, int float_type, int32_t nrow, int32_t ncol, int is_row_major, const char* parameters, const DatesetHandle* reference, DatesetHandle* out) { OverallConfig config; config.LoadFromString(parameters); DatasetLoader loader(config.io_config, nullptr); Dataset* ret = nullptr; auto get_row_fun = Common::GetRowFunctionFromMat(float_type, is_row_major); if (reference == nullptr) { // sample data first Random rand(config.io_config.data_random_seed); const size_t sample_cnt = static_cast(nrow < config.io_config.bin_construct_sample_cnt ? nrow : config.io_config.bin_construct_sample_cnt); auto sample_indices = rand.Sample(nrow, sample_cnt); std::vector> sample_data(ncol); for (size_t i = 0; i < sample_indices.size(); i++) { auto idx = sample_indices[i]; auto row = get_row_fun(data, nrow, ncol, static_cast(idx)); for (size_t j = 0; j < row.size(); j++) { sample_data[j].push_back(row[j]); } } ret = loader.CostructFromSampleData(sample_data, nrow); } else { ret = new Dataset(); // need to set num_data first ret->SetNumData(nrow); reinterpret_cast(*reference)->CopyFeatureMetadataTo(ret, config.io_config.is_enable_sparse); } #pragma omp parallel for schedule(guided) for (int i = 0; i < nrow; ++i) { const int tid = omp_get_thread_num(); auto one_row = get_row_fun(data, nrow, ncol, i); ret->PushOneRow(tid, i, one_row); } ret->FinishLoad(); *out = ret; return 1; }