#include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include "./application/predictor.hpp" namespace LightGBM { class Booster { public: explicit Booster(const char* filename): boosting_(Boosting::CreateBoosting(filename)), predictor_(nullptr) { } Booster(const Dataset* train_data, std::vector valid_data, std::vector valid_names, const char* parameters) :train_data_(train_data), valid_datas_(valid_data), predictor_(nullptr) { config_.LoadFromString(parameters); // create boosting if (config_.io_config.input_model.size() > 0) { Log::Warning("continued train from model is not support for c_api, \ please use continued train with input score"); } boosting_ = Boosting::CreateBoosting(config_.boosting_type, ""); // create objective function objective_fun_ = ObjectiveFunction::CreateObjectiveFunction(config_.objective_type, config_.objective_config); // create training metric for (auto metric_type : config_.metric_types) { Metric* metric = Metric::CreateMetric(metric_type, config_.metric_config); if (metric == nullptr) { continue; } metric->Init("training", train_data_->metadata(), train_data_->num_data()); train_metric_.push_back(metric); } // add metric for validation data for (size_t i = 0; i < valid_datas_.size(); ++i) { valid_metrics_.emplace_back(); for (auto metric_type : config_.metric_types) { Metric* metric = Metric::CreateMetric(metric_type, config_.metric_config); if (metric == nullptr) { continue; } metric->Init(valid_names[i].c_str(), valid_datas_[i]->metadata(), valid_datas_[i]->num_data()); valid_metrics_.back().push_back(metric); } } // initialize the objective function objective_fun_->Init(train_data_->metadata(), train_data_->num_data()); // initialize the boosting boosting_->Init(config_.boosting_config, train_data_, objective_fun_, ConstPtrInVectorWarpper(train_metric_)); // add validation data into boosting for (size_t i = 0; i < valid_datas_.size(); ++i) { boosting_->AddDataset(valid_datas_[i], ConstPtrInVectorWarpper(valid_metrics_[i])); } } ~Booster() { for (auto& metric : train_metric_) { if (metric != nullptr) { delete metric; } } for (auto& metric : valid_metrics_) { for (auto& sub_metric : metric) { if (sub_metric != nullptr) { delete sub_metric; } } } valid_metrics_.clear(); if (boosting_ != nullptr) { delete boosting_; } if (objective_fun_ != nullptr) { delete objective_fun_; } if (predictor_ != nullptr) { delete predictor_; } } bool TrainOneIter() { return boosting_->TrainOneIter(nullptr, nullptr, false); } bool TrainOneIter(const float* gradients, const float* hessians) { return boosting_->TrainOneIter(gradients, hessians, false); } void PrepareForPrediction(int num_used_model, int predict_type) { boosting_->SetNumUsedModel(num_used_model); if (predictor_ != nullptr) { delete predictor_; } bool is_predict_leaf = false; bool is_raw_score = false; if (predict_type == 2) { is_predict_leaf = true; } else if (predict_type == 1) { is_raw_score = false; } else { is_raw_score = true; } predictor_ = new Predictor(boosting_, is_raw_score, is_predict_leaf); } std::vector Predict(const std::vector>& features) { return predictor_->GetPredictFunction()(features); } void SaveModelToFile(int num_used_model, const char* filename) { boosting_->SaveModelToFile(num_used_model, true, filename); } const Boosting* GetBoosting() const { return boosting_; } const inline int NumberOfClasses() const { return boosting_->NumberOfClasses(); } private: Boosting* boosting_; /*! \brief All configs */ OverallConfig config_; /*! \brief Training data */ const Dataset* train_data_; /*! \brief Validation data */ std::vector valid_datas_; /*! \brief Metric for training data */ std::vector train_metric_; /*! \brief Metrics for validation data */ std::vector> valid_metrics_; /*! \brief Training objective function */ ObjectiveFunction* objective_fun_; /*! \brief Using predictor for prediction task */ Predictor* predictor_; }; } using namespace LightGBM; DllExport const char* LGBM_GetLastError() { return "Not error msg now, will support soon"; } DllExport int LGBM_CreateDatasetFromFile(const char* filename, const char* parameters, const DatesetHandle* reference, DatesetHandle* out) { OverallConfig config; config.LoadFromString(parameters); DatasetLoader loader(config.io_config, nullptr); if (reference == nullptr) { *out = loader.LoadFromFile(filename); } else { *out = loader.LoadFromFileAlignWithOtherDataset(filename, reinterpret_cast(*reference)); } return 0; } DllExport int LGBM_CreateDatasetFromBinaryFile(const char* filename, DatesetHandle* out) { OverallConfig config; DatasetLoader loader(config.io_config, nullptr); *out = loader.LoadFromBinFile(filename, 0, 1); return 0; } DllExport int LGBM_CreateDatasetFromMat(const void* data, int float_type, int32_t nrow, int32_t ncol, int is_row_major, const char* parameters, const DatesetHandle* reference, DatesetHandle* out) { OverallConfig config; config.LoadFromString(parameters); DatasetLoader loader(config.io_config, nullptr); Dataset* ret = nullptr; auto get_row_fun = Common::RowFunctionFromDenseMatric(data, nrow, ncol, float_type, is_row_major); if (reference == nullptr) { // sample data first Random rand(config.io_config.data_random_seed); const size_t sample_cnt = static_cast(nrow < config.io_config.bin_construct_sample_cnt ? nrow : config.io_config.bin_construct_sample_cnt); auto sample_indices = rand.Sample(nrow, sample_cnt); std::vector> sample_values(ncol); for (size_t i = 0; i < sample_indices.size(); ++i) { auto idx = sample_indices[i]; auto row = get_row_fun(static_cast(idx)); for (size_t j = 0; j < row.size(); ++j) { sample_values[j].push_back(row[j]); } } ret = loader.CostructFromSampleData(sample_values, nrow); } else { ret = new Dataset(nrow, config.io_config.num_class); reinterpret_cast(*reference)->CopyFeatureBinMapperTo(ret, config.io_config.is_enable_sparse); } #pragma omp parallel for schedule(guided) for (int i = 0; i < nrow; ++i) { const int tid = omp_get_thread_num(); auto one_row = get_row_fun(i); ret->PushOneRow(tid, i, one_row); } ret->FinishLoad(); *out = ret; return 0; } DllExport int LGBM_CreateDatasetFromCSR(const int32_t* indptr, const int32_t* indices, const void* data, int float_type, uint64_t nindptr, uint64_t nelem, uint64_t num_col, const char* parameters, const DatesetHandle* reference, DatesetHandle* out) { OverallConfig config; config.LoadFromString(parameters); DatasetLoader loader(config.io_config, nullptr); Dataset* ret = nullptr; auto get_row_fun = Common::RowFunctionFromCSR(indptr, indices, data, float_type, nindptr, nelem); int32_t nrow = static_cast(nindptr - 1); if (reference == nullptr) { // sample data first Random rand(config.io_config.data_random_seed); const size_t sample_cnt = static_cast(nrow < config.io_config.bin_construct_sample_cnt ? nrow : config.io_config.bin_construct_sample_cnt); auto sample_indices = rand.Sample(nrow, sample_cnt); std::vector> sample_values; for (size_t i = 0; i < sample_indices.size(); ++i) { auto idx = sample_indices[i]; auto row = get_row_fun(static_cast(idx)); // push 0 first, then edit the value according existing feature values for (auto& feature_values : sample_values) { feature_values.push_back(0.0); } for (std::pair& inner_data : row) { if (static_cast(inner_data.first) >= sample_values.size()) { // if need expand feature set size_t need_size = inner_data.first - sample_values.size() + 1; for (size_t j = 0; j < need_size; ++j) { // push i+1 0 sample_values.emplace_back(i + 1, 0.0f); } } // edit the feature value sample_values[inner_data.first][i] = inner_data.second; } } CHECK(num_col >= sample_values.size()); ret = loader.CostructFromSampleData(sample_values, nrow); } else { ret = new Dataset(nrow, config.io_config.num_class); reinterpret_cast(*reference)->CopyFeatureBinMapperTo(ret, config.io_config.is_enable_sparse); } #pragma omp parallel for schedule(guided) for (int i = 0; i < nindptr - 1; ++i) { const int tid = omp_get_thread_num(); auto one_row = get_row_fun(i); ret->PushOneRow(tid, i, one_row); } ret->FinishLoad(); *out = ret; return 0; } DllExport int LGBM_CreateDatasetFromCSC(const int32_t* col_ptr, const int32_t* indices, const void* data, int float_type, uint64_t ncol_ptr, uint64_t nelem, uint64_t num_row, const char* parameters, const DatesetHandle* reference, DatesetHandle* out) { OverallConfig config; config.LoadFromString(parameters); DatasetLoader loader(config.io_config, nullptr); Dataset* ret = nullptr; auto get_col_fun = Common::ColumnFunctionFromCSC(col_ptr, indices, data, float_type, ncol_ptr, nelem); int32_t nrow = static_cast(num_row); if (reference == nullptr) { Log::Warning("Construct from CSC format is not efficient"); // sample data first Random rand(config.io_config.data_random_seed); const size_t sample_cnt = static_cast(nrow < config.io_config.bin_construct_sample_cnt ? nrow : config.io_config.bin_construct_sample_cnt); auto sample_indices = rand.Sample(nrow, sample_cnt); std::vector> sample_values(ncol_ptr - 1); #pragma omp parallel for schedule(guided) for (int i = 0; i < static_cast(sample_values.size()); ++i) { auto cur_col = get_col_fun(i); sample_values[i] = Common::SampleFromOneColumn(cur_col, sample_indices); } ret = loader.CostructFromSampleData(sample_values, nrow); } else { ret = new Dataset(nrow, config.io_config.num_class); reinterpret_cast(*reference)->CopyFeatureBinMapperTo(ret, config.io_config.is_enable_sparse); } #pragma omp parallel for schedule(guided) for (int i = 0; i < ncol_ptr - 1; ++i) { const int tid = omp_get_thread_num(); auto one_col = get_col_fun(i); ret->PushOneColumn(tid, i, one_col); } ret->FinishLoad(); *out = ret; return 0; } DllExport int LGBM_DatasetFree(DatesetHandle* handle) { auto dataset = reinterpret_cast(*handle); delete dataset; return 0; } DllExport int LGBM_DatasetSaveBinary(DatesetHandle handle, const char* filename) { auto dataset = reinterpret_cast(handle); dataset->SaveBinaryFile(filename); return 0; } DllExport int LGBM_DatasetSetField(DatesetHandle handle, const char* field_name, const void* field_data, uint64_t num_element, int type) { auto dataset = reinterpret_cast(handle); dataset->SetField(field_name, field_data, static_cast(num_element), type); return 0; } DllExport int LGBM_DatasetGetField(DatesetHandle handle, const char* field_name, uint64_t* out_len, const void** out_ptr, int* out_type) { auto dataset = reinterpret_cast(handle); dataset->GetField(field_name, out_len, out_ptr, out_type); return 0; } DllExport int LGBM_DatasetGetNumData(DatesetHandle handle, uint64_t* out) { auto dataset = reinterpret_cast(handle); *out = dataset->num_data(); return 0; } DllExport int LGBM_DatasetGetNumFeature(DatesetHandle handle, uint64_t* out) { auto dataset = reinterpret_cast(handle); *out = dataset->num_total_features(); return 0; } // ---- start of booster DllExport int LGBM_BoosterCreate(const DatesetHandle train_data, const DatesetHandle valid_datas[], const char* valid_names[], int n_valid_datas, const char* parameters, BoosterHandle* out) { const Dataset* p_train_data = reinterpret_cast(train_data); std::vector p_valid_datas; std::vector p_valid_names; for (int i = 0; i < n_valid_datas; ++i) { p_valid_datas.emplace_back(reinterpret_cast(valid_datas[i])); p_valid_names.emplace_back(valid_names[i]); } *out = new Booster(p_train_data, p_valid_datas, p_valid_names, parameters); return 0; } DllExport int LGBM_BoosterLoadFromModelfile( const char* filename, BoosterHandle* out) { *out = new Booster(filename); return 0; } DllExport int LGBM_BoosterFree(BoosterHandle handle) { Booster* ref_booster = reinterpret_cast(handle); delete ref_booster; return 0; } DllExport int LGBM_BoosterUpdateOneIter(BoosterHandle handle, int* is_finished) { Booster* ref_booster = reinterpret_cast(handle); if (ref_booster->TrainOneIter()) { *is_finished = 1; } else { *is_finished = 0; } return 0; } DllExport int LGBM_BoosterUpdateOneIterCustom(BoosterHandle handle, const float* grad, const float* hess, int* is_finished) { Booster* ref_booster = reinterpret_cast(handle); if (ref_booster->TrainOneIter(grad, hess)) { *is_finished = 1; } else { *is_finished = 0; } return 0; } DllExport int LGBM_BoosterEval(BoosterHandle handle, int data, uint64_t* out_len, float* out_results) { Booster* ref_booster = reinterpret_cast(handle); auto boosting = ref_booster->GetBoosting(); auto result_buf = boosting->GetEvalAt(data); *out_len = static_cast(result_buf.size()); for (size_t i = 0; i < result_buf.size(); ++i) { (out_results)[i] = static_cast(result_buf[i]); } return 0; } DllExport int LGBM_BoosterGetScore(BoosterHandle handle, uint64_t* out_len, const float** out_result) { Booster* ref_booster = reinterpret_cast(handle); auto boosting = ref_booster->GetBoosting(); int len = 0; *out_result = boosting->GetTrainingScore(&len); *out_len = static_cast(len); return 0; } DllExport int LGBM_BoosterGetPredict(BoosterHandle handle, int data, uint64_t* out_len, float* out_result) { Booster* ref_booster = reinterpret_cast(handle); auto boosting = ref_booster->GetBoosting(); int len = 0; boosting->GetPredictAt(data, out_result, &len); *out_len = static_cast(len); return 0; } DllExport int LGBM_BoosterPredictForCSR(BoosterHandle handle, const int32_t* indptr, const int32_t* indices, const void* data, int float_type, uint64_t nindptr, uint64_t nelem, uint64_t, int predict_type, uint64_t n_used_trees, double* out_result) { Booster* ref_booster = reinterpret_cast(handle); ref_booster->PrepareForPrediction(static_cast(n_used_trees), predict_type); auto get_row_fun = Common::RowFunctionFromCSR(indptr, indices, data, float_type, nindptr, nelem); int num_class = ref_booster->NumberOfClasses(); int nrow = static_cast(nindptr - 1); #pragma omp parallel for schedule(guided) for (int i = 0; i < nrow; ++i) { auto one_row = get_row_fun(i); auto predicton_result = ref_booster->Predict(one_row); for (int j = 0; j < num_class; ++j) { out_result[i * num_class + j] = predicton_result[j]; } } return 0; } DllExport int LGBM_BoosterPredictForMat(BoosterHandle handle, const void* data, int float_type, int32_t nrow, int32_t ncol, int is_row_major, int predict_type, uint64_t n_used_trees, double* out_result) { Booster* ref_booster = reinterpret_cast(handle); ref_booster->PrepareForPrediction(static_cast(n_used_trees), predict_type); auto get_row_fun = Common::RowPairFunctionFromDenseMatric(data, nrow, ncol, float_type, is_row_major); int num_class = ref_booster->NumberOfClasses(); #pragma omp parallel for schedule(guided) for (int i = 0; i < nrow; ++i) { auto one_row = get_row_fun(i); auto predicton_result = ref_booster->Predict(one_row); for (int j = 0; j < num_class; ++j) { out_result[i * num_class + j] = predicton_result[j]; } } return 0; } DllExport int LGBM_BoosterSaveModel(BoosterHandle handle, int num_used_model, const char* filename) { Booster* ref_booster = reinterpret_cast(handle); ref_booster->SaveModelToFile(num_used_model, filename); return 0; }