/*! * Copyright (c) 2016 Microsoft Corporation. All rights reserved. * Licensed under the MIT License. See LICENSE file in the project root for license information. */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include "application/predictor.hpp" namespace LightGBM { inline int LGBM_APIHandleException(const std::exception& ex) { LGBM_SetLastError(ex.what()); return -1; } inline int LGBM_APIHandleException(const std::string& ex) { LGBM_SetLastError(ex.c_str()); return -1; } #define API_BEGIN() try { #define API_END() } \ catch(std::exception& ex) { return LGBM_APIHandleException(ex); } \ catch(std::string& ex) { return LGBM_APIHandleException(ex); } \ catch(...) { return LGBM_APIHandleException("unknown exception"); } \ return 0; const int PREDICTOR_TYPES = 4; // Single row predictor to abstract away caching logic class SingleRowPredictor { public: PredictFunction predict_function; int64_t num_pred_in_one_row; SingleRowPredictor(int predict_type, Boosting* boosting, const Config& config, int iter) { bool is_predict_leaf = false; bool is_raw_score = false; bool predict_contrib = false; if (predict_type == C_API_PREDICT_LEAF_INDEX) { is_predict_leaf = true; } else if (predict_type == C_API_PREDICT_RAW_SCORE) { is_raw_score = true; } else if (predict_type == C_API_PREDICT_CONTRIB) { predict_contrib = true; } else { is_raw_score = false; } early_stop_ = config.pred_early_stop; early_stop_freq_ = config.pred_early_stop_freq; early_stop_margin_ = config.pred_early_stop_margin; iter_ = iter; predictor_.reset(new Predictor(boosting, iter_, is_raw_score, is_predict_leaf, predict_contrib, early_stop_, early_stop_freq_, early_stop_margin_)); num_pred_in_one_row = boosting->NumPredictOneRow(iter_, is_predict_leaf, predict_contrib); predict_function = predictor_->GetPredictFunction(); num_total_model_ = boosting->NumberOfTotalModel(); } ~SingleRowPredictor() {} bool IsPredictorEqual(const Config& config, int iter, Boosting* boosting) { return early_stop_ == config.pred_early_stop && early_stop_freq_ == config.pred_early_stop_freq && early_stop_margin_ == config.pred_early_stop_margin && iter_ == iter && num_total_model_ == boosting->NumberOfTotalModel(); } private: std::unique_ptr predictor_; bool early_stop_; int early_stop_freq_; double early_stop_margin_; int iter_; int num_total_model_; }; class Booster { public: explicit Booster(const char* filename) { boosting_.reset(Boosting::CreateBoosting("gbdt", filename)); } Booster(const Dataset* train_data, const char* parameters) { auto param = Config::Str2Map(parameters); config_.Set(param); if (config_.num_threads > 0) { omp_set_num_threads(config_.num_threads); } // create boosting if (config_.input_model.size() > 0) { Log::Warning("Continued train from model is not supported for c_api,\n" "please use continued train with input score"); } boosting_.reset(Boosting::CreateBoosting(config_.boosting, nullptr)); train_data_ = train_data; CreateObjectiveAndMetrics(); // initialize the boosting if (config_.tree_learner == std::string("feature")) { Log::Fatal("Do not support feature parallel in c api"); } if (Network::num_machines() == 1 && config_.tree_learner != std::string("serial")) { Log::Warning("Only find one worker, will switch to serial tree learner"); config_.tree_learner = "serial"; } boosting_->Init(&config_, train_data_, objective_fun_.get(), Common::ConstPtrInVectorWrapper(train_metric_)); } void MergeFrom(const Booster* other) { std::lock_guard lock(mutex_); boosting_->MergeFrom(other->boosting_.get()); } ~Booster() { } void CreateObjectiveAndMetrics() { // create objective function objective_fun_.reset(ObjectiveFunction::CreateObjectiveFunction(config_.objective, config_)); if (objective_fun_ == nullptr) { Log::Warning("Using self-defined objective function"); } // initialize the objective function if (objective_fun_ != nullptr) { objective_fun_->Init(train_data_->metadata(), train_data_->num_data()); } // create training metric train_metric_.clear(); for (auto metric_type : config_.metric) { auto metric = std::unique_ptr( Metric::CreateMetric(metric_type, config_)); if (metric == nullptr) { continue; } metric->Init(train_data_->metadata(), train_data_->num_data()); train_metric_.push_back(std::move(metric)); } train_metric_.shrink_to_fit(); } void ResetTrainingData(const Dataset* train_data) { if (train_data != train_data_) { std::lock_guard lock(mutex_); train_data_ = train_data; CreateObjectiveAndMetrics(); // reset the boosting boosting_->ResetTrainingData(train_data_, objective_fun_.get(), Common::ConstPtrInVectorWrapper(train_metric_)); } } static void CheckDatasetResetConfig( const Config& old_config, const std::unordered_map& new_param) { Config new_config; new_config.Set(new_param); if (new_param.count("data_random_seed") && new_config.data_random_seed != old_config.data_random_seed) { Log::Fatal("Cannot change data_random_seed after constructed Dataset handle."); } if (new_param.count("max_bin") && new_config.max_bin != old_config.max_bin) { Log::Fatal("Cannot change max_bin after constructed Dataset handle."); } if (new_param.count("max_bin_by_feature") && new_config.max_bin_by_feature != old_config.max_bin_by_feature) { Log::Fatal( "Cannot change max_bin_by_feature after constructed Dataset handle."); } if (new_param.count("bin_construct_sample_cnt") && new_config.bin_construct_sample_cnt != old_config.bin_construct_sample_cnt) { Log::Fatal( "Cannot change bin_construct_sample_cnt after constructed Dataset " "handle."); } if (new_param.count("min_data_in_bin") && new_config.min_data_in_bin != old_config.min_data_in_bin) { Log::Fatal( "Cannot change min_data_in_bin after constructed Dataset handle."); } if (new_param.count("use_missing") && new_config.use_missing != old_config.use_missing) { Log::Fatal("Cannot change use_missing after constructed Dataset handle."); } if (new_param.count("zero_as_missing") && new_config.zero_as_missing != old_config.zero_as_missing) { Log::Fatal( "Cannot change zero_as_missing after constructed Dataset handle."); } if (new_param.count("categorical_feature") && new_config.categorical_feature != old_config.categorical_feature) { Log::Fatal( "Cannot change categorical_feature after constructed Dataset " "handle."); } if (new_param.count("feature_pre_filter") && new_config.feature_pre_filter != old_config.feature_pre_filter) { Log::Fatal( "Cannot change feature_pre_filter after constructed Dataset handle."); } if (new_param.count("is_enable_sparse") && new_config.is_enable_sparse != old_config.is_enable_sparse) { Log::Fatal( "Cannot change is_enable_sparse after constructed Dataset handle."); } if (new_param.count("pre_partition") && new_config.pre_partition != old_config.pre_partition) { Log::Fatal( "Cannot change pre_partition after constructed Dataset handle."); } if (new_param.count("enable_bundle") && new_config.enable_bundle != old_config.enable_bundle) { Log::Fatal( "Cannot change enable_bundle after constructed Dataset handle."); } if (new_param.count("header") && new_config.header != old_config.header) { Log::Fatal("Cannot change header after constructed Dataset handle."); } if (new_param.count("two_round") && new_config.two_round != old_config.two_round) { Log::Fatal("Cannot change two_round after constructed Dataset handle."); } if (new_param.count("label_column") && new_config.label_column != old_config.label_column) { Log::Fatal( "Cannot change label_column after constructed Dataset handle."); } if (new_param.count("weight_column") && new_config.weight_column != old_config.weight_column) { Log::Fatal( "Cannot change weight_column after constructed Dataset handle."); } if (new_param.count("group_column") && new_config.group_column != old_config.group_column) { Log::Fatal( "Cannot change group_column after constructed Dataset handle."); } if (new_param.count("ignore_column") && new_config.ignore_column != old_config.ignore_column) { Log::Fatal( "Cannot change ignore_column after constructed Dataset handle."); } if (new_param.count("forcedbins_filename")) { Log::Fatal("Cannot change forced bins after constructed Dataset handle."); } if (new_param.count("min_data_in_leaf") && new_config.min_data_in_leaf < old_config.min_data_in_leaf && old_config.feature_pre_filter) { Log::Fatal( "Reducing `min_data_in_leaf` with `feature_pre_filter=true` may " "cause unexpected behaviour " "for features that were pre-filtered by the larger " "`min_data_in_leaf`.\n" "You need to set `feature_pre_filter=false` to dynamically change " "the `min_data_in_leaf`."); } } void ResetConfig(const char* parameters) { std::lock_guard lock(mutex_); auto param = Config::Str2Map(parameters); if (param.count("num_class")) { Log::Fatal("Cannot change num_class during training"); } if (param.count("boosting")) { Log::Fatal("Cannot change boosting during training"); } if (param.count("metric")) { Log::Fatal("Cannot change metric during training"); } CheckDatasetResetConfig(config_, param); config_.Set(param); if (config_.num_threads > 0) { omp_set_num_threads(config_.num_threads); } if (param.count("objective")) { // create objective function objective_fun_.reset(ObjectiveFunction::CreateObjectiveFunction(config_.objective, config_)); if (objective_fun_ == nullptr) { Log::Warning("Using self-defined objective function"); } // initialize the objective function if (objective_fun_ != nullptr) { objective_fun_->Init(train_data_->metadata(), train_data_->num_data()); } boosting_->ResetTrainingData(train_data_, objective_fun_.get(), Common::ConstPtrInVectorWrapper(train_metric_)); } boosting_->ResetConfig(&config_); } void AddValidData(const Dataset* valid_data) { std::lock_guard lock(mutex_); valid_metrics_.emplace_back(); for (auto metric_type : config_.metric) { auto metric = std::unique_ptr(Metric::CreateMetric(metric_type, config_)); if (metric == nullptr) { continue; } metric->Init(valid_data->metadata(), valid_data->num_data()); valid_metrics_.back().push_back(std::move(metric)); } valid_metrics_.back().shrink_to_fit(); boosting_->AddValidDataset(valid_data, Common::ConstPtrInVectorWrapper(valid_metrics_.back())); } bool TrainOneIter() { std::lock_guard lock(mutex_); return boosting_->TrainOneIter(nullptr, nullptr); } void Refit(const int32_t* leaf_preds, int32_t nrow, int32_t ncol) { std::lock_guard lock(mutex_); std::vector> v_leaf_preds(nrow, std::vector(ncol, 0)); for (int i = 0; i < nrow; ++i) { for (int j = 0; j < ncol; ++j) { v_leaf_preds[i][j] = leaf_preds[i * ncol + j]; } } boosting_->RefitTree(v_leaf_preds); } bool TrainOneIter(const score_t* gradients, const score_t* hessians) { std::lock_guard lock(mutex_); return boosting_->TrainOneIter(gradients, hessians); } void RollbackOneIter() { std::lock_guard lock(mutex_); boosting_->RollbackOneIter(); } void PredictSingleRow(int num_iteration, int predict_type, int ncol, std::function>(int row_idx)> get_row_fun, const Config& config, double* out_result, int64_t* out_len) { if (!config.predict_disable_shape_check && ncol != boosting_->MaxFeatureIdx() + 1) { Log::Fatal("The number of features in data (%d) is not the same as it was in training data (%d).\n"\ "You can set ``predict_disable_shape_check=true`` to discard this error, but please be aware what you are doing.", ncol, boosting_->MaxFeatureIdx() + 1); } std::lock_guard lock(mutex_); if (single_row_predictor_[predict_type].get() == nullptr || !single_row_predictor_[predict_type]->IsPredictorEqual(config, num_iteration, boosting_.get())) { single_row_predictor_[predict_type].reset(new SingleRowPredictor(predict_type, boosting_.get(), config, num_iteration)); } auto one_row = get_row_fun(0); auto pred_wrt_ptr = out_result; single_row_predictor_[predict_type]->predict_function(one_row, pred_wrt_ptr); *out_len = single_row_predictor_[predict_type]->num_pred_in_one_row; } void Predict(int num_iteration, int predict_type, int nrow, int ncol, std::function>(int row_idx)> get_row_fun, const Config& config, double* out_result, int64_t* out_len) { if (!config.predict_disable_shape_check && ncol != boosting_->MaxFeatureIdx() + 1) { Log::Fatal("The number of features in data (%d) is not the same as it was in training data (%d).\n" \ "You can set ``predict_disable_shape_check=true`` to discard this error, but please be aware what you are doing.", ncol, boosting_->MaxFeatureIdx() + 1); } std::lock_guard lock(mutex_); bool is_predict_leaf = false; bool is_raw_score = false; bool predict_contrib = false; if (predict_type == C_API_PREDICT_LEAF_INDEX) { is_predict_leaf = true; } else if (predict_type == C_API_PREDICT_RAW_SCORE) { is_raw_score = true; } else if (predict_type == C_API_PREDICT_CONTRIB) { predict_contrib = true; } else { is_raw_score = false; } Predictor predictor(boosting_.get(), num_iteration, is_raw_score, is_predict_leaf, predict_contrib, config.pred_early_stop, config.pred_early_stop_freq, config.pred_early_stop_margin); int64_t num_pred_in_one_row = boosting_->NumPredictOneRow(num_iteration, is_predict_leaf, predict_contrib); auto pred_fun = predictor.GetPredictFunction(); OMP_INIT_EX(); #pragma omp parallel for schedule(static) for (int i = 0; i < nrow; ++i) { OMP_LOOP_EX_BEGIN(); auto one_row = get_row_fun(i); auto pred_wrt_ptr = out_result + static_cast(num_pred_in_one_row) * i; pred_fun(one_row, pred_wrt_ptr); OMP_LOOP_EX_END(); } OMP_THROW_EX(); *out_len = num_pred_in_one_row * nrow; } void Predict(int num_iteration, int predict_type, const char* data_filename, int data_has_header, const Config& config, const char* result_filename) { std::lock_guard lock(mutex_); bool is_predict_leaf = false; bool is_raw_score = false; bool predict_contrib = false; if (predict_type == C_API_PREDICT_LEAF_INDEX) { is_predict_leaf = true; } else if (predict_type == C_API_PREDICT_RAW_SCORE) { is_raw_score = true; } else if (predict_type == C_API_PREDICT_CONTRIB) { predict_contrib = true; } else { is_raw_score = false; } Predictor predictor(boosting_.get(), num_iteration, is_raw_score, is_predict_leaf, predict_contrib, config.pred_early_stop, config.pred_early_stop_freq, config.pred_early_stop_margin); bool bool_data_has_header = data_has_header > 0 ? true : false; predictor.Predict(data_filename, result_filename, bool_data_has_header, config.predict_disable_shape_check); } void GetPredictAt(int data_idx, double* out_result, int64_t* out_len) { boosting_->GetPredictAt(data_idx, out_result, out_len); } void SaveModelToFile(int start_iteration, int num_iteration, const char* filename) { boosting_->SaveModelToFile(start_iteration, num_iteration, filename); } void LoadModelFromString(const char* model_str) { size_t len = std::strlen(model_str); boosting_->LoadModelFromString(model_str, len); } std::string SaveModelToString(int start_iteration, int num_iteration) { return boosting_->SaveModelToString(start_iteration, num_iteration); } std::string DumpModel(int start_iteration, int num_iteration) { return boosting_->DumpModel(start_iteration, num_iteration); } std::vector FeatureImportance(int num_iteration, int importance_type) { return boosting_->FeatureImportance(num_iteration, importance_type); } double UpperBoundValue() const { std::lock_guard lock(mutex_); return boosting_->GetUpperBoundValue(); } double LowerBoundValue() const { std::lock_guard lock(mutex_); return boosting_->GetLowerBoundValue(); } double GetLeafValue(int tree_idx, int leaf_idx) const { return dynamic_cast(boosting_.get())->GetLeafValue(tree_idx, leaf_idx); } void SetLeafValue(int tree_idx, int leaf_idx, double val) { std::lock_guard lock(mutex_); dynamic_cast(boosting_.get())->SetLeafValue(tree_idx, leaf_idx, val); } void ShuffleModels(int start_iter, int end_iter) { std::lock_guard lock(mutex_); boosting_->ShuffleModels(start_iter, end_iter); } int GetEvalCounts() const { int ret = 0; for (const auto& metric : train_metric_) { ret += static_cast(metric->GetName().size()); } return ret; } int GetEvalNames(char** out_strs) const { int idx = 0; for (const auto& metric : train_metric_) { for (const auto& name : metric->GetName()) { std::memcpy(out_strs[idx], name.c_str(), name.size() + 1); ++idx; } } return idx; } int GetFeatureNames(char** out_strs) const { int idx = 0; for (const auto& name : boosting_->FeatureNames()) { std::memcpy(out_strs[idx], name.c_str(), name.size() + 1); ++idx; } return idx; } const Boosting* GetBoosting() const { return boosting_.get(); } private: const Dataset* train_data_; std::unique_ptr boosting_; std::unique_ptr single_row_predictor_[PREDICTOR_TYPES]; /*! \brief All configs */ Config config_; /*! \brief Metric for training data */ std::vector> train_metric_; /*! \brief Metrics for validation data */ std::vector>> valid_metrics_; /*! \brief Training objective function */ std::unique_ptr objective_fun_; /*! \brief mutex for threading safe call */ mutable std::mutex mutex_; }; } // namespace LightGBM using namespace LightGBM; // some help functions used to convert data std::function(int row_idx)> RowFunctionFromDenseMatric(const void* data, int num_row, int num_col, int data_type, int is_row_major); std::function>(int row_idx)> RowPairFunctionFromDenseMatric(const void* data, int num_row, int num_col, int data_type, int is_row_major); std::function>(int row_idx)> RowPairFunctionFromDenseRows(const void** data, int num_col, int data_type); std::function>(int idx)> RowFunctionFromCSR(const void* indptr, int indptr_type, const int32_t* indices, const void* data, int data_type, int64_t nindptr, int64_t nelem); // Row iterator of on column for CSC matrix class CSC_RowIterator { public: CSC_RowIterator(const void* col_ptr, int col_ptr_type, const int32_t* indices, const void* data, int data_type, int64_t ncol_ptr, int64_t nelem, int col_idx); ~CSC_RowIterator() {} // return value at idx, only can access by ascent order double Get(int idx); // return next non-zero pair, if index < 0, means no more data std::pair NextNonZero(); private: int nonzero_idx_ = 0; int cur_idx_ = -1; double cur_val_ = 0.0f; bool is_end_ = false; std::function(int idx)> iter_fun_; }; // start of c_api functions const char* LGBM_GetLastError() { return LastErrorMsg(); } int LGBM_DatasetCreateFromFile(const char* filename, const char* parameters, const DatasetHandle reference, DatasetHandle* out) { API_BEGIN(); auto param = Config::Str2Map(parameters); Config config; config.Set(param); if (config.num_threads > 0) { omp_set_num_threads(config.num_threads); } DatasetLoader loader(config, nullptr, 1, filename); if (reference == nullptr) { if (Network::num_machines() == 1) { *out = loader.LoadFromFile(filename); } else { *out = loader.LoadFromFile(filename, Network::rank(), Network::num_machines()); } } else { *out = loader.LoadFromFileAlignWithOtherDataset(filename, reinterpret_cast(reference)); } API_END(); } int LGBM_DatasetCreateFromSampledColumn(double** sample_data, int** sample_indices, int32_t ncol, const int* num_per_col, int32_t num_sample_row, int32_t num_total_row, const char* parameters, DatasetHandle* out) { API_BEGIN(); auto param = Config::Str2Map(parameters); Config config; config.Set(param); if (config.num_threads > 0) { omp_set_num_threads(config.num_threads); } DatasetLoader loader(config, nullptr, 1, nullptr); *out = loader.CostructFromSampleData(sample_data, sample_indices, ncol, num_per_col, num_sample_row, static_cast(num_total_row)); API_END(); } int LGBM_DatasetCreateByReference(const DatasetHandle reference, int64_t num_total_row, DatasetHandle* out) { API_BEGIN(); std::unique_ptr ret; ret.reset(new Dataset(static_cast(num_total_row))); ret->CreateValid(reinterpret_cast(reference)); *out = ret.release(); API_END(); } int LGBM_DatasetPushRows(DatasetHandle dataset, const void* data, int data_type, int32_t nrow, int32_t ncol, int32_t start_row) { API_BEGIN(); auto p_dataset = reinterpret_cast(dataset); auto get_row_fun = RowFunctionFromDenseMatric(data, nrow, ncol, data_type, 1); OMP_INIT_EX(); #pragma omp parallel for schedule(static) for (int i = 0; i < nrow; ++i) { OMP_LOOP_EX_BEGIN(); const int tid = omp_get_thread_num(); auto one_row = get_row_fun(i); p_dataset->PushOneRow(tid, start_row + i, one_row); OMP_LOOP_EX_END(); } OMP_THROW_EX(); if (start_row + nrow == p_dataset->num_data()) { p_dataset->FinishLoad(); } API_END(); } int LGBM_DatasetPushRowsByCSR(DatasetHandle dataset, const void* indptr, int indptr_type, const int32_t* indices, const void* data, int data_type, int64_t nindptr, int64_t nelem, int64_t, int64_t start_row) { API_BEGIN(); auto p_dataset = reinterpret_cast(dataset); auto get_row_fun = RowFunctionFromCSR(indptr, indptr_type, indices, data, data_type, nindptr, nelem); int32_t nrow = static_cast(nindptr - 1); OMP_INIT_EX(); #pragma omp parallel for schedule(static) for (int i = 0; i < nrow; ++i) { OMP_LOOP_EX_BEGIN(); const int tid = omp_get_thread_num(); auto one_row = get_row_fun(i); p_dataset->PushOneRow(tid, static_cast(start_row + i), one_row); OMP_LOOP_EX_END(); } OMP_THROW_EX(); if (start_row + nrow == static_cast(p_dataset->num_data())) { p_dataset->FinishLoad(); } API_END(); } int LGBM_DatasetCreateFromMat(const void* data, int data_type, int32_t nrow, int32_t ncol, int is_row_major, const char* parameters, const DatasetHandle reference, DatasetHandle* out) { return LGBM_DatasetCreateFromMats(1, &data, data_type, &nrow, ncol, is_row_major, parameters, reference, out); } int LGBM_DatasetCreateFromMats(int32_t nmat, const void** data, int data_type, int32_t* nrow, int32_t ncol, int is_row_major, const char* parameters, const DatasetHandle reference, DatasetHandle* out) { API_BEGIN(); auto param = Config::Str2Map(parameters); Config config; config.Set(param); if (config.num_threads > 0) { omp_set_num_threads(config.num_threads); } std::unique_ptr ret; int32_t total_nrow = 0; for (int j = 0; j < nmat; ++j) { total_nrow += nrow[j]; } std::vector(int row_idx)>> get_row_fun; for (int j = 0; j < nmat; ++j) { get_row_fun.push_back(RowFunctionFromDenseMatric(data[j], nrow[j], ncol, data_type, is_row_major)); } if (reference == nullptr) { // sample data first Random rand(config.data_random_seed); int sample_cnt = static_cast(total_nrow < config.bin_construct_sample_cnt ? total_nrow : config.bin_construct_sample_cnt); auto sample_indices = rand.Sample(total_nrow, sample_cnt); sample_cnt = static_cast(sample_indices.size()); std::vector> sample_values(ncol); std::vector> sample_idx(ncol); int offset = 0; int j = 0; for (size_t i = 0; i < sample_indices.size(); ++i) { auto idx = sample_indices[i]; while ((idx - offset) >= nrow[j]) { offset += nrow[j]; ++j; } auto row = get_row_fun[j](static_cast(idx - offset)); for (size_t k = 0; k < row.size(); ++k) { if (std::fabs(row[k]) > kZeroThreshold || std::isnan(row[k])) { sample_values[k].emplace_back(row[k]); sample_idx[k].emplace_back(static_cast(i)); } } } DatasetLoader loader(config, nullptr, 1, nullptr); ret.reset(loader.CostructFromSampleData(Common::Vector2Ptr(&sample_values).data(), Common::Vector2Ptr(&sample_idx).data(), ncol, Common::VectorSize(sample_values).data(), sample_cnt, total_nrow)); } else { ret.reset(new Dataset(total_nrow)); ret->CreateValid( reinterpret_cast(reference)); } int32_t start_row = 0; for (int j = 0; j < nmat; ++j) { OMP_INIT_EX(); #pragma omp parallel for schedule(static) for (int i = 0; i < nrow[j]; ++i) { OMP_LOOP_EX_BEGIN(); const int tid = omp_get_thread_num(); auto one_row = get_row_fun[j](i); ret->PushOneRow(tid, start_row + i, one_row); OMP_LOOP_EX_END(); } OMP_THROW_EX(); start_row += nrow[j]; } ret->FinishLoad(); *out = ret.release(); API_END(); } int LGBM_DatasetCreateFromCSR(const void* indptr, int indptr_type, const int32_t* indices, const void* data, int data_type, int64_t nindptr, int64_t nelem, int64_t num_col, const char* parameters, const DatasetHandle reference, DatasetHandle* out) { API_BEGIN(); if (num_col <= 0) { Log::Fatal("The number of columns should be greater than zero."); } else if (num_col >= INT32_MAX) { Log::Fatal("The number of columns should be smaller than INT32_MAX."); } auto param = Config::Str2Map(parameters); Config config; config.Set(param); if (config.num_threads > 0) { omp_set_num_threads(config.num_threads); } std::unique_ptr ret; auto get_row_fun = RowFunctionFromCSR(indptr, indptr_type, indices, data, data_type, nindptr, nelem); int32_t nrow = static_cast(nindptr - 1); if (reference == nullptr) { // sample data first Random rand(config.data_random_seed); int sample_cnt = static_cast(nrow < config.bin_construct_sample_cnt ? nrow : config.bin_construct_sample_cnt); auto sample_indices = rand.Sample(nrow, sample_cnt); sample_cnt = static_cast(sample_indices.size()); std::vector> sample_values(num_col); std::vector> sample_idx(num_col); for (size_t i = 0; i < sample_indices.size(); ++i) { auto idx = sample_indices[i]; auto row = get_row_fun(static_cast(idx)); for (std::pair& inner_data : row) { CHECK(inner_data.first < num_col); if (std::fabs(inner_data.second) > kZeroThreshold || std::isnan(inner_data.second)) { sample_values[inner_data.first].emplace_back(inner_data.second); sample_idx[inner_data.first].emplace_back(static_cast(i)); } } } DatasetLoader loader(config, nullptr, 1, nullptr); ret.reset(loader.CostructFromSampleData(Common::Vector2Ptr(&sample_values).data(), Common::Vector2Ptr(&sample_idx).data(), static_cast(num_col), Common::VectorSize(sample_values).data(), sample_cnt, nrow)); } else { ret.reset(new Dataset(nrow)); ret->CreateValid( reinterpret_cast(reference)); } OMP_INIT_EX(); #pragma omp parallel for schedule(static) for (int i = 0; i < nindptr - 1; ++i) { OMP_LOOP_EX_BEGIN(); const int tid = omp_get_thread_num(); auto one_row = get_row_fun(i); ret->PushOneRow(tid, i, one_row); OMP_LOOP_EX_END(); } OMP_THROW_EX(); ret->FinishLoad(); *out = ret.release(); API_END(); } int LGBM_DatasetCreateFromCSRFunc(void* get_row_funptr, int num_rows, int64_t num_col, const char* parameters, const DatasetHandle reference, DatasetHandle* out) { API_BEGIN(); if (num_col <= 0) { Log::Fatal("The number of columns should be greater than zero."); } else if (num_col >= INT32_MAX) { Log::Fatal("The number of columns should be smaller than INT32_MAX."); } auto get_row_fun = *static_cast>&)>*>(get_row_funptr); auto param = Config::Str2Map(parameters); Config config; config.Set(param); if (config.num_threads > 0) { omp_set_num_threads(config.num_threads); } std::unique_ptr ret; int32_t nrow = num_rows; if (reference == nullptr) { // sample data first Random rand(config.data_random_seed); int sample_cnt = static_cast(nrow < config.bin_construct_sample_cnt ? nrow : config.bin_construct_sample_cnt); auto sample_indices = rand.Sample(nrow, sample_cnt); sample_cnt = static_cast(sample_indices.size()); std::vector> sample_values(num_col); std::vector> sample_idx(num_col); // local buffer to re-use memory std::vector> buffer; for (size_t i = 0; i < sample_indices.size(); ++i) { auto idx = sample_indices[i]; get_row_fun(static_cast(idx), buffer); for (std::pair& inner_data : buffer) { CHECK(inner_data.first < num_col); if (std::fabs(inner_data.second) > kZeroThreshold || std::isnan(inner_data.second)) { sample_values[inner_data.first].emplace_back(inner_data.second); sample_idx[inner_data.first].emplace_back(static_cast(i)); } } } DatasetLoader loader(config, nullptr, 1, nullptr); ret.reset(loader.CostructFromSampleData(Common::Vector2Ptr(&sample_values).data(), Common::Vector2Ptr(&sample_idx).data(), static_cast(num_col), Common::VectorSize(sample_values).data(), sample_cnt, nrow)); } else { ret.reset(new Dataset(nrow)); ret->CreateValid( reinterpret_cast(reference)); } OMP_INIT_EX(); std::vector> thread_buffer; #pragma omp parallel for schedule(static) private(thread_buffer) for (int i = 0; i < num_rows; ++i) { OMP_LOOP_EX_BEGIN(); { const int tid = omp_get_thread_num(); get_row_fun(i, thread_buffer); ret->PushOneRow(tid, i, thread_buffer); } OMP_LOOP_EX_END(); } OMP_THROW_EX(); ret->FinishLoad(); *out = ret.release(); API_END(); } int LGBM_DatasetCreateFromCSC(const void* col_ptr, int col_ptr_type, const int32_t* indices, const void* data, int data_type, int64_t ncol_ptr, int64_t nelem, int64_t num_row, const char* parameters, const DatasetHandle reference, DatasetHandle* out) { API_BEGIN(); auto param = Config::Str2Map(parameters); Config config; config.Set(param); if (config.num_threads > 0) { omp_set_num_threads(config.num_threads); } std::unique_ptr ret; int32_t nrow = static_cast(num_row); if (reference == nullptr) { // sample data first Random rand(config.data_random_seed); int sample_cnt = static_cast(nrow < config.bin_construct_sample_cnt ? nrow : config.bin_construct_sample_cnt); auto sample_indices = rand.Sample(nrow, sample_cnt); sample_cnt = static_cast(sample_indices.size()); std::vector> sample_values(ncol_ptr - 1); std::vector> sample_idx(ncol_ptr - 1); OMP_INIT_EX(); #pragma omp parallel for schedule(static) for (int i = 0; i < static_cast(sample_values.size()); ++i) { OMP_LOOP_EX_BEGIN(); CSC_RowIterator col_it(col_ptr, col_ptr_type, indices, data, data_type, ncol_ptr, nelem, i); for (int j = 0; j < sample_cnt; j++) { auto val = col_it.Get(sample_indices[j]); if (std::fabs(val) > kZeroThreshold || std::isnan(val)) { sample_values[i].emplace_back(val); sample_idx[i].emplace_back(j); } } OMP_LOOP_EX_END(); } OMP_THROW_EX(); DatasetLoader loader(config, nullptr, 1, nullptr); ret.reset(loader.CostructFromSampleData(Common::Vector2Ptr(&sample_values).data(), Common::Vector2Ptr(&sample_idx).data(), static_cast(sample_values.size()), Common::VectorSize(sample_values).data(), sample_cnt, nrow)); } else { ret.reset(new Dataset(nrow)); ret->CreateValid( reinterpret_cast(reference)); } OMP_INIT_EX(); #pragma omp parallel for schedule(static) for (int i = 0; i < ncol_ptr - 1; ++i) { OMP_LOOP_EX_BEGIN(); const int tid = omp_get_thread_num(); int feature_idx = ret->InnerFeatureIndex(i); if (feature_idx < 0) { continue; } int group = ret->Feature2Group(feature_idx); int sub_feature = ret->Feture2SubFeature(feature_idx); CSC_RowIterator col_it(col_ptr, col_ptr_type, indices, data, data_type, ncol_ptr, nelem, i); auto bin_mapper = ret->FeatureBinMapper(feature_idx); if (bin_mapper->GetDefaultBin() == bin_mapper->GetMostFreqBin()) { int row_idx = 0; while (row_idx < nrow) { auto pair = col_it.NextNonZero(); row_idx = pair.first; // no more data if (row_idx < 0) { break; } ret->PushOneData(tid, row_idx, group, sub_feature, pair.second); } } else { for (int row_idx = 0; row_idx < nrow; ++row_idx) { auto val = col_it.Get(row_idx); ret->PushOneData(tid, row_idx, group, sub_feature, val); } } OMP_LOOP_EX_END(); } OMP_THROW_EX(); ret->FinishLoad(); *out = ret.release(); API_END(); } int LGBM_DatasetGetSubset( const DatasetHandle handle, const int32_t* used_row_indices, int32_t num_used_row_indices, const char* parameters, DatasetHandle* out) { API_BEGIN(); auto param = Config::Str2Map(parameters); Config config; config.Set(param); if (config.num_threads > 0) { omp_set_num_threads(config.num_threads); } auto full_dataset = reinterpret_cast(handle); CHECK_GT(num_used_row_indices, 0); const int32_t lower = 0; const int32_t upper = full_dataset->num_data() - 1; Common::CheckElementsIntervalClosed(used_row_indices, lower, upper, num_used_row_indices, "Used indices of subset"); if (!std::is_sorted(used_row_indices, used_row_indices + num_used_row_indices)) { Log::Fatal("used_row_indices should be sorted in Subset"); } auto ret = std::unique_ptr(new Dataset(num_used_row_indices)); ret->CopyFeatureMapperFrom(full_dataset); ret->CopySubset(full_dataset, used_row_indices, num_used_row_indices, true); *out = ret.release(); API_END(); } int LGBM_DatasetSetFeatureNames( DatasetHandle handle, const char** feature_names, int num_feature_names) { API_BEGIN(); auto dataset = reinterpret_cast(handle); std::vector feature_names_str; for (int i = 0; i < num_feature_names; ++i) { feature_names_str.emplace_back(feature_names[i]); } dataset->set_feature_names(feature_names_str); API_END(); } int LGBM_DatasetGetFeatureNames( DatasetHandle handle, char** feature_names, int* num_feature_names) { API_BEGIN(); auto dataset = reinterpret_cast(handle); auto inside_feature_name = dataset->feature_names(); *num_feature_names = static_cast(inside_feature_name.size()); for (int i = 0; i < *num_feature_names; ++i) { std::memcpy(feature_names[i], inside_feature_name[i].c_str(), inside_feature_name[i].size() + 1); } API_END(); } #pragma warning(disable : 4702) int LGBM_DatasetFree(DatasetHandle handle) { API_BEGIN(); delete reinterpret_cast(handle); API_END(); } int LGBM_DatasetSaveBinary(DatasetHandle handle, const char* filename) { API_BEGIN(); auto dataset = reinterpret_cast(handle); dataset->SaveBinaryFile(filename); API_END(); } int LGBM_DatasetDumpText(DatasetHandle handle, const char* filename) { API_BEGIN(); auto dataset = reinterpret_cast(handle); dataset->DumpTextFile(filename); API_END(); } int LGBM_DatasetSetField(DatasetHandle handle, const char* field_name, const void* field_data, int num_element, int type) { API_BEGIN(); auto dataset = reinterpret_cast(handle); bool is_success = false; if (type == C_API_DTYPE_FLOAT32) { is_success = dataset->SetFloatField(field_name, reinterpret_cast(field_data), static_cast(num_element)); } else if (type == C_API_DTYPE_INT32) { is_success = dataset->SetIntField(field_name, reinterpret_cast(field_data), static_cast(num_element)); } else if (type == C_API_DTYPE_FLOAT64) { is_success = dataset->SetDoubleField(field_name, reinterpret_cast(field_data), static_cast(num_element)); } if (!is_success) { Log::Fatal("Input data type error or field not found"); } API_END(); } int LGBM_DatasetGetField(DatasetHandle handle, const char* field_name, int* out_len, const void** out_ptr, int* out_type) { API_BEGIN(); auto dataset = reinterpret_cast(handle); bool is_success = false; if (dataset->GetFloatField(field_name, out_len, reinterpret_cast(out_ptr))) { *out_type = C_API_DTYPE_FLOAT32; is_success = true; } else if (dataset->GetIntField(field_name, out_len, reinterpret_cast(out_ptr))) { *out_type = C_API_DTYPE_INT32; is_success = true; } else if (dataset->GetDoubleField(field_name, out_len, reinterpret_cast(out_ptr))) { *out_type = C_API_DTYPE_FLOAT64; is_success = true; } if (!is_success) { Log::Fatal("Field not found"); } if (*out_ptr == nullptr) { *out_len = 0; } API_END(); } int LGBM_DatasetUpdateParamChecking(const char* old_parameters, const char* new_parameters) { API_BEGIN(); auto old_param = Config::Str2Map(old_parameters); Config old_config; old_config.Set(old_param); auto new_param = Config::Str2Map(new_parameters); Booster::CheckDatasetResetConfig(old_config, new_param); API_END(); } int LGBM_DatasetGetNumData(DatasetHandle handle, int* out) { API_BEGIN(); auto dataset = reinterpret_cast(handle); *out = dataset->num_data(); API_END(); } int LGBM_DatasetGetNumFeature(DatasetHandle handle, int* out) { API_BEGIN(); auto dataset = reinterpret_cast(handle); *out = dataset->num_total_features(); API_END(); } int LGBM_DatasetAddFeaturesFrom(DatasetHandle target, DatasetHandle source) { API_BEGIN(); auto target_d = reinterpret_cast(target); auto source_d = reinterpret_cast(source); target_d->AddFeaturesFrom(source_d); API_END(); } // ---- start of booster int LGBM_BoosterCreate(const DatasetHandle train_data, const char* parameters, BoosterHandle* out) { API_BEGIN(); const Dataset* p_train_data = reinterpret_cast(train_data); auto ret = std::unique_ptr(new Booster(p_train_data, parameters)); *out = ret.release(); API_END(); } int LGBM_BoosterCreateFromModelfile( const char* filename, int* out_num_iterations, BoosterHandle* out) { API_BEGIN(); auto ret = std::unique_ptr(new Booster(filename)); *out_num_iterations = ret->GetBoosting()->GetCurrentIteration(); *out = ret.release(); API_END(); } int LGBM_BoosterLoadModelFromString( const char* model_str, int* out_num_iterations, BoosterHandle* out) { API_BEGIN(); auto ret = std::unique_ptr(new Booster(nullptr)); ret->LoadModelFromString(model_str); *out_num_iterations = ret->GetBoosting()->GetCurrentIteration(); *out = ret.release(); API_END(); } #pragma warning(disable : 4702) int LGBM_BoosterFree(BoosterHandle handle) { API_BEGIN(); delete reinterpret_cast(handle); API_END(); } int LGBM_BoosterShuffleModels(BoosterHandle handle, int start_iter, int end_iter) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); ref_booster->ShuffleModels(start_iter, end_iter); API_END(); } int LGBM_BoosterMerge(BoosterHandle handle, BoosterHandle other_handle) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); Booster* ref_other_booster = reinterpret_cast(other_handle); ref_booster->MergeFrom(ref_other_booster); API_END(); } int LGBM_BoosterAddValidData(BoosterHandle handle, const DatasetHandle valid_data) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); const Dataset* p_dataset = reinterpret_cast(valid_data); ref_booster->AddValidData(p_dataset); API_END(); } int LGBM_BoosterResetTrainingData(BoosterHandle handle, const DatasetHandle train_data) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); const Dataset* p_dataset = reinterpret_cast(train_data); ref_booster->ResetTrainingData(p_dataset); API_END(); } int LGBM_BoosterResetParameter(BoosterHandle handle, const char* parameters) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); ref_booster->ResetConfig(parameters); API_END(); } int LGBM_BoosterGetNumClasses(BoosterHandle handle, int* out_len) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); *out_len = ref_booster->GetBoosting()->NumberOfClasses(); API_END(); } int LGBM_BoosterRefit(BoosterHandle handle, const int32_t* leaf_preds, int32_t nrow, int32_t ncol) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); ref_booster->Refit(leaf_preds, nrow, ncol); API_END(); } int LGBM_BoosterUpdateOneIter(BoosterHandle handle, int* is_finished) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); if (ref_booster->TrainOneIter()) { *is_finished = 1; } else { *is_finished = 0; } API_END(); } int LGBM_BoosterUpdateOneIterCustom(BoosterHandle handle, const float* grad, const float* hess, int* is_finished) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); #ifdef SCORE_T_USE_DOUBLE Log::Fatal("Don't support custom loss function when SCORE_T_USE_DOUBLE is enabled"); #else if (ref_booster->TrainOneIter(grad, hess)) { *is_finished = 1; } else { *is_finished = 0; } #endif API_END(); } int LGBM_BoosterRollbackOneIter(BoosterHandle handle) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); ref_booster->RollbackOneIter(); API_END(); } int LGBM_BoosterGetCurrentIteration(BoosterHandle handle, int* out_iteration) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); *out_iteration = ref_booster->GetBoosting()->GetCurrentIteration(); API_END(); } int LGBM_BoosterNumModelPerIteration(BoosterHandle handle, int* out_tree_per_iteration) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); *out_tree_per_iteration = ref_booster->GetBoosting()->NumModelPerIteration(); API_END(); } int LGBM_BoosterNumberOfTotalModel(BoosterHandle handle, int* out_models) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); *out_models = ref_booster->GetBoosting()->NumberOfTotalModel(); API_END(); } int LGBM_BoosterGetEvalCounts(BoosterHandle handle, int* out_len) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); *out_len = ref_booster->GetEvalCounts(); API_END(); } int LGBM_BoosterGetEvalNames(BoosterHandle handle, int* out_len, char** out_strs) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); *out_len = ref_booster->GetEvalNames(out_strs); API_END(); } int LGBM_BoosterGetFeatureNames(BoosterHandle handle, int* out_len, char** out_strs) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); *out_len = ref_booster->GetFeatureNames(out_strs); API_END(); } int LGBM_BoosterGetNumFeature(BoosterHandle handle, int* out_len) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); *out_len = ref_booster->GetBoosting()->MaxFeatureIdx() + 1; API_END(); } int LGBM_BoosterGetEval(BoosterHandle handle, int data_idx, int* out_len, double* out_results) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); auto boosting = ref_booster->GetBoosting(); auto result_buf = boosting->GetEvalAt(data_idx); *out_len = static_cast(result_buf.size()); for (size_t i = 0; i < result_buf.size(); ++i) { (out_results)[i] = static_cast(result_buf[i]); } API_END(); } int LGBM_BoosterGetNumPredict(BoosterHandle handle, int data_idx, int64_t* out_len) { API_BEGIN(); auto boosting = reinterpret_cast(handle)->GetBoosting(); *out_len = boosting->GetNumPredictAt(data_idx); API_END(); } int LGBM_BoosterGetPredict(BoosterHandle handle, int data_idx, int64_t* out_len, double* out_result) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); ref_booster->GetPredictAt(data_idx, out_result, out_len); API_END(); } int LGBM_BoosterPredictForFile(BoosterHandle handle, const char* data_filename, int data_has_header, int predict_type, int num_iteration, const char* parameter, const char* result_filename) { API_BEGIN(); auto param = Config::Str2Map(parameter); Config config; config.Set(param); if (config.num_threads > 0) { omp_set_num_threads(config.num_threads); } Booster* ref_booster = reinterpret_cast(handle); ref_booster->Predict(num_iteration, predict_type, data_filename, data_has_header, config, result_filename); API_END(); } int LGBM_BoosterCalcNumPredict(BoosterHandle handle, int num_row, int predict_type, int num_iteration, int64_t* out_len) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); *out_len = static_cast(num_row) * ref_booster->GetBoosting()->NumPredictOneRow( num_iteration, predict_type == C_API_PREDICT_LEAF_INDEX, predict_type == C_API_PREDICT_CONTRIB); API_END(); } int LGBM_BoosterPredictForCSR(BoosterHandle handle, const void* indptr, int indptr_type, const int32_t* indices, const void* data, int data_type, int64_t nindptr, int64_t nelem, int64_t num_col, int predict_type, int num_iteration, const char* parameter, int64_t* out_len, double* out_result) { API_BEGIN(); if (num_col <= 0) { Log::Fatal("The number of columns should be greater than zero."); } else if (num_col >= INT32_MAX) { Log::Fatal("The number of columns should be smaller than INT32_MAX."); } auto param = Config::Str2Map(parameter); Config config; config.Set(param); if (config.num_threads > 0) { omp_set_num_threads(config.num_threads); } Booster* ref_booster = reinterpret_cast(handle); auto get_row_fun = RowFunctionFromCSR(indptr, indptr_type, indices, data, data_type, nindptr, nelem); int nrow = static_cast(nindptr - 1); ref_booster->Predict(num_iteration, predict_type, nrow, static_cast(num_col), get_row_fun, config, out_result, out_len); API_END(); } int LGBM_BoosterPredictForCSRSingleRow(BoosterHandle handle, const void* indptr, int indptr_type, const int32_t* indices, const void* data, int data_type, int64_t nindptr, int64_t nelem, int64_t num_col, int predict_type, int num_iteration, const char* parameter, int64_t* out_len, double* out_result) { API_BEGIN(); if (num_col <= 0) { Log::Fatal("The number of columns should be greater than zero."); } else if (num_col >= INT32_MAX) { Log::Fatal("The number of columns should be smaller than INT32_MAX."); } auto param = Config::Str2Map(parameter); Config config; config.Set(param); if (config.num_threads > 0) { omp_set_num_threads(config.num_threads); } Booster* ref_booster = reinterpret_cast(handle); auto get_row_fun = RowFunctionFromCSR(indptr, indptr_type, indices, data, data_type, nindptr, nelem); ref_booster->PredictSingleRow(num_iteration, predict_type, static_cast(num_col), get_row_fun, config, out_result, out_len); API_END(); } int LGBM_BoosterPredictForCSC(BoosterHandle handle, const void* col_ptr, int col_ptr_type, const int32_t* indices, const void* data, int data_type, int64_t ncol_ptr, int64_t nelem, int64_t num_row, int predict_type, int num_iteration, const char* parameter, int64_t* out_len, double* out_result) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); auto param = Config::Str2Map(parameter); Config config; config.Set(param); if (config.num_threads > 0) { omp_set_num_threads(config.num_threads); } int num_threads = OMP_NUM_THREADS(); int ncol = static_cast(ncol_ptr - 1); std::vector> iterators(num_threads, std::vector()); for (int i = 0; i < num_threads; ++i) { for (int j = 0; j < ncol; ++j) { iterators[i].emplace_back(col_ptr, col_ptr_type, indices, data, data_type, ncol_ptr, nelem, j); } } std::function>(int row_idx)> get_row_fun = [&iterators, ncol](int i) { std::vector> one_row; one_row.reserve(ncol); const int tid = omp_get_thread_num(); for (int j = 0; j < ncol; ++j) { auto val = iterators[tid][j].Get(i); if (std::fabs(val) > kZeroThreshold || std::isnan(val)) { one_row.emplace_back(j, val); } } return one_row; }; ref_booster->Predict(num_iteration, predict_type, static_cast(num_row), ncol, get_row_fun, config, out_result, out_len); API_END(); } int LGBM_BoosterPredictForMat(BoosterHandle handle, const void* data, int data_type, int32_t nrow, int32_t ncol, int is_row_major, int predict_type, int num_iteration, const char* parameter, int64_t* out_len, double* out_result) { API_BEGIN(); auto param = Config::Str2Map(parameter); Config config; config.Set(param); if (config.num_threads > 0) { omp_set_num_threads(config.num_threads); } Booster* ref_booster = reinterpret_cast(handle); auto get_row_fun = RowPairFunctionFromDenseMatric(data, nrow, ncol, data_type, is_row_major); ref_booster->Predict(num_iteration, predict_type, nrow, ncol, get_row_fun, config, out_result, out_len); API_END(); } int LGBM_BoosterPredictForMatSingleRow(BoosterHandle handle, const void* data, int data_type, int32_t ncol, int is_row_major, int predict_type, int num_iteration, const char* parameter, int64_t* out_len, double* out_result) { API_BEGIN(); auto param = Config::Str2Map(parameter); Config config; config.Set(param); if (config.num_threads > 0) { omp_set_num_threads(config.num_threads); } Booster* ref_booster = reinterpret_cast(handle); auto get_row_fun = RowPairFunctionFromDenseMatric(data, 1, ncol, data_type, is_row_major); ref_booster->PredictSingleRow(num_iteration, predict_type, ncol, get_row_fun, config, out_result, out_len); API_END(); } int LGBM_BoosterPredictForMats(BoosterHandle handle, const void** data, int data_type, int32_t nrow, int32_t ncol, int predict_type, int num_iteration, const char* parameter, int64_t* out_len, double* out_result) { API_BEGIN(); auto param = Config::Str2Map(parameter); Config config; config.Set(param); if (config.num_threads > 0) { omp_set_num_threads(config.num_threads); } Booster* ref_booster = reinterpret_cast(handle); auto get_row_fun = RowPairFunctionFromDenseRows(data, ncol, data_type); ref_booster->Predict(num_iteration, predict_type, nrow, ncol, get_row_fun, config, out_result, out_len); API_END(); } int LGBM_BoosterSaveModel(BoosterHandle handle, int start_iteration, int num_iteration, const char* filename) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); ref_booster->SaveModelToFile(start_iteration, num_iteration, filename); API_END(); } int LGBM_BoosterSaveModelToString(BoosterHandle handle, int start_iteration, int num_iteration, int64_t buffer_len, int64_t* out_len, char* out_str) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); std::string model = ref_booster->SaveModelToString(start_iteration, num_iteration); *out_len = static_cast(model.size()) + 1; if (*out_len <= buffer_len) { std::memcpy(out_str, model.c_str(), *out_len); } API_END(); } int LGBM_BoosterDumpModel(BoosterHandle handle, int start_iteration, int num_iteration, int64_t buffer_len, int64_t* out_len, char* out_str) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); std::string model = ref_booster->DumpModel(start_iteration, num_iteration); *out_len = static_cast(model.size()) + 1; if (*out_len <= buffer_len) { std::memcpy(out_str, model.c_str(), *out_len); } API_END(); } int LGBM_BoosterGetLeafValue(BoosterHandle handle, int tree_idx, int leaf_idx, double* out_val) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); *out_val = static_cast(ref_booster->GetLeafValue(tree_idx, leaf_idx)); API_END(); } int LGBM_BoosterSetLeafValue(BoosterHandle handle, int tree_idx, int leaf_idx, double val) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); ref_booster->SetLeafValue(tree_idx, leaf_idx, val); API_END(); } int LGBM_BoosterFeatureImportance(BoosterHandle handle, int num_iteration, int importance_type, double* out_results) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); std::vector feature_importances = ref_booster->FeatureImportance(num_iteration, importance_type); for (size_t i = 0; i < feature_importances.size(); ++i) { (out_results)[i] = feature_importances[i]; } API_END(); } int LGBM_BoosterGetUpperBoundValue(BoosterHandle handle, double* out_results) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); double max_value = ref_booster->UpperBoundValue(); *out_results = max_value; API_END(); } int LGBM_BoosterGetLowerBoundValue(BoosterHandle handle, double* out_results) { API_BEGIN(); Booster* ref_booster = reinterpret_cast(handle); double min_value = ref_booster->LowerBoundValue(); *out_results = min_value; API_END(); } int LGBM_NetworkInit(const char* machines, int local_listen_port, int listen_time_out, int num_machines) { API_BEGIN(); Config config; config.machines = Common::RemoveQuotationSymbol(std::string(machines)); config.local_listen_port = local_listen_port; config.num_machines = num_machines; config.time_out = listen_time_out; if (num_machines > 1) { Network::Init(config); } API_END(); } int LGBM_NetworkFree() { API_BEGIN(); Network::Dispose(); API_END(); } int LGBM_NetworkInitWithFunctions(int num_machines, int rank, void* reduce_scatter_ext_fun, void* allgather_ext_fun) { API_BEGIN(); if (num_machines > 1) { Network::Init(num_machines, rank, (ReduceScatterFunction)reduce_scatter_ext_fun, (AllgatherFunction)allgather_ext_fun); } API_END(); } // ---- start of some help functions std::function(int row_idx)> RowFunctionFromDenseMatric(const void* data, int num_row, int num_col, int data_type, int is_row_major) { if (data_type == C_API_DTYPE_FLOAT32) { const float* data_ptr = reinterpret_cast(data); if (is_row_major) { return [=] (int row_idx) { std::vector ret(num_col); auto tmp_ptr = data_ptr + static_cast(num_col) * row_idx; for (int i = 0; i < num_col; ++i) { ret[i] = static_cast(*(tmp_ptr + i)); } return ret; }; } else { return [=] (int row_idx) { std::vector ret(num_col); for (int i = 0; i < num_col; ++i) { ret[i] = static_cast(*(data_ptr + static_cast(num_row) * i + row_idx)); } return ret; }; } } else if (data_type == C_API_DTYPE_FLOAT64) { const double* data_ptr = reinterpret_cast(data); if (is_row_major) { return [=] (int row_idx) { std::vector ret(num_col); auto tmp_ptr = data_ptr + static_cast(num_col) * row_idx; for (int i = 0; i < num_col; ++i) { ret[i] = static_cast(*(tmp_ptr + i)); } return ret; }; } else { return [=] (int row_idx) { std::vector ret(num_col); for (int i = 0; i < num_col; ++i) { ret[i] = static_cast(*(data_ptr + static_cast(num_row) * i + row_idx)); } return ret; }; } } Log::Fatal("Unknown data type in RowFunctionFromDenseMatric"); } std::function>(int row_idx)> RowPairFunctionFromDenseMatric(const void* data, int num_row, int num_col, int data_type, int is_row_major) { auto inner_function = RowFunctionFromDenseMatric(data, num_row, num_col, data_type, is_row_major); if (inner_function != nullptr) { return [inner_function] (int row_idx) { auto raw_values = inner_function(row_idx); std::vector> ret; ret.reserve(raw_values.size()); for (int i = 0; i < static_cast(raw_values.size()); ++i) { if (std::fabs(raw_values[i]) > kZeroThreshold || std::isnan(raw_values[i])) { ret.emplace_back(i, raw_values[i]); } } return ret; }; } return nullptr; } // data is array of pointers to individual rows std::function>(int row_idx)> RowPairFunctionFromDenseRows(const void** data, int num_col, int data_type) { return [=](int row_idx) { auto inner_function = RowFunctionFromDenseMatric(data[row_idx], 1, num_col, data_type, /* is_row_major */ true); auto raw_values = inner_function(0); std::vector> ret; ret.reserve(raw_values.size()); for (int i = 0; i < static_cast(raw_values.size()); ++i) { if (std::fabs(raw_values[i]) > kZeroThreshold || std::isnan(raw_values[i])) { ret.emplace_back(i, raw_values[i]); } } return ret; }; } std::function>(int idx)> RowFunctionFromCSR(const void* indptr, int indptr_type, const int32_t* indices, const void* data, int data_type, int64_t , int64_t ) { if (data_type == C_API_DTYPE_FLOAT32) { const float* data_ptr = reinterpret_cast(data); if (indptr_type == C_API_DTYPE_INT32) { const int32_t* ptr_indptr = reinterpret_cast(indptr); return [=] (int idx) { std::vector> ret; int64_t start = ptr_indptr[idx]; int64_t end = ptr_indptr[idx + 1]; if (end - start > 0) { ret.reserve(end - start); } for (int64_t i = start; i < end; ++i) { ret.emplace_back(indices[i], data_ptr[i]); } return ret; }; } else if (indptr_type == C_API_DTYPE_INT64) { const int64_t* ptr_indptr = reinterpret_cast(indptr); return [=] (int idx) { std::vector> ret; int64_t start = ptr_indptr[idx]; int64_t end = ptr_indptr[idx + 1]; if (end - start > 0) { ret.reserve(end - start); } for (int64_t i = start; i < end; ++i) { ret.emplace_back(indices[i], data_ptr[i]); } return ret; }; } } else if (data_type == C_API_DTYPE_FLOAT64) { const double* data_ptr = reinterpret_cast(data); if (indptr_type == C_API_DTYPE_INT32) { const int32_t* ptr_indptr = reinterpret_cast(indptr); return [=] (int idx) { std::vector> ret; int64_t start = ptr_indptr[idx]; int64_t end = ptr_indptr[idx + 1]; if (end - start > 0) { ret.reserve(end - start); } for (int64_t i = start; i < end; ++i) { ret.emplace_back(indices[i], data_ptr[i]); } return ret; }; } else if (indptr_type == C_API_DTYPE_INT64) { const int64_t* ptr_indptr = reinterpret_cast(indptr); return [=] (int idx) { std::vector> ret; int64_t start = ptr_indptr[idx]; int64_t end = ptr_indptr[idx + 1]; if (end - start > 0) { ret.reserve(end - start); } for (int64_t i = start; i < end; ++i) { ret.emplace_back(indices[i], data_ptr[i]); } return ret; }; } } Log::Fatal("Unknown data type in RowFunctionFromCSR"); } std::function(int idx)> IterateFunctionFromCSC(const void* col_ptr, int col_ptr_type, const int32_t* indices, const void* data, int data_type, int64_t ncol_ptr, int64_t , int col_idx) { CHECK(col_idx < ncol_ptr && col_idx >= 0); if (data_type == C_API_DTYPE_FLOAT32) { const float* data_ptr = reinterpret_cast(data); if (col_ptr_type == C_API_DTYPE_INT32) { const int32_t* ptr_col_ptr = reinterpret_cast(col_ptr); int64_t start = ptr_col_ptr[col_idx]; int64_t end = ptr_col_ptr[col_idx + 1]; return [=] (int offset) { int64_t i = static_cast(start + offset); if (i >= end) { return std::make_pair(-1, 0.0); } int idx = static_cast(indices[i]); double val = static_cast(data_ptr[i]); return std::make_pair(idx, val); }; } else if (col_ptr_type == C_API_DTYPE_INT64) { const int64_t* ptr_col_ptr = reinterpret_cast(col_ptr); int64_t start = ptr_col_ptr[col_idx]; int64_t end = ptr_col_ptr[col_idx + 1]; return [=] (int offset) { int64_t i = static_cast(start + offset); if (i >= end) { return std::make_pair(-1, 0.0); } int idx = static_cast(indices[i]); double val = static_cast(data_ptr[i]); return std::make_pair(idx, val); }; } } else if (data_type == C_API_DTYPE_FLOAT64) { const double* data_ptr = reinterpret_cast(data); if (col_ptr_type == C_API_DTYPE_INT32) { const int32_t* ptr_col_ptr = reinterpret_cast(col_ptr); int64_t start = ptr_col_ptr[col_idx]; int64_t end = ptr_col_ptr[col_idx + 1]; return [=] (int offset) { int64_t i = static_cast(start + offset); if (i >= end) { return std::make_pair(-1, 0.0); } int idx = static_cast(indices[i]); double val = static_cast(data_ptr[i]); return std::make_pair(idx, val); }; } else if (col_ptr_type == C_API_DTYPE_INT64) { const int64_t* ptr_col_ptr = reinterpret_cast(col_ptr); int64_t start = ptr_col_ptr[col_idx]; int64_t end = ptr_col_ptr[col_idx + 1]; return [=] (int offset) { int64_t i = static_cast(start + offset); if (i >= end) { return std::make_pair(-1, 0.0); } int idx = static_cast(indices[i]); double val = static_cast(data_ptr[i]); return std::make_pair(idx, val); }; } } Log::Fatal("Unknown data type in CSC matrix"); } CSC_RowIterator::CSC_RowIterator(const void* col_ptr, int col_ptr_type, const int32_t* indices, const void* data, int data_type, int64_t ncol_ptr, int64_t nelem, int col_idx) { iter_fun_ = IterateFunctionFromCSC(col_ptr, col_ptr_type, indices, data, data_type, ncol_ptr, nelem, col_idx); } double CSC_RowIterator::Get(int idx) { while (idx > cur_idx_ && !is_end_) { auto ret = iter_fun_(nonzero_idx_); if (ret.first < 0) { is_end_ = true; break; } cur_idx_ = ret.first; cur_val_ = ret.second; ++nonzero_idx_; } if (idx == cur_idx_) { return cur_val_; } else { return 0.0f; } } std::pair CSC_RowIterator::NextNonZero() { if (!is_end_) { auto ret = iter_fun_(nonzero_idx_); ++nonzero_idx_; if (ret.first < 0) { is_end_ = true; } return ret; } else { return std::make_pair(-1, 0.0); } }