/*! * Copyright (c) 2016 Microsoft Corporation. All rights reserved. * Licensed under the MIT License. See LICENSE file in the project root for license information. */ #include #include #include #include #include namespace LightGBM { void Config::KV2Map(std::unordered_map* params, const char* kv) { std::vector tmp_strs = Common::Split(kv, '='); if (tmp_strs.size() == 2 || tmp_strs.size() == 1) { std::string key = Common::RemoveQuotationSymbol(Common::Trim(tmp_strs[0])); std::string value = ""; if (tmp_strs.size() == 2) { value = Common::RemoveQuotationSymbol(Common::Trim(tmp_strs[1])); } if (key.size() > 0) { auto value_search = params->find(key); if (value_search == params->end()) { // not set params->emplace(key, value); } else { Log::Warning("%s is set=%s, %s=%s will be ignored. Current value: %s=%s", key.c_str(), value_search->second.c_str(), key.c_str(), value.c_str(), key.c_str(), value_search->second.c_str()); } } } else { Log::Warning("Unknown parameter %s", kv); } } std::unordered_map Config::Str2Map(const char* parameters) { std::unordered_map params; auto args = Common::Split(parameters, " \t\n\r"); for (auto arg : args) { KV2Map(¶ms, Common::Trim(arg).c_str()); } ParameterAlias::KeyAliasTransform(¶ms); return params; } void GetBoostingType(const std::unordered_map& params, std::string* boosting) { std::string value; if (Config::GetString(params, "boosting", &value)) { std::transform(value.begin(), value.end(), value.begin(), Common::tolower); if (value == std::string("gbdt") || value == std::string("gbrt")) { *boosting = "gbdt"; } else if (value == std::string("dart")) { *boosting = "dart"; } else if (value == std::string("goss")) { *boosting = "goss"; } else if (value == std::string("rf") || value == std::string("random_forest")) { *boosting = "rf"; } else { Log::Fatal("Unknown boosting type %s", value.c_str()); } } } void ParseMetrics(const std::string& value, std::vector* out_metric) { std::unordered_set metric_sets; out_metric->clear(); std::vector metrics = Common::Split(value.c_str(), ','); for (auto& met : metrics) { auto type = ParseMetricAlias(met); if (metric_sets.count(type) <= 0) { out_metric->push_back(type); metric_sets.insert(type); } } } void GetObjectiveType(const std::unordered_map& params, std::string* objective) { std::string value; if (Config::GetString(params, "objective", &value)) { std::transform(value.begin(), value.end(), value.begin(), Common::tolower); *objective = ParseObjectiveAlias(value); } } void GetMetricType(const std::unordered_map& params, std::vector* metric) { std::string value; if (Config::GetString(params, "metric", &value)) { std::transform(value.begin(), value.end(), value.begin(), Common::tolower); ParseMetrics(value, metric); } // add names of objective function if not providing metric if (metric->empty() && value.size() == 0) { if (Config::GetString(params, "objective", &value)) { std::transform(value.begin(), value.end(), value.begin(), Common::tolower); ParseMetrics(value, metric); } } } void GetTaskType(const std::unordered_map& params, TaskType* task) { std::string value; if (Config::GetString(params, "task", &value)) { std::transform(value.begin(), value.end(), value.begin(), Common::tolower); if (value == std::string("train") || value == std::string("training")) { *task = TaskType::kTrain; } else if (value == std::string("predict") || value == std::string("prediction") || value == std::string("test")) { *task = TaskType::kPredict; } else if (value == std::string("convert_model")) { *task = TaskType::kConvertModel; } else if (value == std::string("refit") || value == std::string("refit_tree")) { *task = TaskType::KRefitTree; } else { Log::Fatal("Unknown task type %s", value.c_str()); } } } void GetDeviceType(const std::unordered_map& params, std::string* device_type) { std::string value; if (Config::GetString(params, "device_type", &value)) { std::transform(value.begin(), value.end(), value.begin(), Common::tolower); if (value == std::string("cpu")) { *device_type = "cpu"; } else if (value == std::string("gpu")) { *device_type = "gpu"; } else { Log::Fatal("Unknown device type %s", value.c_str()); } } } void GetTreeLearnerType(const std::unordered_map& params, std::string* tree_learner) { std::string value; if (Config::GetString(params, "tree_learner", &value)) { std::transform(value.begin(), value.end(), value.begin(), Common::tolower); if (value == std::string("serial")) { *tree_learner = "serial"; } else if (value == std::string("feature") || value == std::string("feature_parallel")) { *tree_learner = "feature"; } else if (value == std::string("data") || value == std::string("data_parallel")) { *tree_learner = "data"; } else if (value == std::string("voting") || value == std::string("voting_parallel")) { *tree_learner = "voting"; } else { Log::Fatal("Unknown tree learner type %s", value.c_str()); } } } void Config::GetAucMuWeights() { if (auc_mu_weights.empty()) { // equal weights for all classes auc_mu_weights_matrix = std::vector> (num_class, std::vector(num_class, 1)); for (size_t i = 0; i < static_cast(num_class); ++i) { auc_mu_weights_matrix[i][i] = 0; } } else { auc_mu_weights_matrix = std::vector> (num_class, std::vector(num_class, 0)); if (auc_mu_weights.size() != static_cast(num_class * num_class)) { Log::Fatal("auc_mu_weights must have %d elements, but found %d", num_class * num_class, auc_mu_weights.size()); } for (size_t i = 0; i < static_cast(num_class); ++i) { for (size_t j = 0; j < static_cast(num_class); ++j) { if (i == j) { auc_mu_weights_matrix[i][j] = 0; if (std::fabs(auc_mu_weights[i * num_class + j]) > kZeroThreshold) { Log::Info("AUC-mu matrix must have zeros on diagonal. Overwriting value in position %d of auc_mu_weights with 0.", i * num_class + j); } } else { if (std::fabs(auc_mu_weights[i * num_class + j]) < kZeroThreshold) { Log::Fatal("AUC-mu matrix must have non-zero values for non-diagonal entries. Found zero value in position %d of auc_mu_weights.", i * num_class + j); } auc_mu_weights_matrix[i][j] = auc_mu_weights[i * num_class + j]; } } } } } void Config::GetInteractionConstraints() { if (interaction_constraints == "") { interaction_constraints_vector = std::vector>(); } else { interaction_constraints_vector = Common::StringToArrayofArrays(interaction_constraints, '[', ']', ','); } } void Config::Set(const std::unordered_map& params) { // generate seeds by seed. if (GetInt(params, "seed", &seed)) { Random rand(seed); int int_max = std::numeric_limits::max(); data_random_seed = static_cast(rand.NextShort(0, int_max)); bagging_seed = static_cast(rand.NextShort(0, int_max)); drop_seed = static_cast(rand.NextShort(0, int_max)); feature_fraction_seed = static_cast(rand.NextShort(0, int_max)); objective_seed = static_cast(rand.NextShort(0, int_max)); extra_seed = static_cast(rand.NextShort(0, int_max)); } GetTaskType(params, &task); GetBoostingType(params, &boosting); GetMetricType(params, &metric); GetObjectiveType(params, &objective); GetDeviceType(params, &device_type); GetTreeLearnerType(params, &tree_learner); GetMembersFromString(params); GetAucMuWeights(); GetInteractionConstraints(); // sort eval_at std::sort(eval_at.begin(), eval_at.end()); std::vector new_valid; for (size_t i = 0; i < valid.size(); ++i) { if (valid[i] != data) { // Only push the non-training data new_valid.push_back(valid[i]); } else { is_provide_training_metric = true; } } valid = new_valid; // check for conflicts CheckParamConflict(); if (verbosity == 1) { LightGBM::Log::ResetLogLevel(LightGBM::LogLevel::Info); } else if (verbosity == 0) { LightGBM::Log::ResetLogLevel(LightGBM::LogLevel::Warning); } else if (verbosity >= 2) { LightGBM::Log::ResetLogLevel(LightGBM::LogLevel::Debug); } else { LightGBM::Log::ResetLogLevel(LightGBM::LogLevel::Fatal); } } bool CheckMultiClassObjective(const std::string& objective) { return (objective == std::string("multiclass") || objective == std::string("multiclassova")); } void Config::CheckParamConflict() { // check if objective, metric, and num_class match int num_class_check = num_class; bool objective_type_multiclass = CheckMultiClassObjective(objective) || (objective == std::string("custom") && num_class_check > 1); if (objective_type_multiclass) { if (num_class_check <= 1) { Log::Fatal("Number of classes should be specified and greater than 1 for multiclass training"); } } else { if (task == TaskType::kTrain && num_class_check != 1) { Log::Fatal("Number of classes must be 1 for non-multiclass training"); } } for (std::string metric_type : metric) { bool metric_type_multiclass = (CheckMultiClassObjective(metric_type) || metric_type == std::string("multi_logloss") || metric_type == std::string("multi_error") || metric_type == std::string("auc_mu") || (metric_type == std::string("custom") && num_class_check > 1)); if ((objective_type_multiclass && !metric_type_multiclass) || (!objective_type_multiclass && metric_type_multiclass)) { Log::Fatal("Multiclass objective and metrics don't match"); } } if (num_machines > 1) { is_parallel = true; } else { is_parallel = false; tree_learner = "serial"; } bool is_single_tree_learner = tree_learner == std::string("serial"); if (is_single_tree_learner) { is_parallel = false; num_machines = 1; } if (is_single_tree_learner || tree_learner == std::string("feature")) { is_data_based_parallel = false; } else if (tree_learner == std::string("data") || tree_learner == std::string("voting")) { is_data_based_parallel = true; if (histogram_pool_size >= 0 && tree_learner == std::string("data")) { Log::Warning("Histogram LRU queue was enabled (histogram_pool_size=%f).\n" "Will disable this to reduce communication costs", histogram_pool_size); // Change pool size to -1 (no limit) when using data parallel to reduce communication costs histogram_pool_size = -1; } } if (is_data_based_parallel) { if (!forcedsplits_filename.empty()) { Log::Fatal("Don't support forcedsplits in %s tree learner", tree_learner.c_str()); } } // Check max_depth and num_leaves if (max_depth > 0) { double full_num_leaves = std::pow(2, max_depth); if (full_num_leaves > num_leaves && num_leaves == kDefaultNumLeaves) { Log::Warning("Accuracy may be bad since you didn't set num_leaves and 2^max_depth > num_leaves"); } if (full_num_leaves < num_leaves) { // Fits in an int, and is more restrictive than the current num_leaves num_leaves = static_cast(full_num_leaves); } } // force col-wise for gpu if (device_type == std::string("gpu")) { force_col_wise = true; force_row_wise = false; } // min_data_in_leaf must be at least 2 if path smoothing is active. This is because when the split is calculated // the count is calculated using the proportion of hessian in the leaf which is rounded up to nearest int, so it can // be 1 when there is actually no data in the leaf. In rare cases this can cause a bug because with path smoothing the // calculated split gain can be positive even with zero gradient and hessian. if (path_smooth > kEpsilon && min_data_in_leaf < 2) { min_data_in_leaf = 2; Log::Warning("min_data_in_leaf has been increased to 2 because this is required when path smoothing is active."); } if (is_parallel && monotone_constraints_method == std::string("intermediate")) { // In distributed mode, local node doesn't have histograms on all features, cannot perform "intermediate" monotone constraints. Log::Warning("Cannot use \"intermediate\" monotone constraints in parallel learning, auto set to \"basic\" method."); monotone_constraints_method = "basic"; } if (feature_fraction_bynode != 1.0 && monotone_constraints_method == std::string("intermediate")) { // "intermediate" monotone constraints need to recompute splits. If the features are sampled when computing the // split initially, then the sampling needs to be recorded or done once again, which is currently not supported Log::Warning("Cannot use \"intermediate\" monotone constraints with feature fraction different from 1, auto set monotone constraints to \"basic\" method."); monotone_constraints_method = "basic"; } if (max_depth > 0 && monotone_penalty >= max_depth) { Log::Warning("Monotone penalty greater than tree depth. Monotone features won't be used."); } } std::string Config::ToString() const { std::stringstream str_buf; str_buf << "[boosting: " << boosting << "]\n"; str_buf << "[objective: " << objective << "]\n"; str_buf << "[metric: " << Common::Join(metric, ",") << "]\n"; str_buf << "[tree_learner: " << tree_learner << "]\n"; str_buf << "[device_type: " << device_type << "]\n"; str_buf << SaveMembersToString(); return str_buf.str(); } } // namespace LightGBM