Unverified Commit c676a7ea authored by david-cortes's avatar david-cortes Committed by GitHub
Browse files

[R-package] Accept factor labels and use their levels (#5341)

parent 9713ff40
...@@ -63,4 +63,4 @@ Imports: ...@@ -63,4 +63,4 @@ Imports:
utils utils
SystemRequirements: SystemRequirements:
C++11 C++11
RoxygenNote: 7.2.1 RoxygenNote: 7.2.3
...@@ -78,3 +78,27 @@ ...@@ -78,3 +78,27 @@
) )
) )
} }
.MULTICLASS_OBJECTIVES <- function() {
return(
c(
"multi_logloss"
, "multiclass"
, "softmax"
, "multiclassova"
, "multiclass_ova"
, "ova"
, "ovr"
)
)
}
.BINARY_OBJECTIVES <- function() {
return(
c(
"binary_logloss"
, "binary"
, "binary_error"
)
)
}
...@@ -9,6 +9,7 @@ Booster <- R6::R6Class( ...@@ -9,6 +9,7 @@ Booster <- R6::R6Class(
best_score = NA_real_, best_score = NA_real_,
params = list(), params = list(),
record_evals = list(), record_evals = list(),
data_processor = NULL,
# Finalize will free up the handles # Finalize will free up the handles
finalize = function() { finalize = function() {
...@@ -837,6 +838,11 @@ Booster <- R6::R6Class( ...@@ -837,6 +838,11 @@ Booster <- R6::R6Class(
#' #'
#' Note that, if using custom objectives, types "class" and "response" will not be available and will #' Note that, if using custom objectives, types "class" and "response" will not be available and will
#' default towards using "raw" instead. #' default towards using "raw" instead.
#'
#' If the model was fit through function \link{lightgbm} and it was passed a factor as labels,
#' passing the prediction type through \code{params} instead of through this argument might
#' result in factor levels for classification objectives not being applied correctly to the
#' resulting output.
#' @param start_iteration int or None, optional (default=None) #' @param start_iteration int or None, optional (default=None)
#' Start index of the iteration to predict. #' Start index of the iteration to predict.
#' If None or <= 0, starts from the first iteration. #' If None or <= 0, starts from the first iteration.
...@@ -895,6 +901,11 @@ NULL ...@@ -895,6 +901,11 @@ NULL
#' in the order "feature contributions for first class, feature contributions for second class, feature #' in the order "feature contributions for first class, feature contributions for second class, feature
#' contributions for third class, etc.". #' contributions for third class, etc.".
#' #'
#' If the model was fit through function \link{lightgbm} and it was passed a factor as labels, predictions
#' returned from this function will retain the factor levels (either as values for \code{type="class"}, or
#' as column names for \code{type="response"} and \code{type="raw"} for multi-class objectives). Note that
#' passing the requested prediction type under \code{params} instead of through \code{type} might result in
#' the factor levels not being present in the output.
#' @examples #' @examples
#' \donttest{ #' \donttest{
#' data(agaricus.train, package = "lightgbm") #' data(agaricus.train, package = "lightgbm")
...@@ -996,12 +1007,18 @@ predict.lgb.Booster <- function(object, ...@@ -996,12 +1007,18 @@ predict.lgb.Booster <- function(object,
, params = params , params = params
) )
if (type == "class") { if (type == "class") {
if (object$params$objective == "binary") { if (object$params$objective %in% .BINARY_OBJECTIVES()) {
pred <- as.integer(pred >= 0.5) pred <- as.integer(pred >= 0.5)
} else if (object$params$objective %in% c("multiclass", "multiclassova")) { } else if (object$params$objective %in% .MULTICLASS_OBJECTIVES()) {
pred <- max.col(pred) - 1L pred <- max.col(pred) - 1L
} }
} }
if (!is.null(object$data_processor)) {
pred <- object$data_processor$process_predictions(
pred = pred
, type = type
)
}
return(pred) return(pred)
} }
......
DataProcessor <- R6::R6Class(
classname = "lgb.DataProcessor",
public = list(
factor_levels = NULL,
process_label = function(label, objective, params) {
if (is.character(label)) {
label <- factor(label)
}
if (is.factor(label)) {
self$factor_levels <- levels(label)
if (length(self$factor_levels) <= 1L) {
stop("Labels to predict is a factor with <2 possible values.")
}
label <- as.numeric(label) - 1.0
out <- list(label = label)
if (length(self$factor_levels) == 2L) {
if (objective == "auto") {
objective <- "binary"
}
if (!(objective %in% .BINARY_OBJECTIVES())) {
stop("Two-level factors as labels only allowed for objective='binary' or objective='auto'.")
}
} else {
if (objective == "auto") {
objective <- "multiclass"
}
if (!(objective %in% .MULTICLASS_OBJECTIVES())) {
stop(
sprintf(
"Factors with >2 levels as labels only allowed for multi-class objectives. Got: %s (allowed: %s)"
, objective
, toString(.MULTICLASS_OBJECTIVES())
)
)
}
data_num_class <- length(self$factor_levels)
params <- lgb.check.wrapper_param(
main_param_name = "num_class"
, params = params
, alternative_kwarg_value = data_num_class
)
if (params[["num_class"]] != data_num_class) {
warning(
sprintf(
"Found num_class=%d in params, but 'label' is a factor with %d levels. 'num_class' will be ignored."
, params[["num_class"]]
, data_num_class
)
)
params$num_class <- data_num_class
}
}
out$objective <- objective
out$params <- params
return(out)
} else {
label <- as.numeric(label)
if (objective == "auto") {
objective <- "regression"
}
out <- list(
label = label
, objective = objective
, params = params
)
return(out)
}
},
process_predictions = function(pred, type) {
if (NROW(self$factor_levels)) {
if (type == "class") {
pred <- as.integer(pred) + 1L
attributes(pred)$levels <- self$factor_levels
attributes(pred)$class <- "factor"
} else if (type %in% c("response", "raw")) {
if (is.matrix(pred) && ncol(pred) == length(self$factor_levels)) {
colnames(pred) <- self$factor_levels
}
}
}
return(pred)
}
)
)
...@@ -103,6 +103,15 @@ NULL ...@@ -103,6 +103,15 @@ NULL
#' For a list of accepted objectives, see #' For a list of accepted objectives, see
#' \href{https://lightgbm.readthedocs.io/en/latest/Parameters.html#objective}{ #' \href{https://lightgbm.readthedocs.io/en/latest/Parameters.html#objective}{
#' the "objective" item of the "Parameters" section of the documentation}. #' the "objective" item of the "Parameters" section of the documentation}.
#'
#' If passing \code{"auto"} and \code{data} is not of type \code{lgb.Dataset}, the objective will
#' be determined according to what is passed for \code{label}:\itemize{
#' \item If passing a factor with two variables, will use objective \code{"binary"}.
#' \item If passing a factor with more than two variables, will use objective \code{"multiclass"}
#' (note that parameter \code{num_class} in this case will also be determined automatically from
#' \code{label}).
#' \item Otherwise, will use objective \code{"regression"}.
#' }
#' @param init_score initial score is the base prediction lightgbm will boost from #' @param init_score initial score is the base prediction lightgbm will boost from
#' @param num_threads Number of parallel threads to use. For best speed, this should be set to the number of #' @param num_threads Number of parallel threads to use. For best speed, this should be set to the number of
#' physical cores in the CPU - in a typical x86-64 machine, this corresponds to half the #' physical cores in the CPU - in a typical x86-64 machine, this corresponds to half the
...@@ -149,7 +158,7 @@ lightgbm <- function(data, ...@@ -149,7 +158,7 @@ lightgbm <- function(data,
init_model = NULL, init_model = NULL,
callbacks = list(), callbacks = list(),
serializable = TRUE, serializable = TRUE,
objective = "regression", objective = "auto",
init_score = NULL, init_score = NULL,
num_threads = NULL, num_threads = NULL,
...) { ...) {
...@@ -173,6 +182,22 @@ lightgbm <- function(data, ...@@ -173,6 +182,22 @@ lightgbm <- function(data,
, alternative_kwarg_value = verbose , alternative_kwarg_value = verbose
) )
# Process factors as labels and auto-determine objective
if (!lgb.is.Dataset(data)) {
data_processor <- DataProcessor$new()
temp <- data_processor$process_label(
label = label
, objective = objective
, params = params
)
label <- temp$label
objective <- temp$objective
params <- temp$params
rm(temp)
} else {
data_processor <- NULL
}
# Set data to a temporary variable # Set data to a temporary variable
dtrain <- data dtrain <- data
...@@ -204,6 +229,7 @@ lightgbm <- function(data, ...@@ -204,6 +229,7 @@ lightgbm <- function(data,
what = lgb.train what = lgb.train
, args = train_args , args = train_args
) )
bst$data_processor <- data_processor
return(bst) return(bst)
} }
......
...@@ -51,7 +51,12 @@ If <= 0, all iterations from start_iteration are used (no limits).} ...@@ -51,7 +51,12 @@ If <= 0, all iterations from start_iteration are used (no limits).}
} }
Note that, if using custom objectives, types "class" and "response" will not be available and will Note that, if using custom objectives, types "class" and "response" will not be available and will
default towards using "raw" instead.} default towards using "raw" instead.
If the model was fit through function \link{lightgbm} and it was passed a factor as labels,
passing the prediction type through \code{params} instead of through this argument might
result in factor levels for classification objectives not being applied correctly to the
resulting output.}
\item{params}{a list of additional named parameters. See \item{params}{a list of additional named parameters. See
\href{https://lightgbm.readthedocs.io/en/latest/Parameters.html#predict-parameters}{ \href{https://lightgbm.readthedocs.io/en/latest/Parameters.html#predict-parameters}{
......
...@@ -16,7 +16,7 @@ lightgbm( ...@@ -16,7 +16,7 @@ lightgbm(
init_model = NULL, init_model = NULL,
callbacks = list(), callbacks = list(),
serializable = TRUE, serializable = TRUE,
objective = "regression", objective = "auto",
init_score = NULL, init_score = NULL,
num_threads = NULL, num_threads = NULL,
... ...
...@@ -56,9 +56,18 @@ set to the iteration number of the best iteration.} ...@@ -56,9 +56,18 @@ set to the iteration number of the best iteration.}
\code{save} or \code{saveRDS} (see section "Model serialization").} \code{save} or \code{saveRDS} (see section "Model serialization").}
\item{objective}{Optimization objective (e.g. `"regression"`, `"binary"`, etc.). \item{objective}{Optimization objective (e.g. `"regression"`, `"binary"`, etc.).
For a list of accepted objectives, see For a list of accepted objectives, see
\href{https://lightgbm.readthedocs.io/en/latest/Parameters.html#objective}{ \href{https://lightgbm.readthedocs.io/en/latest/Parameters.html#objective}{
the "objective" item of the "Parameters" section of the documentation}.} the "objective" item of the "Parameters" section of the documentation}.
If passing \code{"auto"} and \code{data} is not of type \code{lgb.Dataset}, the objective will
be determined according to what is passed for \code{label}:\itemize{
\item If passing a factor with two variables, will use objective \code{"binary"}.
\item If passing a factor with more than two variables, will use objective \code{"multiclass"}
(note that parameter \code{num_class} in this case will also be determined automatically from
\code{label}).
\item Otherwise, will use objective \code{"regression"}.
}}
\item{init_score}{initial score is the base prediction lightgbm will boost from} \item{init_score}{initial score is the base prediction lightgbm will boost from}
......
...@@ -49,7 +49,12 @@ ...@@ -49,7 +49,12 @@
} }
Note that, if using custom objectives, types "class" and "response" will not be available and will Note that, if using custom objectives, types "class" and "response" will not be available and will
default towards using "raw" instead.} default towards using "raw" instead.
If the model was fit through function \link{lightgbm} and it was passed a factor as labels,
passing the prediction type through \code{params} instead of through this argument might
result in factor levels for classification objectives not being applied correctly to the
resulting output.}
\item{start_iteration}{int or None, optional (default=None) \item{start_iteration}{int or None, optional (default=None)
Start index of the iteration to predict. Start index of the iteration to predict.
...@@ -92,6 +97,12 @@ For prediction types that are meant to always return one output per observation ...@@ -92,6 +97,12 @@ For prediction types that are meant to always return one output per observation
Shapley base value. For multiclass objectives, this matrix will represent \code{num_classes} such matrices, Shapley base value. For multiclass objectives, this matrix will represent \code{num_classes} such matrices,
in the order "feature contributions for first class, feature contributions for second class, feature in the order "feature contributions for first class, feature contributions for second class, feature
contributions for third class, etc.". contributions for third class, etc.".
If the model was fit through function \link{lightgbm} and it was passed a factor as labels, predictions
returned from this function will retain the factor levels (either as values for \code{type="class"}, or
as column names for \code{type="response"} and \code{type="raw"} for multi-class objectives). Note that
passing the requested prediction type under \code{params} instead of through \code{type} might result in
the factor levels not being present in the output.
} }
\description{ \description{
Predicted values based on class \code{lgb.Booster} Predicted values based on class \code{lgb.Booster}
......
...@@ -3529,3 +3529,120 @@ test_that("lgb.cv() only prints eval metrics when expected to", { ...@@ -3529,3 +3529,120 @@ test_that("lgb.cv() only prints eval metrics when expected to", {
fitted_model = out[["booster"]] fitted_model = out[["booster"]]
) )
}) })
test_that("lightgbm() changes objective='auto' appropriately", {
# Regression
data("mtcars")
y <- mtcars$mpg
x <- as.matrix(mtcars[, -1L])
model <- lightgbm(x, y, objective = "auto", verbose = VERBOSITY, nrounds = 5L)
expect_equal(model$params$objective, "regression")
model_txt_lines <- strsplit(
x = model$save_model_to_string()
, split = "\n"
, fixed = TRUE
)[[1L]]
expect_true(any(grepl("objective=regression", model_txt_lines, fixed = TRUE)))
expect_false(any(grepl("objective=regression_l1", model_txt_lines, fixed = TRUE)))
# Binary classification
x <- train$data
y <- factor(train$label)
model <- lightgbm(x, y, objective = "auto", verbose = VERBOSITY, nrounds = 5L)
expect_equal(model$params$objective, "binary")
model_txt_lines <- strsplit(
x = model$save_model_to_string()
, split = "\n"
, fixed = TRUE
)[[1L]]
expect_true(any(grepl("objective=binary", model_txt_lines, fixed = TRUE)))
# Multi-class classification
data("iris")
y <- factor(iris$Species)
x <- as.matrix(iris[, -5L])
model <- lightgbm(x, y, objective = "auto", verbose = VERBOSITY, nrounds = 5L)
expect_equal(model$params$objective, "multiclass")
expect_equal(model$params$num_class, 3L)
model_txt_lines <- strsplit(
x = model$save_model_to_string()
, split = "\n"
, fixed = TRUE
)[[1L]]
expect_true(any(grepl("objective=multiclass", model_txt_lines, fixed = TRUE)))
})
test_that("lightgbm() determines number of classes for non-default multiclass objectives", {
data("iris")
y <- factor(iris$Species)
x <- as.matrix(iris[, -5L])
model <- lightgbm(x, y, objective = "multiclassova", verbose = VERBOSITY, nrounds = 5L)
expect_equal(model$params$objective, "multiclassova")
expect_equal(model$params$num_class, 3L)
model_txt_lines <- strsplit(
x = model$save_model_to_string()
, split = "\n"
, fixed = TRUE
)[[1L]]
expect_true(any(grepl("objective=multiclassova", model_txt_lines, fixed = TRUE)))
})
test_that("lightgbm() doesn't accept binary classification with non-binary factors", {
data("iris")
y <- factor(iris$Species)
x <- as.matrix(iris[, -5L])
expect_error({
lightgbm(x, y, objective = "binary", verbose = VERBOSITY, nrounds = 5L)
}, regexp = "Factors with >2 levels as labels only allowed for multi-class objectives")
})
test_that("lightgbm() doesn't accept multi-class classification with binary factors", {
data("iris")
y <- as.character(iris$Species)
y[y == "setosa"] <- "versicolor"
y <- factor(y)
x <- as.matrix(iris[, -5L])
expect_error({
lightgbm(x, y, objective = "multiclass", verbose = VERBOSITY, nrounds = 5L)
}, regexp = "Two-level factors as labels only allowed for objective='binary'")
})
test_that("lightgbm() model predictions retain factor levels for multiclass classification", {
data("iris")
y <- factor(iris$Species)
x <- as.matrix(iris[, -5L])
model <- lightgbm(x, y, objective = "auto", verbose = VERBOSITY, nrounds = 5L)
pred <- predict(model, x, type = "class")
expect_true(is.factor(pred))
expect_equal(levels(pred), levels(y))
pred <- predict(model, x, type = "response")
expect_equal(colnames(pred), levels(y))
pred <- predict(model, x, type = "raw")
expect_equal(colnames(pred), levels(y))
})
test_that("lightgbm() model predictions retain factor levels for binary classification", {
data("iris")
y <- as.character(iris$Species)
y[y == "setosa"] <- "versicolor"
y <- factor(y)
x <- as.matrix(iris[, -5L])
model <- lightgbm(x, y, objective = "auto", verbose = VERBOSITY, nrounds = 5L)
pred <- predict(model, x, type = "class")
expect_true(is.factor(pred))
expect_equal(levels(pred), levels(y))
pred <- predict(model, x, type = "response")
expect_true(is.vector(pred))
expect_true(is.numeric(pred))
expect_false(any(pred %in% y))
pred <- predict(model, x, type = "raw")
expect_true(is.vector(pred))
expect_true(is.numeric(pred))
expect_false(any(pred %in% y))
})
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment