Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
tianlh
LightGBM-DCU
Commits
f65164f6
Commit
f65164f6
authored
Nov 30, 2016
by
Guolin Ke
Browse files
less verbose in test
parent
c67d2890
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
7 additions
and
7 deletions
+7
-7
tests/python_package_test/test_sklearn.py
tests/python_package_test/test_sklearn.py
+7
-7
No files found.
tests/python_package_test/test_sklearn.py
View file @
f65164f6
...
...
@@ -11,13 +11,13 @@ def test_binary_classification():
X
,
y
=
datasets
.
make_classification
(
n_samples
=
10000
,
n_features
=
100
)
x_train
,
x_test
,
y_train
,
y_test
=
model_selection
.
train_test_split
(
X
,
y
,
test_size
=
0.1
)
lgb_model
=
lgb
.
LGBMClassifier
().
fit
(
x_train
,
y_train
,
eval_set
=
[[
x_train
,
y_train
],(
x_test
,
y_test
)],
eval_metric
=
'binary_logloss'
)
lgb_model
=
lgb
.
LGBMClassifier
().
fit
(
x_train
,
y_train
)
from
sklearn.datasets
import
load_digits
digits
=
load_digits
(
2
)
y
=
digits
[
'target'
]
X
=
digits
[
'data'
]
x_train
,
x_test
,
y_train
,
y_test
=
model_selection
.
train_test_split
(
X
,
y
,
test_size
=
0.2
)
lgb_model
=
lgb
.
LGBMClassifier
().
fit
(
x_train
,
y_train
,
eval_set
=
[[
x_train
,
y_train
],(
x_test
,
y_test
)],
eval_metric
=
'binary_logloss'
)
lgb_model
=
lgb
.
LGBMClassifier
().
fit
(
x_train
,
y_train
)
preds
=
lgb_model
.
predict
(
x_test
)
err
=
sum
(
1
for
i
in
range
(
len
(
preds
))
if
int
(
preds
[
i
]
>
0.5
)
!=
y_test
[
i
])
/
float
(
len
(
preds
))
...
...
@@ -37,7 +37,7 @@ def test_multiclass_classification():
x_train
,
x_test
,
y_train
,
y_test
=
model_selection
.
train_test_split
(
X
,
y
,
test_size
=
0.1
)
lgb_model
=
lgb
.
LGBMClassifier
().
fit
(
x_train
,
y_train
,
eval_set
=
[[
x_train
,
y_train
],(
x_test
,
y_test
)],
eval_metric
=
'multi_logloss'
)
lgb_model
=
lgb
.
LGBMClassifier
().
fit
(
x_train
,
y_train
)
preds
=
lgb_model
.
predict
(
x_test
)
check_pred
(
preds
,
y_test
)
...
...
@@ -52,7 +52,7 @@ def test_regression():
y
=
boston
[
'target'
]
X
=
boston
[
'data'
]
x_train
,
x_test
,
y_train
,
y_test
=
model_selection
.
train_test_split
(
X
,
y
,
test_size
=
0.1
)
lgb_model
=
lgb
.
LGBMRegressor
().
fit
(
x_train
,
y_train
,
eval_set
=
[[
x_train
,
y_train
],(
x_test
,
y_test
)],
eval_metric
=
'l2'
)
lgb_model
=
lgb
.
LGBMRegressor
().
fit
(
x_train
,
y_train
)
preds
=
lgb_model
.
predict
(
x_test
)
assert
mean_squared_error
(
preds
,
y_test
)
<
40
...
...
@@ -69,7 +69,7 @@ def test_regression_with_custom_objective():
y
=
boston
[
'target'
]
X
=
boston
[
'data'
]
x_train
,
x_test
,
y_train
,
y_test
=
model_selection
.
train_test_split
(
X
,
y
,
test_size
=
0.1
)
lgb_model
=
lgb
.
LGBMRegressor
(
objective
=
objective_ls
).
fit
(
x_train
,
y_train
,
eval_set
=
[[
x_train
,
y_train
],(
x_test
,
y_test
)],
eval_metric
=
'l2'
)
lgb_model
=
lgb
.
LGBMRegressor
(
objective
=
objective_ls
).
fit
(
x_train
,
y_train
)
preds
=
lgb_model
.
predict
(
x_test
)
assert
mean_squared_error
(
preds
,
y_test
)
<
40
...
...
@@ -84,13 +84,13 @@ def test_binary_classification_with_custom_objective():
return
grad
,
hess
X
,
y
=
datasets
.
make_classification
(
n_samples
=
10000
,
n_features
=
100
)
x_train
,
x_test
,
y_train
,
y_test
=
model_selection
.
train_test_split
(
X
,
y
,
test_size
=
0.1
)
lgb_model
=
lgb
.
LGBMClassifier
(
objective
=
logregobj
).
fit
(
x_train
,
y_train
,
eval_set
=
[[
x_train
,
y_train
],(
x_test
,
y_test
)],
eval_metric
=
'binary_logloss'
)
lgb_model
=
lgb
.
LGBMClassifier
(
objective
=
logregobj
).
fit
(
x_train
,
y_train
)
from
sklearn.datasets
import
load_digits
digits
=
load_digits
(
2
)
y
=
digits
[
'target'
]
X
=
digits
[
'data'
]
x_train
,
x_test
,
y_train
,
y_test
=
model_selection
.
train_test_split
(
X
,
y
,
test_size
=
0.2
)
lgb_model
=
lgb
.
LGBMClassifier
(
objective
=
logregobj
).
fit
(
x_train
,
y_train
,
eval_set
=
[[
x_train
,
y_train
],(
x_test
,
y_test
)],
eval_metric
=
'binary_logloss'
)
lgb_model
=
lgb
.
LGBMClassifier
(
objective
=
logregobj
).
fit
(
x_train
,
y_train
)
preds
=
lgb_model
.
predict
(
x_test
)
err
=
sum
(
1
for
i
in
range
(
len
(
preds
))
if
int
(
preds
[
i
]
>
0.5
)
!=
y_test
[
i
])
/
float
(
len
(
preds
))
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment