Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
tianlh
LightGBM-DCU
Commits
05f4a26a
Commit
05f4a26a
authored
Apr 17, 2017
by
Guolin Ke
Browse files
revert the multi-class loss.
parent
66b7f032
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
6 additions
and
25 deletions
+6
-25
src/objective/multiclass_objective.hpp
src/objective/multiclass_objective.hpp
+6
-25
No files found.
src/objective/multiclass_objective.hpp
View file @
05f4a26a
...
@@ -17,7 +17,6 @@ class MulticlassSoftmax: public ObjectiveFunction {
...
@@ -17,7 +17,6 @@ class MulticlassSoftmax: public ObjectiveFunction {
public:
public:
explicit
MulticlassSoftmax
(
const
ObjectiveConfig
&
config
)
{
explicit
MulticlassSoftmax
(
const
ObjectiveConfig
&
config
)
{
num_class_
=
config
.
num_class
;
num_class_
=
config
.
num_class
;
softmax_weight_decay_
=
1e-3
;
}
}
explicit
MulticlassSoftmax
(
const
std
::
vector
<
std
::
string
>&
strs
)
{
explicit
MulticlassSoftmax
(
const
std
::
vector
<
std
::
string
>&
strs
)
{
...
@@ -44,25 +43,12 @@ public:
...
@@ -44,25 +43,12 @@ public:
label_
=
metadata
.
label
();
label_
=
metadata
.
label
();
weights_
=
metadata
.
weights
();
weights_
=
metadata
.
weights
();
label_int_
.
resize
(
num_data_
);
label_int_
.
resize
(
num_data_
);
std
::
vector
<
data_size_t
>
cnt_per_class
(
num_class_
,
0
);
for
(
int
i
=
0
;
i
<
num_data_
;
++
i
)
{
for
(
int
i
=
0
;
i
<
num_data_
;
++
i
)
{
label_int_
[
i
]
=
static_cast
<
int
>
(
label_
[
i
]);
label_int_
[
i
]
=
static_cast
<
int
>
(
label_
[
i
]);
if
(
label_int_
[
i
]
<
0
||
label_int_
[
i
]
>=
num_class_
)
{
if
(
label_int_
[
i
]
<
0
||
label_int_
[
i
]
>=
num_class_
)
{
Log
::
Fatal
(
"Label must be in [0, %d), but found %d in label"
,
num_class_
,
label_int_
[
i
]);
Log
::
Fatal
(
"Label must be in [0, %d), but found %d in label"
,
num_class_
,
label_int_
[
i
]);
}
}
++
cnt_per_class
[
label_int_
[
i
]];
}
}
int
non_empty_class
=
0
;
is_empty_class_
=
std
::
vector
<
bool
>
(
num_class_
,
false
);
for
(
int
i
=
0
;
i
<
num_class_
;
++
i
)
{
if
(
cnt_per_class
[
i
]
>
0
)
{
++
non_empty_class
;
}
else
{
is_empty_class_
[
i
]
=
true
;
}
}
if
(
non_empty_class
<
2
)
{
non_empty_class
=
2
;
}
hessian_nor_
=
static_cast
<
score_t
>
(
non_empty_class
)
/
(
non_empty_class
-
1
);
}
}
void
GetGradients
(
const
double
*
score
,
score_t
*
gradients
,
score_t
*
hessians
)
const
override
{
void
GetGradients
(
const
double
*
score
,
score_t
*
gradients
,
score_t
*
hessians
)
const
override
{
...
@@ -77,15 +63,14 @@ public:
...
@@ -77,15 +63,14 @@ public:
}
}
Common
::
Softmax
(
&
rec
);
Common
::
Softmax
(
&
rec
);
for
(
int
k
=
0
;
k
<
num_class_
;
++
k
)
{
for
(
int
k
=
0
;
k
<
num_class_
;
++
k
)
{
if
(
is_empty_class_
[
k
])
{
continue
;
}
auto
p
=
rec
[
k
];
auto
p
=
rec
[
k
];
size_t
idx
=
static_cast
<
size_t
>
(
num_data_
)
*
k
+
i
;
size_t
idx
=
static_cast
<
size_t
>
(
num_data_
)
*
k
+
i
;
if
(
label_int_
[
i
]
==
k
)
{
if
(
label_int_
[
i
]
==
k
)
{
gradients
[
idx
]
=
static_cast
<
score_t
>
(
p
-
1.0
f
+
softmax_weight_decay_
*
score
[
idx
]
);
gradients
[
idx
]
=
static_cast
<
score_t
>
(
p
-
1.0
f
);
}
else
{
}
else
{
gradients
[
idx
]
=
static_cast
<
score_t
>
(
p
+
softmax_weight_decay_
*
score
[
idx
]
);
gradients
[
idx
]
=
static_cast
<
score_t
>
(
p
);
}
}
hessians
[
idx
]
=
static_cast
<
score_t
>
(
hessian_nor_
*
p
*
(
1.0
f
-
p
)
+
softmax_weight_decay_
);
hessians
[
idx
]
=
static_cast
<
score_t
>
(
2.0
f
*
p
*
(
1.0
f
-
p
));
}
}
}
}
}
else
{
}
else
{
...
@@ -99,15 +84,14 @@ public:
...
@@ -99,15 +84,14 @@ public:
}
}
Common
::
Softmax
(
&
rec
);
Common
::
Softmax
(
&
rec
);
for
(
int
k
=
0
;
k
<
num_class_
;
++
k
)
{
for
(
int
k
=
0
;
k
<
num_class_
;
++
k
)
{
if
(
is_empty_class_
[
k
])
{
continue
;
}
auto
p
=
rec
[
k
];
auto
p
=
rec
[
k
];
size_t
idx
=
static_cast
<
size_t
>
(
num_data_
)
*
k
+
i
;
size_t
idx
=
static_cast
<
size_t
>
(
num_data_
)
*
k
+
i
;
if
(
label_int_
[
i
]
==
k
)
{
if
(
label_int_
[
i
]
==
k
)
{
gradients
[
idx
]
=
static_cast
<
score_t
>
((
p
-
1.0
f
+
softmax_weight_decay_
*
score
[
idx
]
)
*
weights_
[
i
]);
gradients
[
idx
]
=
static_cast
<
score_t
>
((
p
-
1.0
f
)
*
weights_
[
i
]);
}
else
{
}
else
{
gradients
[
idx
]
=
static_cast
<
score_t
>
((
p
+
softmax_weight_decay_
*
score
[
idx
]
)
*
weights_
[
i
]);
gradients
[
idx
]
=
static_cast
<
score_t
>
((
p
)
*
weights_
[
i
]);
}
}
hessians
[
idx
]
=
static_cast
<
score_t
>
((
hessian_nor_
*
p
*
(
1.0
f
-
p
)
+
softmax_weight_decay_
)
*
weights_
[
i
]);
hessians
[
idx
]
=
static_cast
<
score_t
>
((
2.0
f
*
p
*
(
1.0
f
-
p
))
*
weights_
[
i
]);
}
}
}
}
}
}
...
@@ -145,9 +129,6 @@ private:
...
@@ -145,9 +129,6 @@ private:
std
::
vector
<
int
>
label_int_
;
std
::
vector
<
int
>
label_int_
;
/*! \brief Weights for data */
/*! \brief Weights for data */
const
float
*
weights_
;
const
float
*
weights_
;
std
::
vector
<
bool
>
is_empty_class_
;
double
softmax_weight_decay_
;
score_t
hessian_nor_
;
};
};
/*!
/*!
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment