I'm doing research for cost-sensitive neural network based on Tensorflow. But because of the static graph structure of Tensorflow. Some NN structure couldn't be realized by myself.
My loss function(cost) ,cost matrix and the computational progress is described as follow and my target is to compute the total cost and then optimize the NN :
Approximately computational progress:
- the
y_
is the last full-connect output of a CNN which has shape(1024,5)
- the
y
is a Tensor which has shape(1024) and indicates the ground truth ofx[i]
- the
y_soft[i] [j]
indicates the probability ofx[i]
to be classj
How can I realize this in Tensorflow?
cost_matrix:
label:
y*:
y(prediction):
label,cost_matrix-->cost_embedding:
It obvious 0.3 in [0.2,0.3,0.5] refers to right lable probility of [0,1,0], so it should not contibute to loss.
0.7 in [0.1,0.2,0.7] is the same. In other words, the pos with value 1 in y* not contibute to loss.
So I have (1-y*):
Then the entropy is target*log(predict) + (1-target) * log(1-predict),and value 0 in y*,should use (1-target)*log(1-predict), so I use (1-predict) said (1-y)
1-y:
(italic num is useless)
the custom loss is
and you can see the (1-y*) can be drop here
so the loss is -tf.reduce_mean(cost_embedding*log(1-y)) ,to make it applicable , should be:
the demo is below