In the function to calculate ZILN loss,
The classification loss and regression loss is combined together, however, the scale of two loss is different meaning, the binary crossentropy loss is often much smaller than regression loss. Shouldn't we compensate for this difference in scale?
In addition, why do we take the negative of regression loss?
In the function to calculate ZILN loss,
The classification loss and regression loss is combined together, however, the scale of two loss is different meaning, the binary crossentropy loss is often much smaller than regression loss. Shouldn't we compensate for this difference in scale?
In addition, why do we take the negative of regression loss?