Hi There,
I got some results from using nnet on a two-class problem, and I'd like
to hear your comments to understand well about the algorithm. In the training
set, the ratio of class 1 to class 2 is about 23:77. I did a 5-fold cross
validation. The networks were trained twice, one with 'weights=1', one
with 'weights=ifelse(species=="class1", 77/33, 1)'(pointed out
by Prof. Brian Ripley).All other settings are same. The average Matthew
Correlation Coeffience for the one with weights=1 is 0.80, significantly larger
than that of the other, 0.74. So, it seems weighting the unbalanced samples does
not help performance on evaluations, which is against my initial thoughts. My
question would be, does that mean the training data is not unbalanced enough?
then how unbalanced is enough? Or it was totally just a signal event? Or it was
just some suboptimal results? Any references regarding this issue in particular?
Thanks!
Best regards,
Baoqiang Cao