Alternative to softmax function for Neural Network predicting fractions of a whole
3 visualizaciones (últimos 30 días)
Mostrar comentarios más antiguos
Hi, I created a feed forward Regression Neural Network to predict variables which are fractions of a whole (i.e. they sum up to 1). In order to have the network fullfil this criterion perfectly, I am using the softmax transfer function. Unfortunately, I realize that the network predicts smaller fractions very poorly, and I think this is due to the fact that the softmax transfer function normalizes my target fractions by dividing exponent of the fractions minus the largest fraction by its sum (exp(n-nmax)/sum(exp(n-nmax))), which results in much larger values for very small fractions. It wouldn't have to do that, since my fractions are already between 0 and 1. Can I change that somehow in the softmax transfer function, or is there an alternative to it that doesn't do this normailzation?
6 comentarios
Ajay Pattassery
el 29 de Ag. de 2019
Did you tried creating a custom layer which can force your output to one like the one I mentioned above instead of having a softmax layer.
Please refer the following link for creating custom layers.
Respuestas (0)
Ver también
Categorías
Más información sobre Classification en Help Center y File Exchange.
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!