How to use different transfer functions within the same layer of a neural network?
2 visualizaciones (últimos 30 días)
Mostrar comentarios más antiguos
For example, hidden layer with 3 different neurons: 1.tansig, 2.logsig, 3.elliotsig. I understand that this influences normalization range and that training might not resault in good performance but i need it to test some other aspects. Thank you.
0 comentarios
Respuestas (1)
Greg Heath
el 24 de Abr. de 2013
If it were possible with the NNTBX, you would have to design a custom net.
The only way I can see doing it is to have 3 hidden layers that are only connected to the input and output but not to each other.
See the custom network section in the documentation.
Hope this helps.
Thank you for formally accepting my answer
Greg
Ver también
Categorías
Más información sobre Sequence and Numeric Feature Data Workflows en Help Center y File Exchange.
Productos
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!