What happend if the validation performance was greater than the test performance?
3 visualizaciones (últimos 30 días)
Mostrar comentarios más antiguos
Rita
el 22 de Feb. de 2017
Comentada: Rita
el 3 de Mzo. de 2017
I have trained several networks with different hidden neurons. Now my question is what happens if I select a network with the validation performance higher than the test performance.I have read somewhere that in this situation the distribution of data between the training set and test set is not correct.Thanks for any suggestion.
0 comentarios
Respuesta aceptada
Greg Heath
el 28 de Feb. de 2017
There is no rule governing the order of the val and tst performances. That is why it is worthwhile to design a number of nets differing only by the random initial weights and/or if not a time series, the order of the input/target pairs.
Thank you for formally accepting my answer
Greg
3 comentarios
Greg Heath
el 1 de Mzo. de 2017
Sometimes it is dangerous to rely on the results of just a few designs. That is why, for challenging cases, I typically design at least 10 or 15 nets for each candidate value of H, the number of hidden nodes.
Then by tabulating the 3 NMSE values with the column for NMSEval monotonically decreasing, you can detect how well trends in NMSEval can indicate the acceptability of NMSEtst.
Of course there are other tabulations and/or plots that reveal design trends( e.g., NMSE vs H). The important thing is to design enough nets so that you are convinced your choice is reliable.
Más respuestas (0)
Ver también
Categorías
Más información sobre Deep Learning Toolbox en Help Center y File Exchange.
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!