Why do the training curve fall sharp suddenly?
1 visualización (últimos 30 días)
Mostrar comentarios más antiguos
Saugata Bose
el 31 de Ag. de 2019
Comentada: Saugata Bose
el 31 de Ag. de 2019
I am training a CNN classifier on a binary balanced dataset. The dataset has 4500 numbers of tweet data along with the class of the tweet. During training, I am applying, GLOVE embedding of 300 dimensions, 'adam' solver to run the model for 33 times of epochs. Besides, the sequence length I have considered is 31.
I have applied 200 filters which include a number of convolution2dlayers,batch normalization layers, relu layers, dropout layers and max-pooling layers. The drop out I have considered is 0.2 and the max pool layer is of size [1 sequence length].
The training curve is approaching smoothly until the end period where it has fallen sharply. Here, I have attached the training plot I receive:
Would you please explain to me why does this sudden fall occur? And how could I get rid of this?
Thanks,
0 comentarios
Respuesta aceptada
Matt J
el 31 de Ag. de 2019
The final validation metrics are labeled Final in the plots. If your network contains batch normalization layers, then the final validation metrics are often different from the validation metrics evaluated during training. This is because batch normalization layers in the final network perform different operations than during training.
Más respuestas (0)
Ver también
Categorías
Más información sobre Sequence and Numeric Feature Data Workflows en Help Center y File Exchange.
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!