Validation loss increases and validation accuracy decreasesHigh model accuracy vs very low validation accuarcyWhat are the possible approaches to fixing Overfitting on a CNN?Can overfitting occur even with validation loss still dropping?Accuracy drops if more layers trainable - weirdHow to set input for proper fit with lstm?Multi-label classification, recall and precision increase but accuracy decrease, why?Training Accuracy stuck in KerasTraining multi-label classifier with unbalanced samples in Kerashow to consider some miss classifications “half correct” using as base categorical_crossentropy - for a trading systemkeras model only predicts one class for all the test imagesAccuracy and Loss in MLPnormalization in auto-encoder

Why "had" in "[something] we would have made had we used [something]"?

Can I still be respawned if I die by falling off the map?

Temporarily disable WLAN internet access for children, but allow it for adults

Calculating total slots

Why is this estimator biased?

Quoting Keynes in a lecture

How do you respond to a colleague from another team when they're wrongly expecting that you'll help them?

Limits and Infinite Integration by Parts

What does "Scientists rise up against statistical significance" mean? (Comment in Nature)

Can the US President recognize Israel’s sovereignty over the Golan Heights for the USA or does that need an act of Congress?

Why is the "ls" command showing permissions of files in a FAT32 partition?

Pre-mixing cryogenic fuels and using only one fuel tank

Why does AES have exactly 10 rounds for a 128-bit key, 12 for 192 bits and 14 for a 256-bit key size?

Lowest total scrabble score

Plot of a tornado-shaped surface

Strong empirical falsification of quantum mechanics based on vacuum energy density

Fear of getting stuck on one programming language / technology that is not used in my country

How to fade a semiplane defined by line?

Does Doodling or Improvising on the Piano Have Any Benefits?

Can I say "fingers" when referring to toes?

On a tidally locked planet, would time be quantized?

Extract more than nine arguments that occur periodically in a sentence to use in macros in order to typset

Does IPv6 have similar concept of network mask?

Has any country ever had 2 former presidents in jail simultaneously?



Validation loss increases and validation accuracy decreases


High model accuracy vs very low validation accuarcyWhat are the possible approaches to fixing Overfitting on a CNN?Can overfitting occur even with validation loss still dropping?Accuracy drops if more layers trainable - weirdHow to set input for proper fit with lstm?Multi-label classification, recall and precision increase but accuracy decrease, why?Training Accuracy stuck in KerasTraining multi-label classifier with unbalanced samples in Kerashow to consider some miss classifications “half correct” using as base categorical_crossentropy - for a trading systemkeras model only predicts one class for all the test imagesAccuracy and Loss in MLPnormalization in auto-encoder













2












$begingroup$


I have an issue with my model. I'm trying to use the most basic Conv1D model to analyze review data and output a rating of 1-5 class, therefore the loss is categorical_crossentropy. Model structure is as below



# define model
model = Sequential()
model.add(Embedding(vocab_size, 100, input_length=max_length))
model.add(Conv1D(filters=32, kernel_size=8, activation='relu'))
model.add(MaxPooling1D(pool_size=2))
model.add(Flatten())
model.add(Dense(10, activation='relu'))
model.add(Dense(5, activation='softmax'))

# compile network
model.compile(loss='categorical_crossentropy', optimizer='adam', metrics=['accuracy'])

# fit network
model.fit(final_X_train, final_Y_train, epochs=5, batch_size=50, validation_data=(final_X_val, final_Y_val), callbacks=callback)


Total params: 14,977,717



Trainable params: 14,977,717



Non-trainable params: 0



Train on 212135 samples, validate on 69472 samples



Epoch 1/5
loss: 0.9452 - acc: 0.5781 - val_loss: 0.8440 - val_acc: 0.6309



Epoch 2/5
loss: 0.7707 - acc: 0.6711 - val_loss: 0.8234 - val_acc: 0.6433



Epoch 3/5
loss: 0.5807 - acc: 0.7657 - val_loss: 0.9144 - val_acc: 0.6345



Epoch 4/5
loss: 0.3736 - acc: 0.8575 - val_loss: 1.1982 - val_acc: 0.6194



Epoch 5/5
loss: 0.2285 - acc: 0.9173 - val_loss: 1.5770 - val_acc: 0.6073



Training acc increases and loss decreases as expected. But validation loss and validation acc decrease straight after the 2nd epoch itself. The overall testing after training gives an accuracy around 60s.



The total accuracy is : 0.6046845041714888




I've already cleaned, shuffled, down-sampled (all classes have 42427 number of data samples) and split the data properly to training(70%) / validation(10%) / testing(20%).




If you see any improvements to fix this problem, please let me know. :)










share|improve this question







New contributor




stranger is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
Check out our Code of Conduct.







$endgroup$
















    2












    $begingroup$


    I have an issue with my model. I'm trying to use the most basic Conv1D model to analyze review data and output a rating of 1-5 class, therefore the loss is categorical_crossentropy. Model structure is as below



    # define model
    model = Sequential()
    model.add(Embedding(vocab_size, 100, input_length=max_length))
    model.add(Conv1D(filters=32, kernel_size=8, activation='relu'))
    model.add(MaxPooling1D(pool_size=2))
    model.add(Flatten())
    model.add(Dense(10, activation='relu'))
    model.add(Dense(5, activation='softmax'))

    # compile network
    model.compile(loss='categorical_crossentropy', optimizer='adam', metrics=['accuracy'])

    # fit network
    model.fit(final_X_train, final_Y_train, epochs=5, batch_size=50, validation_data=(final_X_val, final_Y_val), callbacks=callback)


    Total params: 14,977,717



    Trainable params: 14,977,717



    Non-trainable params: 0



    Train on 212135 samples, validate on 69472 samples



    Epoch 1/5
    loss: 0.9452 - acc: 0.5781 - val_loss: 0.8440 - val_acc: 0.6309



    Epoch 2/5
    loss: 0.7707 - acc: 0.6711 - val_loss: 0.8234 - val_acc: 0.6433



    Epoch 3/5
    loss: 0.5807 - acc: 0.7657 - val_loss: 0.9144 - val_acc: 0.6345



    Epoch 4/5
    loss: 0.3736 - acc: 0.8575 - val_loss: 1.1982 - val_acc: 0.6194



    Epoch 5/5
    loss: 0.2285 - acc: 0.9173 - val_loss: 1.5770 - val_acc: 0.6073



    Training acc increases and loss decreases as expected. But validation loss and validation acc decrease straight after the 2nd epoch itself. The overall testing after training gives an accuracy around 60s.



    The total accuracy is : 0.6046845041714888




    I've already cleaned, shuffled, down-sampled (all classes have 42427 number of data samples) and split the data properly to training(70%) / validation(10%) / testing(20%).




    If you see any improvements to fix this problem, please let me know. :)










    share|improve this question







    New contributor




    stranger is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
    Check out our Code of Conduct.







    $endgroup$














      2












      2








      2





      $begingroup$


      I have an issue with my model. I'm trying to use the most basic Conv1D model to analyze review data and output a rating of 1-5 class, therefore the loss is categorical_crossentropy. Model structure is as below



      # define model
      model = Sequential()
      model.add(Embedding(vocab_size, 100, input_length=max_length))
      model.add(Conv1D(filters=32, kernel_size=8, activation='relu'))
      model.add(MaxPooling1D(pool_size=2))
      model.add(Flatten())
      model.add(Dense(10, activation='relu'))
      model.add(Dense(5, activation='softmax'))

      # compile network
      model.compile(loss='categorical_crossentropy', optimizer='adam', metrics=['accuracy'])

      # fit network
      model.fit(final_X_train, final_Y_train, epochs=5, batch_size=50, validation_data=(final_X_val, final_Y_val), callbacks=callback)


      Total params: 14,977,717



      Trainable params: 14,977,717



      Non-trainable params: 0



      Train on 212135 samples, validate on 69472 samples



      Epoch 1/5
      loss: 0.9452 - acc: 0.5781 - val_loss: 0.8440 - val_acc: 0.6309



      Epoch 2/5
      loss: 0.7707 - acc: 0.6711 - val_loss: 0.8234 - val_acc: 0.6433



      Epoch 3/5
      loss: 0.5807 - acc: 0.7657 - val_loss: 0.9144 - val_acc: 0.6345



      Epoch 4/5
      loss: 0.3736 - acc: 0.8575 - val_loss: 1.1982 - val_acc: 0.6194



      Epoch 5/5
      loss: 0.2285 - acc: 0.9173 - val_loss: 1.5770 - val_acc: 0.6073



      Training acc increases and loss decreases as expected. But validation loss and validation acc decrease straight after the 2nd epoch itself. The overall testing after training gives an accuracy around 60s.



      The total accuracy is : 0.6046845041714888




      I've already cleaned, shuffled, down-sampled (all classes have 42427 number of data samples) and split the data properly to training(70%) / validation(10%) / testing(20%).




      If you see any improvements to fix this problem, please let me know. :)










      share|improve this question







      New contributor




      stranger is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.







      $endgroup$




      I have an issue with my model. I'm trying to use the most basic Conv1D model to analyze review data and output a rating of 1-5 class, therefore the loss is categorical_crossentropy. Model structure is as below



      # define model
      model = Sequential()
      model.add(Embedding(vocab_size, 100, input_length=max_length))
      model.add(Conv1D(filters=32, kernel_size=8, activation='relu'))
      model.add(MaxPooling1D(pool_size=2))
      model.add(Flatten())
      model.add(Dense(10, activation='relu'))
      model.add(Dense(5, activation='softmax'))

      # compile network
      model.compile(loss='categorical_crossentropy', optimizer='adam', metrics=['accuracy'])

      # fit network
      model.fit(final_X_train, final_Y_train, epochs=5, batch_size=50, validation_data=(final_X_val, final_Y_val), callbacks=callback)


      Total params: 14,977,717



      Trainable params: 14,977,717



      Non-trainable params: 0



      Train on 212135 samples, validate on 69472 samples



      Epoch 1/5
      loss: 0.9452 - acc: 0.5781 - val_loss: 0.8440 - val_acc: 0.6309



      Epoch 2/5
      loss: 0.7707 - acc: 0.6711 - val_loss: 0.8234 - val_acc: 0.6433



      Epoch 3/5
      loss: 0.5807 - acc: 0.7657 - val_loss: 0.9144 - val_acc: 0.6345



      Epoch 4/5
      loss: 0.3736 - acc: 0.8575 - val_loss: 1.1982 - val_acc: 0.6194



      Epoch 5/5
      loss: 0.2285 - acc: 0.9173 - val_loss: 1.5770 - val_acc: 0.6073



      Training acc increases and loss decreases as expected. But validation loss and validation acc decrease straight after the 2nd epoch itself. The overall testing after training gives an accuracy around 60s.



      The total accuracy is : 0.6046845041714888




      I've already cleaned, shuffled, down-sampled (all classes have 42427 number of data samples) and split the data properly to training(70%) / validation(10%) / testing(20%).




      If you see any improvements to fix this problem, please let me know. :)







      python deep-learning classification keras overfitting






      share|improve this question







      New contributor




      stranger is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.











      share|improve this question







      New contributor




      stranger is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.









      share|improve this question




      share|improve this question






      New contributor




      stranger is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.









      asked yesterday









      strangerstranger

      112




      112




      New contributor




      stranger is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.





      New contributor





      stranger is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.






      stranger is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.




















          1 Answer
          1






          active

          oldest

          votes


















          0












          $begingroup$

          What you are experiencing is known as overfitting, and it’s a common problem in machine learning and data science.



          Overfitting happens when a model begins to focus on the noise in the training dataset and extracts features based on it. This helps the model to improve its performance on the training set, but hurts its ability to generalize, so the accuracy of the validation set decreases.



          To deal with overfitting, you need to use regularization during the training. You ca try adding dropout layers or batch-normalization layers, adding weight regularization, or you can artificially increase the size of your training set by performing some data augmentation.



          You can read more about it in the following post:



          What are the possible approaches to fixing Overfitting on a CNN?






          share|improve this answer









          $endgroup$












          • $begingroup$
            Thanks, I tried adding regularizers to Conv1D and Dense layers as below. Conv1D(filters=64, kernel_size=8, activation='relu', kernel_regularizer=l2(0.001), bias_regularizer=l2(0.001)) AND Dense(10, activation='relu', kernel_regularizer=l2(0.001)) . But unfortunately none of them did any changes to val_acc and val_loss.
            $endgroup$
            – stranger
            yesterday










          Your Answer





          StackExchange.ifUsing("editor", function ()
          return StackExchange.using("mathjaxEditing", function ()
          StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix)
          StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
          );
          );
          , "mathjax-editing");

          StackExchange.ready(function()
          var channelOptions =
          tags: "".split(" "),
          id: "557"
          ;
          initTagRenderer("".split(" "), "".split(" "), channelOptions);

          StackExchange.using("externalEditor", function()
          // Have to fire editor after snippets, if snippets enabled
          if (StackExchange.settings.snippets.snippetsEnabled)
          StackExchange.using("snippets", function()
          createEditor();
          );

          else
          createEditor();

          );

          function createEditor()
          StackExchange.prepareEditor(
          heartbeatType: 'answer',
          autoActivateHeartbeat: false,
          convertImagesToLinks: false,
          noModals: true,
          showLowRepImageUploadWarning: true,
          reputationToPostImages: null,
          bindNavPrevention: true,
          postfix: "",
          imageUploader:
          brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
          contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
          allowUrls: true
          ,
          onDemand: true,
          discardSelector: ".discard-answer"
          ,immediatelyShowMarkdownHelp:true
          );



          );






          stranger is a new contributor. Be nice, and check out our Code of Conduct.









          draft saved

          draft discarded


















          StackExchange.ready(
          function ()
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fdatascience.stackexchange.com%2fquestions%2f47720%2fvalidation-loss-increases-and-validation-accuracy-decreases%23new-answer', 'question_page');

          );

          Post as a guest















          Required, but never shown

























          1 Answer
          1






          active

          oldest

          votes








          1 Answer
          1






          active

          oldest

          votes









          active

          oldest

          votes






          active

          oldest

          votes









          0












          $begingroup$

          What you are experiencing is known as overfitting, and it’s a common problem in machine learning and data science.



          Overfitting happens when a model begins to focus on the noise in the training dataset and extracts features based on it. This helps the model to improve its performance on the training set, but hurts its ability to generalize, so the accuracy of the validation set decreases.



          To deal with overfitting, you need to use regularization during the training. You ca try adding dropout layers or batch-normalization layers, adding weight regularization, or you can artificially increase the size of your training set by performing some data augmentation.



          You can read more about it in the following post:



          What are the possible approaches to fixing Overfitting on a CNN?






          share|improve this answer









          $endgroup$












          • $begingroup$
            Thanks, I tried adding regularizers to Conv1D and Dense layers as below. Conv1D(filters=64, kernel_size=8, activation='relu', kernel_regularizer=l2(0.001), bias_regularizer=l2(0.001)) AND Dense(10, activation='relu', kernel_regularizer=l2(0.001)) . But unfortunately none of them did any changes to val_acc and val_loss.
            $endgroup$
            – stranger
            yesterday















          0












          $begingroup$

          What you are experiencing is known as overfitting, and it’s a common problem in machine learning and data science.



          Overfitting happens when a model begins to focus on the noise in the training dataset and extracts features based on it. This helps the model to improve its performance on the training set, but hurts its ability to generalize, so the accuracy of the validation set decreases.



          To deal with overfitting, you need to use regularization during the training. You ca try adding dropout layers or batch-normalization layers, adding weight regularization, or you can artificially increase the size of your training set by performing some data augmentation.



          You can read more about it in the following post:



          What are the possible approaches to fixing Overfitting on a CNN?






          share|improve this answer









          $endgroup$












          • $begingroup$
            Thanks, I tried adding regularizers to Conv1D and Dense layers as below. Conv1D(filters=64, kernel_size=8, activation='relu', kernel_regularizer=l2(0.001), bias_regularizer=l2(0.001)) AND Dense(10, activation='relu', kernel_regularizer=l2(0.001)) . But unfortunately none of them did any changes to val_acc and val_loss.
            $endgroup$
            – stranger
            yesterday













          0












          0








          0





          $begingroup$

          What you are experiencing is known as overfitting, and it’s a common problem in machine learning and data science.



          Overfitting happens when a model begins to focus on the noise in the training dataset and extracts features based on it. This helps the model to improve its performance on the training set, but hurts its ability to generalize, so the accuracy of the validation set decreases.



          To deal with overfitting, you need to use regularization during the training. You ca try adding dropout layers or batch-normalization layers, adding weight regularization, or you can artificially increase the size of your training set by performing some data augmentation.



          You can read more about it in the following post:



          What are the possible approaches to fixing Overfitting on a CNN?






          share|improve this answer









          $endgroup$



          What you are experiencing is known as overfitting, and it’s a common problem in machine learning and data science.



          Overfitting happens when a model begins to focus on the noise in the training dataset and extracts features based on it. This helps the model to improve its performance on the training set, but hurts its ability to generalize, so the accuracy of the validation set decreases.



          To deal with overfitting, you need to use regularization during the training. You ca try adding dropout layers or batch-normalization layers, adding weight regularization, or you can artificially increase the size of your training set by performing some data augmentation.



          You can read more about it in the following post:



          What are the possible approaches to fixing Overfitting on a CNN?







          share|improve this answer












          share|improve this answer



          share|improve this answer










          answered yesterday









          Mark.FMark.F

          9991418




          9991418











          • $begingroup$
            Thanks, I tried adding regularizers to Conv1D and Dense layers as below. Conv1D(filters=64, kernel_size=8, activation='relu', kernel_regularizer=l2(0.001), bias_regularizer=l2(0.001)) AND Dense(10, activation='relu', kernel_regularizer=l2(0.001)) . But unfortunately none of them did any changes to val_acc and val_loss.
            $endgroup$
            – stranger
            yesterday
















          • $begingroup$
            Thanks, I tried adding regularizers to Conv1D and Dense layers as below. Conv1D(filters=64, kernel_size=8, activation='relu', kernel_regularizer=l2(0.001), bias_regularizer=l2(0.001)) AND Dense(10, activation='relu', kernel_regularizer=l2(0.001)) . But unfortunately none of them did any changes to val_acc and val_loss.
            $endgroup$
            – stranger
            yesterday















          $begingroup$
          Thanks, I tried adding regularizers to Conv1D and Dense layers as below. Conv1D(filters=64, kernel_size=8, activation='relu', kernel_regularizer=l2(0.001), bias_regularizer=l2(0.001)) AND Dense(10, activation='relu', kernel_regularizer=l2(0.001)) . But unfortunately none of them did any changes to val_acc and val_loss.
          $endgroup$
          – stranger
          yesterday




          $begingroup$
          Thanks, I tried adding regularizers to Conv1D and Dense layers as below. Conv1D(filters=64, kernel_size=8, activation='relu', kernel_regularizer=l2(0.001), bias_regularizer=l2(0.001)) AND Dense(10, activation='relu', kernel_regularizer=l2(0.001)) . But unfortunately none of them did any changes to val_acc and val_loss.
          $endgroup$
          – stranger
          yesterday










          stranger is a new contributor. Be nice, and check out our Code of Conduct.









          draft saved

          draft discarded


















          stranger is a new contributor. Be nice, and check out our Code of Conduct.












          stranger is a new contributor. Be nice, and check out our Code of Conduct.











          stranger is a new contributor. Be nice, and check out our Code of Conduct.














          Thanks for contributing an answer to Data Science Stack Exchange!


          • Please be sure to answer the question. Provide details and share your research!

          But avoid


          • Asking for help, clarification, or responding to other answers.

          • Making statements based on opinion; back them up with references or personal experience.

          Use MathJax to format equations. MathJax reference.


          To learn more, see our tips on writing great answers.




          draft saved


          draft discarded














          StackExchange.ready(
          function ()
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fdatascience.stackexchange.com%2fquestions%2f47720%2fvalidation-loss-increases-and-validation-accuracy-decreases%23new-answer', 'question_page');

          );

          Post as a guest















          Required, but never shown





















































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown

































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown







          Popular posts from this blog

          Adding axes to figuresAdding axes labels to LaTeX figuresLaTeX equivalent of ConTeXt buffersRotate a node but not its content: the case of the ellipse decorationHow to define the default vertical distance between nodes?TikZ scaling graphic and adjust node position and keep font sizeNumerical conditional within tikz keys?adding axes to shapesAlign axes across subfiguresAdding figures with a certain orderLine up nested tikz enviroments or how to get rid of themAdding axes labels to LaTeX figures

          Tähtien Talli Jäsenet | Lähteet | NavigointivalikkoSuomen Hippos – Tähtien Talli

          Do these cracks on my tires look bad? The Next CEO of Stack OverflowDry rot tire should I replace?Having to replace tiresFishtailed so easily? Bad tires? ABS?Filling the tires with something other than air, to avoid puncture hassles?Used Michelin tires safe to install?Do these tyre cracks necessitate replacement?Rumbling noise: tires or mechanicalIs it possible to fix noisy feathered tires?Are bad winter tires still better than summer tires in winter?Torque converter failure - Related to replacing only 2 tires?Why use snow tires on all 4 wheels on 2-wheel-drive cars?