Problem about tuning hyper-parametressklearn - overfitting problemStrategies for automatically tuning the hyper-parameters of deep learning modelsAutomated tuning of HyperparameterWhich parameters are hyper parameters in a linear regression?Hyper parameters and ValidationSetOverfitting problem in modelModel Selection with Oversampling/ Cross-Validation leads to similar test results in 2 approachesHyperparameter tuning for stacked modelsDuring a regression task, I am getting low R^2 values, but elementwise difference between test set and prediction values is hugeHyper-parameter tuning when you don't have an access to the test data

Why does Surtur say that Thor is Asgard's doom?

How do researchers send unsolicited emails asking for feedback on their works?

Are hand made posters acceptable in Academia?

Determine voltage drop over 10G resistors with cheap multimeter

"Marked down as someone wanting to sell shares." What does that mean?

Do people actually use the word "kaputt" in conversation?

Parts of mini page are not placed properly

What is the difference between something being completely legal and being completely decriminalized?

Does fire aspect on a sword, destroy mob drops?

What should be the ideal length of sentences in a blog post for ease of reading?

I got the following comment from a reputed math journal. What does it mean?

Exposing a company lying about themselves in a tightly knit industry (videogames) : Is my career at risk on the long run?

What do the positive and negative (+/-) transmit and receive pins mean on Ethernet cables?

God... independent

Do native speakers use "ultima" and "proxima" frequently in spoken English?

SQL PRINT vs SQL EXEC

What kind of footwear is suitable for walking in micro gravity environment?

What are the consequences of changing the number of hours in a day?

Can other pieces capture a threatening piece and prevent a checkmate?

Exit shell with shortcut (not typing exit) that closes session properly

Why are there no stars visible in cislunar space?

How do you justify more code being written by following clean code practices?

Would this string work as string?

Jem'Hadar, something strange about their life expectancy



Problem about tuning hyper-parametres


sklearn - overfitting problemStrategies for automatically tuning the hyper-parameters of deep learning modelsAutomated tuning of HyperparameterWhich parameters are hyper parameters in a linear regression?Hyper parameters and ValidationSetOverfitting problem in modelModel Selection with Oversampling/ Cross-Validation leads to similar test results in 2 approachesHyperparameter tuning for stacked modelsDuring a regression task, I am getting low R^2 values, but elementwise difference between test set and prediction values is hugeHyper-parameter tuning when you don't have an access to the test data













0












$begingroup$


I have tried GridSearchCV and BayesSearchCV for tuning my lightGBM algorithm (for binary classification).



I have used 10 iterations. And I have indicated scoring ="roc_auc"



In the first iteration, I have got:



best score (e.g :0.71...) 
and best param (e.g: max-depth: 10 , learning-rate: 0.17..., num-leave:175, n-estimators: 176, ....)


In the 10th iteration, I have got :



best score (e.g :0.72...) 
and best param (e.g: max-depth: 9 , learning-rate: 0.19..., num-leave:168, n-estimators: 172, ....)


Then I tried to train my LightGBM classifier with the 10th param (which supposed that it get the best score!!). I have got :



AUC : (0.7541.., 0.6467..)
Accuracy: 0.7338..
RMSE: 0.5216..


Then because I had some curiosity I have tried to train my classifier with (best param) of the First iteration (which is considered as worst score)!. I had surprised by the result that I have got:



AUC : (0.7545., 0.6592..)
Accuracy: 0.7332..
RMSE: 0.5152..


Because I have fixed previously scoring by roc-auc, Generally, I should get AUC in the 10th iteration better than the first iteration but I have got the contrary.



I have supposed that it is considered AUC train of 10th iteration 0.7541 as better than 0.7545 of the 1st iteration because of the overfitting but when I tried to check the 3rd and the 5th iteration I get on the 3rd : 0.7532 and in the 5th : 0.7548.



So I don't know what the best score in this algorithm means. And what is its role exactly if it gets values as the described situation. I had tried then many times with other tuning parameters but I have got the same case. I don't know where is the problem exactly.










share|improve this question











$endgroup$
















    0












    $begingroup$


    I have tried GridSearchCV and BayesSearchCV for tuning my lightGBM algorithm (for binary classification).



    I have used 10 iterations. And I have indicated scoring ="roc_auc"



    In the first iteration, I have got:



    best score (e.g :0.71...) 
    and best param (e.g: max-depth: 10 , learning-rate: 0.17..., num-leave:175, n-estimators: 176, ....)


    In the 10th iteration, I have got :



    best score (e.g :0.72...) 
    and best param (e.g: max-depth: 9 , learning-rate: 0.19..., num-leave:168, n-estimators: 172, ....)


    Then I tried to train my LightGBM classifier with the 10th param (which supposed that it get the best score!!). I have got :



    AUC : (0.7541.., 0.6467..)
    Accuracy: 0.7338..
    RMSE: 0.5216..


    Then because I had some curiosity I have tried to train my classifier with (best param) of the First iteration (which is considered as worst score)!. I had surprised by the result that I have got:



    AUC : (0.7545., 0.6592..)
    Accuracy: 0.7332..
    RMSE: 0.5152..


    Because I have fixed previously scoring by roc-auc, Generally, I should get AUC in the 10th iteration better than the first iteration but I have got the contrary.



    I have supposed that it is considered AUC train of 10th iteration 0.7541 as better than 0.7545 of the 1st iteration because of the overfitting but when I tried to check the 3rd and the 5th iteration I get on the 3rd : 0.7532 and in the 5th : 0.7548.



    So I don't know what the best score in this algorithm means. And what is its role exactly if it gets values as the described situation. I had tried then many times with other tuning parameters but I have got the same case. I don't know where is the problem exactly.










    share|improve this question











    $endgroup$














      0












      0








      0





      $begingroup$


      I have tried GridSearchCV and BayesSearchCV for tuning my lightGBM algorithm (for binary classification).



      I have used 10 iterations. And I have indicated scoring ="roc_auc"



      In the first iteration, I have got:



      best score (e.g :0.71...) 
      and best param (e.g: max-depth: 10 , learning-rate: 0.17..., num-leave:175, n-estimators: 176, ....)


      In the 10th iteration, I have got :



      best score (e.g :0.72...) 
      and best param (e.g: max-depth: 9 , learning-rate: 0.19..., num-leave:168, n-estimators: 172, ....)


      Then I tried to train my LightGBM classifier with the 10th param (which supposed that it get the best score!!). I have got :



      AUC : (0.7541.., 0.6467..)
      Accuracy: 0.7338..
      RMSE: 0.5216..


      Then because I had some curiosity I have tried to train my classifier with (best param) of the First iteration (which is considered as worst score)!. I had surprised by the result that I have got:



      AUC : (0.7545., 0.6592..)
      Accuracy: 0.7332..
      RMSE: 0.5152..


      Because I have fixed previously scoring by roc-auc, Generally, I should get AUC in the 10th iteration better than the first iteration but I have got the contrary.



      I have supposed that it is considered AUC train of 10th iteration 0.7541 as better than 0.7545 of the 1st iteration because of the overfitting but when I tried to check the 3rd and the 5th iteration I get on the 3rd : 0.7532 and in the 5th : 0.7548.



      So I don't know what the best score in this algorithm means. And what is its role exactly if it gets values as the described situation. I had tried then many times with other tuning parameters but I have got the same case. I don't know where is the problem exactly.










      share|improve this question











      $endgroup$




      I have tried GridSearchCV and BayesSearchCV for tuning my lightGBM algorithm (for binary classification).



      I have used 10 iterations. And I have indicated scoring ="roc_auc"



      In the first iteration, I have got:



      best score (e.g :0.71...) 
      and best param (e.g: max-depth: 10 , learning-rate: 0.17..., num-leave:175, n-estimators: 176, ....)


      In the 10th iteration, I have got :



      best score (e.g :0.72...) 
      and best param (e.g: max-depth: 9 , learning-rate: 0.19..., num-leave:168, n-estimators: 172, ....)


      Then I tried to train my LightGBM classifier with the 10th param (which supposed that it get the best score!!). I have got :



      AUC : (0.7541.., 0.6467..)
      Accuracy: 0.7338..
      RMSE: 0.5216..


      Then because I had some curiosity I have tried to train my classifier with (best param) of the First iteration (which is considered as worst score)!. I had surprised by the result that I have got:



      AUC : (0.7545., 0.6592..)
      Accuracy: 0.7332..
      RMSE: 0.5152..


      Because I have fixed previously scoring by roc-auc, Generally, I should get AUC in the 10th iteration better than the first iteration but I have got the contrary.



      I have supposed that it is considered AUC train of 10th iteration 0.7541 as better than 0.7545 of the 1st iteration because of the overfitting but when I tried to check the 3rd and the 5th iteration I get on the 3rd : 0.7532 and in the 5th : 0.7548.



      So I don't know what the best score in this algorithm means. And what is its role exactly if it gets values as the described situation. I had tried then many times with other tuning parameters but I have got the same case. I don't know where is the problem exactly.







      machine-learning data-mining xgboost hyperparameter






      share|improve this question















      share|improve this question













      share|improve this question




      share|improve this question








      edited 6 mins ago







      amal amal

















      asked 11 mins ago









      amal amalamal amal

      225




      225




















          0






          active

          oldest

          votes











          Your Answer





          StackExchange.ifUsing("editor", function ()
          return StackExchange.using("mathjaxEditing", function ()
          StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix)
          StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
          );
          );
          , "mathjax-editing");

          StackExchange.ready(function()
          var channelOptions =
          tags: "".split(" "),
          id: "557"
          ;
          initTagRenderer("".split(" "), "".split(" "), channelOptions);

          StackExchange.using("externalEditor", function()
          // Have to fire editor after snippets, if snippets enabled
          if (StackExchange.settings.snippets.snippetsEnabled)
          StackExchange.using("snippets", function()
          createEditor();
          );

          else
          createEditor();

          );

          function createEditor()
          StackExchange.prepareEditor(
          heartbeatType: 'answer',
          autoActivateHeartbeat: false,
          convertImagesToLinks: false,
          noModals: true,
          showLowRepImageUploadWarning: true,
          reputationToPostImages: null,
          bindNavPrevention: true,
          postfix: "",
          imageUploader:
          brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
          contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
          allowUrls: true
          ,
          onDemand: true,
          discardSelector: ".discard-answer"
          ,immediatelyShowMarkdownHelp:true
          );



          );













          draft saved

          draft discarded


















          StackExchange.ready(
          function ()
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fdatascience.stackexchange.com%2fquestions%2f47606%2fproblem-about-tuning-hyper-parametres%23new-answer', 'question_page');

          );

          Post as a guest















          Required, but never shown

























          0






          active

          oldest

          votes








          0






          active

          oldest

          votes









          active

          oldest

          votes






          active

          oldest

          votes















          draft saved

          draft discarded
















































          Thanks for contributing an answer to Data Science Stack Exchange!


          • Please be sure to answer the question. Provide details and share your research!

          But avoid


          • Asking for help, clarification, or responding to other answers.

          • Making statements based on opinion; back them up with references or personal experience.

          Use MathJax to format equations. MathJax reference.


          To learn more, see our tips on writing great answers.




          draft saved


          draft discarded














          StackExchange.ready(
          function ()
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fdatascience.stackexchange.com%2fquestions%2f47606%2fproblem-about-tuning-hyper-parametres%23new-answer', 'question_page');

          );

          Post as a guest















          Required, but never shown





















































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown

































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown







          Popular posts from this blog

          Adding axes to figuresAdding axes labels to LaTeX figuresLaTeX equivalent of ConTeXt buffersRotate a node but not its content: the case of the ellipse decorationHow to define the default vertical distance between nodes?TikZ scaling graphic and adjust node position and keep font sizeNumerical conditional within tikz keys?adding axes to shapesAlign axes across subfiguresAdding figures with a certain orderLine up nested tikz enviroments or how to get rid of themAdding axes labels to LaTeX figures

          Tähtien Talli Jäsenet | Lähteet | NavigointivalikkoSuomen Hippos – Tähtien Talli

          Do these cracks on my tires look bad? The Next CEO of Stack OverflowDry rot tire should I replace?Having to replace tiresFishtailed so easily? Bad tires? ABS?Filling the tires with something other than air, to avoid puncture hassles?Used Michelin tires safe to install?Do these tyre cracks necessitate replacement?Rumbling noise: tires or mechanicalIs it possible to fix noisy feathered tires?Are bad winter tires still better than summer tires in winter?Torque converter failure - Related to replacing only 2 tires?Why use snow tires on all 4 wheels on 2-wheel-drive cars?