Problem about tuning hyper-parametressklearn - overfitting problemStrategies for automatically tuning the hyper-parameters of deep learning modelsAutomated tuning of HyperparameterWhich parameters are hyper parameters in a linear regression?Hyper parameters and ValidationSetOverfitting problem in modelModel Selection with Oversampling/ Cross-Validation leads to similar test results in 2 approachesHyperparameter tuning for stacked modelsDuring a regression task, I am getting low R^2 values, but elementwise difference between test set and prediction values is hugeHyper-parameter tuning when you don't have an access to the test data
Why does Surtur say that Thor is Asgard's doom?
How do researchers send unsolicited emails asking for feedback on their works?
Are hand made posters acceptable in Academia?
Determine voltage drop over 10G resistors with cheap multimeter
"Marked down as someone wanting to sell shares." What does that mean?
Do people actually use the word "kaputt" in conversation?
Parts of mini page are not placed properly
What is the difference between something being completely legal and being completely decriminalized?
Does fire aspect on a sword, destroy mob drops?
What should be the ideal length of sentences in a blog post for ease of reading?
I got the following comment from a reputed math journal. What does it mean?
Exposing a company lying about themselves in a tightly knit industry (videogames) : Is my career at risk on the long run?
What do the positive and negative (+/-) transmit and receive pins mean on Ethernet cables?
God... independent
Do native speakers use "ultima" and "proxima" frequently in spoken English?
SQL PRINT vs SQL EXEC
What kind of footwear is suitable for walking in micro gravity environment?
What are the consequences of changing the number of hours in a day?
Can other pieces capture a threatening piece and prevent a checkmate?
Exit shell with shortcut (not typing exit) that closes session properly
Why are there no stars visible in cislunar space?
How do you justify more code being written by following clean code practices?
Would this string work as string?
Jem'Hadar, something strange about their life expectancy
Problem about tuning hyper-parametres
sklearn - overfitting problemStrategies for automatically tuning the hyper-parameters of deep learning modelsAutomated tuning of HyperparameterWhich parameters are hyper parameters in a linear regression?Hyper parameters and ValidationSetOverfitting problem in modelModel Selection with Oversampling/ Cross-Validation leads to similar test results in 2 approachesHyperparameter tuning for stacked modelsDuring a regression task, I am getting low R^2 values, but elementwise difference between test set and prediction values is hugeHyper-parameter tuning when you don't have an access to the test data
$begingroup$
I have tried GridSearchCV and BayesSearchCV for tuning my lightGBM algorithm (for binary classification).
I have used 10 iterations. And I have indicated scoring ="roc_auc"
In the first iteration, I have got:
best score (e.g :0.71...)
and best param (e.g: max-depth: 10 , learning-rate: 0.17..., num-leave:175, n-estimators: 176, ....)
In the 10th iteration, I have got :
best score (e.g :0.72...)
and best param (e.g: max-depth: 9 , learning-rate: 0.19..., num-leave:168, n-estimators: 172, ....)
Then I tried to train my LightGBM classifier with the 10th param (which supposed that it get the best score!!). I have got :
AUC : (0.7541.., 0.6467..)
Accuracy: 0.7338..
RMSE: 0.5216..
Then because I had some curiosity I have tried to train my classifier with (best param) of the First iteration (which is considered as worst score)!. I had surprised by the result that I have got:
AUC : (0.7545., 0.6592..)
Accuracy: 0.7332..
RMSE: 0.5152..
Because I have fixed previously scoring by roc-auc, Generally, I should get AUC in the 10th iteration better than the first iteration but I have got the contrary.
I have supposed that it is considered AUC train of 10th iteration 0.7541 as better than 0.7545 of the 1st iteration because of the overfitting but when I tried to check the 3rd and the 5th iteration I get on the 3rd : 0.7532 and in the 5th : 0.7548.
So I don't know what the best score in this algorithm means. And what is its role exactly if it gets values as the described situation. I had tried then many times with other tuning parameters but I have got the same case. I don't know where is the problem exactly.
machine-learning data-mining xgboost hyperparameter
$endgroup$
add a comment |
$begingroup$
I have tried GridSearchCV and BayesSearchCV for tuning my lightGBM algorithm (for binary classification).
I have used 10 iterations. And I have indicated scoring ="roc_auc"
In the first iteration, I have got:
best score (e.g :0.71...)
and best param (e.g: max-depth: 10 , learning-rate: 0.17..., num-leave:175, n-estimators: 176, ....)
In the 10th iteration, I have got :
best score (e.g :0.72...)
and best param (e.g: max-depth: 9 , learning-rate: 0.19..., num-leave:168, n-estimators: 172, ....)
Then I tried to train my LightGBM classifier with the 10th param (which supposed that it get the best score!!). I have got :
AUC : (0.7541.., 0.6467..)
Accuracy: 0.7338..
RMSE: 0.5216..
Then because I had some curiosity I have tried to train my classifier with (best param) of the First iteration (which is considered as worst score)!. I had surprised by the result that I have got:
AUC : (0.7545., 0.6592..)
Accuracy: 0.7332..
RMSE: 0.5152..
Because I have fixed previously scoring by roc-auc, Generally, I should get AUC in the 10th iteration better than the first iteration but I have got the contrary.
I have supposed that it is considered AUC train of 10th iteration 0.7541 as better than 0.7545 of the 1st iteration because of the overfitting but when I tried to check the 3rd and the 5th iteration I get on the 3rd : 0.7532 and in the 5th : 0.7548.
So I don't know what the best score in this algorithm means. And what is its role exactly if it gets values as the described situation. I had tried then many times with other tuning parameters but I have got the same case. I don't know where is the problem exactly.
machine-learning data-mining xgboost hyperparameter
$endgroup$
add a comment |
$begingroup$
I have tried GridSearchCV and BayesSearchCV for tuning my lightGBM algorithm (for binary classification).
I have used 10 iterations. And I have indicated scoring ="roc_auc"
In the first iteration, I have got:
best score (e.g :0.71...)
and best param (e.g: max-depth: 10 , learning-rate: 0.17..., num-leave:175, n-estimators: 176, ....)
In the 10th iteration, I have got :
best score (e.g :0.72...)
and best param (e.g: max-depth: 9 , learning-rate: 0.19..., num-leave:168, n-estimators: 172, ....)
Then I tried to train my LightGBM classifier with the 10th param (which supposed that it get the best score!!). I have got :
AUC : (0.7541.., 0.6467..)
Accuracy: 0.7338..
RMSE: 0.5216..
Then because I had some curiosity I have tried to train my classifier with (best param) of the First iteration (which is considered as worst score)!. I had surprised by the result that I have got:
AUC : (0.7545., 0.6592..)
Accuracy: 0.7332..
RMSE: 0.5152..
Because I have fixed previously scoring by roc-auc, Generally, I should get AUC in the 10th iteration better than the first iteration but I have got the contrary.
I have supposed that it is considered AUC train of 10th iteration 0.7541 as better than 0.7545 of the 1st iteration because of the overfitting but when I tried to check the 3rd and the 5th iteration I get on the 3rd : 0.7532 and in the 5th : 0.7548.
So I don't know what the best score in this algorithm means. And what is its role exactly if it gets values as the described situation. I had tried then many times with other tuning parameters but I have got the same case. I don't know where is the problem exactly.
machine-learning data-mining xgboost hyperparameter
$endgroup$
I have tried GridSearchCV and BayesSearchCV for tuning my lightGBM algorithm (for binary classification).
I have used 10 iterations. And I have indicated scoring ="roc_auc"
In the first iteration, I have got:
best score (e.g :0.71...)
and best param (e.g: max-depth: 10 , learning-rate: 0.17..., num-leave:175, n-estimators: 176, ....)
In the 10th iteration, I have got :
best score (e.g :0.72...)
and best param (e.g: max-depth: 9 , learning-rate: 0.19..., num-leave:168, n-estimators: 172, ....)
Then I tried to train my LightGBM classifier with the 10th param (which supposed that it get the best score!!). I have got :
AUC : (0.7541.., 0.6467..)
Accuracy: 0.7338..
RMSE: 0.5216..
Then because I had some curiosity I have tried to train my classifier with (best param) of the First iteration (which is considered as worst score)!. I had surprised by the result that I have got:
AUC : (0.7545., 0.6592..)
Accuracy: 0.7332..
RMSE: 0.5152..
Because I have fixed previously scoring by roc-auc, Generally, I should get AUC in the 10th iteration better than the first iteration but I have got the contrary.
I have supposed that it is considered AUC train of 10th iteration 0.7541 as better than 0.7545 of the 1st iteration because of the overfitting but when I tried to check the 3rd and the 5th iteration I get on the 3rd : 0.7532 and in the 5th : 0.7548.
So I don't know what the best score in this algorithm means. And what is its role exactly if it gets values as the described situation. I had tried then many times with other tuning parameters but I have got the same case. I don't know where is the problem exactly.
machine-learning data-mining xgboost hyperparameter
machine-learning data-mining xgboost hyperparameter
edited 6 mins ago
amal amal
asked 11 mins ago
amal amalamal amal
225
225
add a comment |
add a comment |
0
active
oldest
votes
Your Answer
StackExchange.ifUsing("editor", function ()
return StackExchange.using("mathjaxEditing", function ()
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix)
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
);
);
, "mathjax-editing");
StackExchange.ready(function()
var channelOptions =
tags: "".split(" "),
id: "557"
;
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function()
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled)
StackExchange.using("snippets", function()
createEditor();
);
else
createEditor();
);
function createEditor()
StackExchange.prepareEditor(
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: false,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: null,
bindNavPrevention: true,
postfix: "",
imageUploader:
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
,
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
);
);
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fdatascience.stackexchange.com%2fquestions%2f47606%2fproblem-about-tuning-hyper-parametres%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
0
active
oldest
votes
0
active
oldest
votes
active
oldest
votes
active
oldest
votes
Thanks for contributing an answer to Data Science Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fdatascience.stackexchange.com%2fquestions%2f47606%2fproblem-about-tuning-hyper-parametres%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown