What is the differences in the Gini Index, Chi-Square, and Information Gain splitting methods? Announcing the arrival of Valued Associate #679: Cesar Manara Planned maintenance scheduled April 23, 2019 at 23:30 UTC (7:30pm US/Eastern) 2019 Moderator Election Q&A - Questionnaire 2019 Community Moderator Election ResultsGini Impurity vs Entropywhat is the difference between “fully developed decision trees” and “shallow decision trees”?Gini Impurity vs EntropyFit Decision Tree to Gradient Boosted Trees for InterpretabilityMore features hurts when underfitting?Decision Trees - C4.5 vs CART - rule setsWhat exactly is a Gini IndexHow Can I Compute Information-Gain for Continuous- Valued AttributesWhen does decision tree perform better than the neural network?Why neural networks do not perform well on structured data?Decision Tree - Preprocessing for very sparse features
Positioning dot before text in math mode
Does silver oxide react with hydrogen sulfide?
I can't produce songs
What is the difference between a "ranged attack" and a "ranged weapon attack"?
"klopfte jemand" or "jemand klopfte"?
Where is the Next Backup Size entry on iOS 12?
Why weren't discrete x86 CPUs ever used in game hardware?
Central Vacuuming: Is it worth it, and how does it compare to normal vacuuming?
Delete free apps from library
Mounting TV on a weird wall that has some material between the drywall and stud
My mentor says to set image to Fine instead of RAW — how is this different from JPG?
Found this skink in my tomato plant bucket. Is he trapped? Or could he leave if he wanted?
Can an iPhone 7 be made to function as a NFC Tag?
Why is it faster to reheat something than it is to cook it?
What is the difference between CTSS and ITS?
Trying to understand entropy as a novice in thermodynamics
Co-worker has annoying ringtone
Would color changing eyes affect vision?
Flight departed from the gate 5 min before scheduled departure time. Refund options
Why do early math courses focus on the cross sections of a cone and not on other 3D objects?
Are the endpoints of the domain of a function counted as critical points?
Why complex landing gears are used instead of simple,reliability and light weight muscle wire or shape memory alloys?
Is there public access to the Meteor Crater in Arizona?
How many time has Arya actually used Needle?
What is the differences in the Gini Index, Chi-Square, and Information Gain splitting methods?
Announcing the arrival of Valued Associate #679: Cesar Manara
Planned maintenance scheduled April 23, 2019 at 23:30 UTC (7:30pm US/Eastern)
2019 Moderator Election Q&A - Questionnaire
2019 Community Moderator Election ResultsGini Impurity vs Entropywhat is the difference between “fully developed decision trees” and “shallow decision trees”?Gini Impurity vs EntropyFit Decision Tree to Gradient Boosted Trees for InterpretabilityMore features hurts when underfitting?Decision Trees - C4.5 vs CART - rule setsWhat exactly is a Gini IndexHow Can I Compute Information-Gain for Continuous- Valued AttributesWhen does decision tree perform better than the neural network?Why neural networks do not perform well on structured data?Decision Tree - Preprocessing for very sparse features
$begingroup$
I am looking through decision trees, and I do not understand what makes each of these methods different. Could someone explain clearly what the difference between these is? Thank you.
decision-trees
$endgroup$
add a comment |
$begingroup$
I am looking through decision trees, and I do not understand what makes each of these methods different. Could someone explain clearly what the difference between these is? Thank you.
decision-trees
$endgroup$
$begingroup$
Welcome to the site! Look at this post on this site, and this post on medium (explains all three with example).
$endgroup$
– Esmailian
Apr 4 at 11:34
add a comment |
$begingroup$
I am looking through decision trees, and I do not understand what makes each of these methods different. Could someone explain clearly what the difference between these is? Thank you.
decision-trees
$endgroup$
I am looking through decision trees, and I do not understand what makes each of these methods different. Could someone explain clearly what the difference between these is? Thank you.
decision-trees
decision-trees
asked Apr 4 at 1:40
cheeztcheezt
61
61
$begingroup$
Welcome to the site! Look at this post on this site, and this post on medium (explains all three with example).
$endgroup$
– Esmailian
Apr 4 at 11:34
add a comment |
$begingroup$
Welcome to the site! Look at this post on this site, and this post on medium (explains all three with example).
$endgroup$
– Esmailian
Apr 4 at 11:34
$begingroup$
Welcome to the site! Look at this post on this site, and this post on medium (explains all three with example).
$endgroup$
– Esmailian
Apr 4 at 11:34
$begingroup$
Welcome to the site! Look at this post on this site, and this post on medium (explains all three with example).
$endgroup$
– Esmailian
Apr 4 at 11:34
add a comment |
1 Answer
1
active
oldest
votes
$begingroup$
As I understand it, all three want to minimize the false classified data points in your data set. (Logically, if you look for what decision trees are used)
But each of them comes from another side to this problem.
gini impurity wants "better as random"
It compares the "I label random data with random labels" against the labeling after possible split by decision tree (Wish is, that you can split the tree with better outcome than "random random random")
information gain wants small trees
It uses knowledge from information theory. It models the difference between "good" and "bad" split with criteria "simple/small trees preferred". As a result of this, it want to split the data in a way, that the daughters are "pure as possible".
For the chi-square ... I have found two things: CHAID, a (seemingly complex) decision tree technique and the chi square to prune decision trees after their building.
The chi square in general has its roots in biological statistics. It gives a characteristic number how the observed distribution conform with the null hypothesis one have about this distribution. (Biology have to act like this a lot. "I observe something, I search for an explanation, I form a hypothesis, I probe if this is statistical confirmable")
For formulas please look in Wikipedia and other sources.
$endgroup$
add a comment |
Your Answer
StackExchange.ready(function()
var channelOptions =
tags: "".split(" "),
id: "557"
;
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function()
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled)
StackExchange.using("snippets", function()
createEditor();
);
else
createEditor();
);
function createEditor()
StackExchange.prepareEditor(
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: false,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: null,
bindNavPrevention: true,
postfix: "",
imageUploader:
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
,
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
);
);
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fdatascience.stackexchange.com%2fquestions%2f48560%2fwhat-is-the-differences-in-the-gini-index-chi-square-and-information-gain-spli%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
1 Answer
1
active
oldest
votes
1 Answer
1
active
oldest
votes
active
oldest
votes
active
oldest
votes
$begingroup$
As I understand it, all three want to minimize the false classified data points in your data set. (Logically, if you look for what decision trees are used)
But each of them comes from another side to this problem.
gini impurity wants "better as random"
It compares the "I label random data with random labels" against the labeling after possible split by decision tree (Wish is, that you can split the tree with better outcome than "random random random")
information gain wants small trees
It uses knowledge from information theory. It models the difference between "good" and "bad" split with criteria "simple/small trees preferred". As a result of this, it want to split the data in a way, that the daughters are "pure as possible".
For the chi-square ... I have found two things: CHAID, a (seemingly complex) decision tree technique and the chi square to prune decision trees after their building.
The chi square in general has its roots in biological statistics. It gives a characteristic number how the observed distribution conform with the null hypothesis one have about this distribution. (Biology have to act like this a lot. "I observe something, I search for an explanation, I form a hypothesis, I probe if this is statistical confirmable")
For formulas please look in Wikipedia and other sources.
$endgroup$
add a comment |
$begingroup$
As I understand it, all three want to minimize the false classified data points in your data set. (Logically, if you look for what decision trees are used)
But each of them comes from another side to this problem.
gini impurity wants "better as random"
It compares the "I label random data with random labels" against the labeling after possible split by decision tree (Wish is, that you can split the tree with better outcome than "random random random")
information gain wants small trees
It uses knowledge from information theory. It models the difference between "good" and "bad" split with criteria "simple/small trees preferred". As a result of this, it want to split the data in a way, that the daughters are "pure as possible".
For the chi-square ... I have found two things: CHAID, a (seemingly complex) decision tree technique and the chi square to prune decision trees after their building.
The chi square in general has its roots in biological statistics. It gives a characteristic number how the observed distribution conform with the null hypothesis one have about this distribution. (Biology have to act like this a lot. "I observe something, I search for an explanation, I form a hypothesis, I probe if this is statistical confirmable")
For formulas please look in Wikipedia and other sources.
$endgroup$
add a comment |
$begingroup$
As I understand it, all three want to minimize the false classified data points in your data set. (Logically, if you look for what decision trees are used)
But each of them comes from another side to this problem.
gini impurity wants "better as random"
It compares the "I label random data with random labels" against the labeling after possible split by decision tree (Wish is, that you can split the tree with better outcome than "random random random")
information gain wants small trees
It uses knowledge from information theory. It models the difference between "good" and "bad" split with criteria "simple/small trees preferred". As a result of this, it want to split the data in a way, that the daughters are "pure as possible".
For the chi-square ... I have found two things: CHAID, a (seemingly complex) decision tree technique and the chi square to prune decision trees after their building.
The chi square in general has its roots in biological statistics. It gives a characteristic number how the observed distribution conform with the null hypothesis one have about this distribution. (Biology have to act like this a lot. "I observe something, I search for an explanation, I form a hypothesis, I probe if this is statistical confirmable")
For formulas please look in Wikipedia and other sources.
$endgroup$
As I understand it, all three want to minimize the false classified data points in your data set. (Logically, if you look for what decision trees are used)
But each of them comes from another side to this problem.
gini impurity wants "better as random"
It compares the "I label random data with random labels" against the labeling after possible split by decision tree (Wish is, that you can split the tree with better outcome than "random random random")
information gain wants small trees
It uses knowledge from information theory. It models the difference between "good" and "bad" split with criteria "simple/small trees preferred". As a result of this, it want to split the data in a way, that the daughters are "pure as possible".
For the chi-square ... I have found two things: CHAID, a (seemingly complex) decision tree technique and the chi square to prune decision trees after their building.
The chi square in general has its roots in biological statistics. It gives a characteristic number how the observed distribution conform with the null hypothesis one have about this distribution. (Biology have to act like this a lot. "I observe something, I search for an explanation, I form a hypothesis, I probe if this is statistical confirmable")
For formulas please look in Wikipedia and other sources.
answered Apr 4 at 10:03
AllerleirauhAllerleirauh
1313
1313
add a comment |
add a comment |
Thanks for contributing an answer to Data Science Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fdatascience.stackexchange.com%2fquestions%2f48560%2fwhat-is-the-differences-in-the-gini-index-chi-square-and-information-gain-spli%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
$begingroup$
Welcome to the site! Look at this post on this site, and this post on medium (explains all three with example).
$endgroup$
– Esmailian
Apr 4 at 11:34