Covariance as inner productExtrapolating GLM coefficients for year a product was sold into future years?Identity covariance matrix, decorrelated data?Handling covariance matrices of changing sizeNeural networks: which cost function to use?Advance Methods of Understanding Significance of Customer BehaviorsA statistic for testing if $mu$ which is known to be in subspace $H$, is also in subspace $H_0subseteq H$How do I convert an L2 norm to a probability?Fast way of computing covariance matrix of nonstationary kernel in Python
Stark VS Thanos
Phrase for the opposite of "foolproof"
Illegal assignment from SObject to Contact
What's the metal clinking sound at the end of credits in Avengers: Endgame?
Multiple options for Pseudonyms
Will tsunami waves travel forever if there was no land?
Modify locally tikzset
Is it possible to measure lightning discharges as Nikola Tesla?
Confusion about capacitors
"ne paelici suspectaretur" (Tacitus)
Reverse the word in a string with the same order in javascript
Has any spacecraft ever had the ability to directly communicate with civilian air traffic control?
How to stop co-workers from teasing me because I know Russian?
How to creep the reader out with what seems like a normal person?
In the time of the mishna, were there Jewish cities without courts?
What is the difference between `a[bc]d` (brackets) and `ab,cd` (braces)?
Why was Germany not as successful as other Europeans in establishing overseas colonies?
Python "triplet" dictionary?
Can a creature tell when it has been affected by a Divination wizard's Portent?
Why does Bran Stark feel that Jon Snow "needs to know" about his lineage?
Does the EU Common Fisheries Policy cover British Overseas Territories?
Is it cheaper to drop cargo drop than to land it?
Options leqno, reqno for documentclass or exist another option?
Pawn Sacrifice Justification
Covariance as inner product
Extrapolating GLM coefficients for year a product was sold into future years?Identity covariance matrix, decorrelated data?Handling covariance matrices of changing sizeNeural networks: which cost function to use?Advance Methods of Understanding Significance of Customer BehaviorsA statistic for testing if $mu$ which is known to be in subspace $H$, is also in subspace $H_0subseteq H$How do I convert an L2 norm to a probability?Fast way of computing covariance matrix of nonstationary kernel in Python
$begingroup$
Why is covariance considered as inner product if there is no projection of one vector onto another?
Right now I perceive this as just a multiplication of $x$ segment of vector($x_i - barx$) and $y$ segment($y_i - bary$) of the same vector in order to understand direction of relationship.
statistics data-analysis
$endgroup$
add a comment |
$begingroup$
Why is covariance considered as inner product if there is no projection of one vector onto another?
Right now I perceive this as just a multiplication of $x$ segment of vector($x_i - barx$) and $y$ segment($y_i - bary$) of the same vector in order to understand direction of relationship.
statistics data-analysis
$endgroup$
add a comment |
$begingroup$
Why is covariance considered as inner product if there is no projection of one vector onto another?
Right now I perceive this as just a multiplication of $x$ segment of vector($x_i - barx$) and $y$ segment($y_i - bary$) of the same vector in order to understand direction of relationship.
statistics data-analysis
$endgroup$
Why is covariance considered as inner product if there is no projection of one vector onto another?
Right now I perceive this as just a multiplication of $x$ segment of vector($x_i - barx$) and $y$ segment($y_i - bary$) of the same vector in order to understand direction of relationship.
statistics data-analysis
statistics data-analysis
edited Apr 8 at 16:11
Stephen Rauch♦
1,53551330
1,53551330
asked Apr 8 at 15:40
user641597user641597
434
434
add a comment |
add a comment |
1 Answer
1
active
oldest
votes
$begingroup$
Definition
A inner product (AKA dot product and scalar product) can be define on two vectors $mathbfx$ and $mathbfy$ $in mathcalR^n $ as
$$ mathbfx.x^T = <mathbfx,mathbfy>_mathcalR^n=<mathbfy,mathbfx>_mathcalR^n = sum_i=1^n x_i times y_i $$
The inner product can be seem as the length of the projection of a vector into another and it is widely used as a similarity measure between two vectors.
Also the inner product have the following properties:
Commutative or symmetric
Distributive (over vector addition)- Bilinear
Positive-definite: i.e $mathbfx.x^T > 0,forall mathbfx $
The covariance of two random variables $X$ and $Y$ can be defined as
$$ E[(X-E[X]) times (Y - E[Y])] $$
the covariance holds the properties of been commutative, bilinear and positive-definite.
These properties imply that the covariance is an Inner Product in a vector space, more specifically the Quotient Space.
Association with the kernel trick
If you are familiar with Support Vector Machines you probably familiar with the Kernel Trick where you implicitly compute the inner product of two vectors into a mapped space, called feature space. Without performing the mapping you can compute the inner product into even a possibly infinite dimensional space given that this mapping.
To perform that inner product, you need to find a function, known as kernel functions, that can perform this inner product without explicitly mapping the vectors.
For a kernel function to exist it needs to have the following atributes:
- It needs to be symmetric
- It needs to be positive-definite
That is sufficient and necessary to a function $kappa(mathbfx,y)$ to be considered a inner product in an arbitrary vector space $mathcalH$.
As the covariance, comply to this definition it is a Kernel Function and consequentially it is an Inner Product in a Vector Space.
$endgroup$
add a comment |
Your Answer
StackExchange.ready(function()
var channelOptions =
tags: "".split(" "),
id: "557"
;
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function()
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled)
StackExchange.using("snippets", function()
createEditor();
);
else
createEditor();
);
function createEditor()
StackExchange.prepareEditor(
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: false,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: null,
bindNavPrevention: true,
postfix: "",
imageUploader:
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
,
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
);
);
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fdatascience.stackexchange.com%2fquestions%2f48885%2fcovariance-as-inner-product%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
1 Answer
1
active
oldest
votes
1 Answer
1
active
oldest
votes
active
oldest
votes
active
oldest
votes
$begingroup$
Definition
A inner product (AKA dot product and scalar product) can be define on two vectors $mathbfx$ and $mathbfy$ $in mathcalR^n $ as
$$ mathbfx.x^T = <mathbfx,mathbfy>_mathcalR^n=<mathbfy,mathbfx>_mathcalR^n = sum_i=1^n x_i times y_i $$
The inner product can be seem as the length of the projection of a vector into another and it is widely used as a similarity measure between two vectors.
Also the inner product have the following properties:
Commutative or symmetric
Distributive (over vector addition)- Bilinear
Positive-definite: i.e $mathbfx.x^T > 0,forall mathbfx $
The covariance of two random variables $X$ and $Y$ can be defined as
$$ E[(X-E[X]) times (Y - E[Y])] $$
the covariance holds the properties of been commutative, bilinear and positive-definite.
These properties imply that the covariance is an Inner Product in a vector space, more specifically the Quotient Space.
Association with the kernel trick
If you are familiar with Support Vector Machines you probably familiar with the Kernel Trick where you implicitly compute the inner product of two vectors into a mapped space, called feature space. Without performing the mapping you can compute the inner product into even a possibly infinite dimensional space given that this mapping.
To perform that inner product, you need to find a function, known as kernel functions, that can perform this inner product without explicitly mapping the vectors.
For a kernel function to exist it needs to have the following atributes:
- It needs to be symmetric
- It needs to be positive-definite
That is sufficient and necessary to a function $kappa(mathbfx,y)$ to be considered a inner product in an arbitrary vector space $mathcalH$.
As the covariance, comply to this definition it is a Kernel Function and consequentially it is an Inner Product in a Vector Space.
$endgroup$
add a comment |
$begingroup$
Definition
A inner product (AKA dot product and scalar product) can be define on two vectors $mathbfx$ and $mathbfy$ $in mathcalR^n $ as
$$ mathbfx.x^T = <mathbfx,mathbfy>_mathcalR^n=<mathbfy,mathbfx>_mathcalR^n = sum_i=1^n x_i times y_i $$
The inner product can be seem as the length of the projection of a vector into another and it is widely used as a similarity measure between two vectors.
Also the inner product have the following properties:
Commutative or symmetric
Distributive (over vector addition)- Bilinear
Positive-definite: i.e $mathbfx.x^T > 0,forall mathbfx $
The covariance of two random variables $X$ and $Y$ can be defined as
$$ E[(X-E[X]) times (Y - E[Y])] $$
the covariance holds the properties of been commutative, bilinear and positive-definite.
These properties imply that the covariance is an Inner Product in a vector space, more specifically the Quotient Space.
Association with the kernel trick
If you are familiar with Support Vector Machines you probably familiar with the Kernel Trick where you implicitly compute the inner product of two vectors into a mapped space, called feature space. Without performing the mapping you can compute the inner product into even a possibly infinite dimensional space given that this mapping.
To perform that inner product, you need to find a function, known as kernel functions, that can perform this inner product without explicitly mapping the vectors.
For a kernel function to exist it needs to have the following atributes:
- It needs to be symmetric
- It needs to be positive-definite
That is sufficient and necessary to a function $kappa(mathbfx,y)$ to be considered a inner product in an arbitrary vector space $mathcalH$.
As the covariance, comply to this definition it is a Kernel Function and consequentially it is an Inner Product in a Vector Space.
$endgroup$
add a comment |
$begingroup$
Definition
A inner product (AKA dot product and scalar product) can be define on two vectors $mathbfx$ and $mathbfy$ $in mathcalR^n $ as
$$ mathbfx.x^T = <mathbfx,mathbfy>_mathcalR^n=<mathbfy,mathbfx>_mathcalR^n = sum_i=1^n x_i times y_i $$
The inner product can be seem as the length of the projection of a vector into another and it is widely used as a similarity measure between two vectors.
Also the inner product have the following properties:
Commutative or symmetric
Distributive (over vector addition)- Bilinear
Positive-definite: i.e $mathbfx.x^T > 0,forall mathbfx $
The covariance of two random variables $X$ and $Y$ can be defined as
$$ E[(X-E[X]) times (Y - E[Y])] $$
the covariance holds the properties of been commutative, bilinear and positive-definite.
These properties imply that the covariance is an Inner Product in a vector space, more specifically the Quotient Space.
Association with the kernel trick
If you are familiar with Support Vector Machines you probably familiar with the Kernel Trick where you implicitly compute the inner product of two vectors into a mapped space, called feature space. Without performing the mapping you can compute the inner product into even a possibly infinite dimensional space given that this mapping.
To perform that inner product, you need to find a function, known as kernel functions, that can perform this inner product without explicitly mapping the vectors.
For a kernel function to exist it needs to have the following atributes:
- It needs to be symmetric
- It needs to be positive-definite
That is sufficient and necessary to a function $kappa(mathbfx,y)$ to be considered a inner product in an arbitrary vector space $mathcalH$.
As the covariance, comply to this definition it is a Kernel Function and consequentially it is an Inner Product in a Vector Space.
$endgroup$
Definition
A inner product (AKA dot product and scalar product) can be define on two vectors $mathbfx$ and $mathbfy$ $in mathcalR^n $ as
$$ mathbfx.x^T = <mathbfx,mathbfy>_mathcalR^n=<mathbfy,mathbfx>_mathcalR^n = sum_i=1^n x_i times y_i $$
The inner product can be seem as the length of the projection of a vector into another and it is widely used as a similarity measure between two vectors.
Also the inner product have the following properties:
Commutative or symmetric
Distributive (over vector addition)- Bilinear
Positive-definite: i.e $mathbfx.x^T > 0,forall mathbfx $
The covariance of two random variables $X$ and $Y$ can be defined as
$$ E[(X-E[X]) times (Y - E[Y])] $$
the covariance holds the properties of been commutative, bilinear and positive-definite.
These properties imply that the covariance is an Inner Product in a vector space, more specifically the Quotient Space.
Association with the kernel trick
If you are familiar with Support Vector Machines you probably familiar with the Kernel Trick where you implicitly compute the inner product of two vectors into a mapped space, called feature space. Without performing the mapping you can compute the inner product into even a possibly infinite dimensional space given that this mapping.
To perform that inner product, you need to find a function, known as kernel functions, that can perform this inner product without explicitly mapping the vectors.
For a kernel function to exist it needs to have the following atributes:
- It needs to be symmetric
- It needs to be positive-definite
That is sufficient and necessary to a function $kappa(mathbfx,y)$ to be considered a inner product in an arbitrary vector space $mathcalH$.
As the covariance, comply to this definition it is a Kernel Function and consequentially it is an Inner Product in a Vector Space.
answered Apr 9 at 23:28
Pedro Henrique MonfortePedro Henrique Monforte
569219
569219
add a comment |
add a comment |
Thanks for contributing an answer to Data Science Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fdatascience.stackexchange.com%2fquestions%2f48885%2fcovariance-as-inner-product%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown