Gradient descent vs. system of equations
Clash Royale CLAN TAG#URR8PPP
up vote
0
down vote
favorite
Given the matrices $mathbft_Mtimes 1$ and $mathbfQ_Mtimes N$, we want to find $mathbfp_Ntimes 1$ that minimizes $epsilon = ||mathbft - mathbfQmathbfp||_2$.
In order to do so, we could use the gradient descent method. My question is, assuming that we get the best possible result with the gradient descent method, will it be equivalent to clearing $mathbfp$ from the equation $mathbft = mathbfQmathbfp$ as
$mathbfQ^T mathbft = mathbfQ^T mathbfQmathbfp$,
$(mathbfQ^T mathbfQ)^-1 mathbfQ^T mathbft = (mathbfQ^T mathbfQ)^-1 (mathbfQ^T mathbfQ) mathbfp$,
and so
$mathbfp = (mathbfQ^T mathbfQ)^-1 mathbfQ^T mathbft$?
I see that by premultiplying by $mathbfQ^T$ we are 'colapsing' our system of equations (M equations, N unknowns) to N equations with N unknowns, but I can't see how this will minimize $epsilon$ (in case it does).
linear-algebra gradient-descent
add a comment |Â
up vote
0
down vote
favorite
Given the matrices $mathbft_Mtimes 1$ and $mathbfQ_Mtimes N$, we want to find $mathbfp_Ntimes 1$ that minimizes $epsilon = ||mathbft - mathbfQmathbfp||_2$.
In order to do so, we could use the gradient descent method. My question is, assuming that we get the best possible result with the gradient descent method, will it be equivalent to clearing $mathbfp$ from the equation $mathbft = mathbfQmathbfp$ as
$mathbfQ^T mathbft = mathbfQ^T mathbfQmathbfp$,
$(mathbfQ^T mathbfQ)^-1 mathbfQ^T mathbft = (mathbfQ^T mathbfQ)^-1 (mathbfQ^T mathbfQ) mathbfp$,
and so
$mathbfp = (mathbfQ^T mathbfQ)^-1 mathbfQ^T mathbft$?
I see that by premultiplying by $mathbfQ^T$ we are 'colapsing' our system of equations (M equations, N unknowns) to N equations with N unknowns, but I can't see how this will minimize $epsilon$ (in case it does).
linear-algebra gradient-descent
Minimizing the error is equivalent to minimizing the the squared error, which is $epsilon^2 = |t-Qp|^2 = t^Tt-2t^TQp+p^TQ^TQp$. The minimum is attained when its gradient, $2Q^TQp-2Q^Tt$, equals zero, which occurs when $p=(Q^TQ)^-1Q^Tt$. Thus the two approaches are equivalent.
â Rahul
Sep 7 at 9:20
add a comment |Â
up vote
0
down vote
favorite
up vote
0
down vote
favorite
Given the matrices $mathbft_Mtimes 1$ and $mathbfQ_Mtimes N$, we want to find $mathbfp_Ntimes 1$ that minimizes $epsilon = ||mathbft - mathbfQmathbfp||_2$.
In order to do so, we could use the gradient descent method. My question is, assuming that we get the best possible result with the gradient descent method, will it be equivalent to clearing $mathbfp$ from the equation $mathbft = mathbfQmathbfp$ as
$mathbfQ^T mathbft = mathbfQ^T mathbfQmathbfp$,
$(mathbfQ^T mathbfQ)^-1 mathbfQ^T mathbft = (mathbfQ^T mathbfQ)^-1 (mathbfQ^T mathbfQ) mathbfp$,
and so
$mathbfp = (mathbfQ^T mathbfQ)^-1 mathbfQ^T mathbft$?
I see that by premultiplying by $mathbfQ^T$ we are 'colapsing' our system of equations (M equations, N unknowns) to N equations with N unknowns, but I can't see how this will minimize $epsilon$ (in case it does).
linear-algebra gradient-descent
Given the matrices $mathbft_Mtimes 1$ and $mathbfQ_Mtimes N$, we want to find $mathbfp_Ntimes 1$ that minimizes $epsilon = ||mathbft - mathbfQmathbfp||_2$.
In order to do so, we could use the gradient descent method. My question is, assuming that we get the best possible result with the gradient descent method, will it be equivalent to clearing $mathbfp$ from the equation $mathbft = mathbfQmathbfp$ as
$mathbfQ^T mathbft = mathbfQ^T mathbfQmathbfp$,
$(mathbfQ^T mathbfQ)^-1 mathbfQ^T mathbft = (mathbfQ^T mathbfQ)^-1 (mathbfQ^T mathbfQ) mathbfp$,
and so
$mathbfp = (mathbfQ^T mathbfQ)^-1 mathbfQ^T mathbft$?
I see that by premultiplying by $mathbfQ^T$ we are 'colapsing' our system of equations (M equations, N unknowns) to N equations with N unknowns, but I can't see how this will minimize $epsilon$ (in case it does).
linear-algebra gradient-descent
linear-algebra gradient-descent
asked Sep 7 at 9:12
Carlos Navarro Astiasarán
454
454
Minimizing the error is equivalent to minimizing the the squared error, which is $epsilon^2 = |t-Qp|^2 = t^Tt-2t^TQp+p^TQ^TQp$. The minimum is attained when its gradient, $2Q^TQp-2Q^Tt$, equals zero, which occurs when $p=(Q^TQ)^-1Q^Tt$. Thus the two approaches are equivalent.
â Rahul
Sep 7 at 9:20
add a comment |Â
Minimizing the error is equivalent to minimizing the the squared error, which is $epsilon^2 = |t-Qp|^2 = t^Tt-2t^TQp+p^TQ^TQp$. The minimum is attained when its gradient, $2Q^TQp-2Q^Tt$, equals zero, which occurs when $p=(Q^TQ)^-1Q^Tt$. Thus the two approaches are equivalent.
â Rahul
Sep 7 at 9:20
Minimizing the error is equivalent to minimizing the the squared error, which is $epsilon^2 = |t-Qp|^2 = t^Tt-2t^TQp+p^TQ^TQp$. The minimum is attained when its gradient, $2Q^TQp-2Q^Tt$, equals zero, which occurs when $p=(Q^TQ)^-1Q^Tt$. Thus the two approaches are equivalent.
â Rahul
Sep 7 at 9:20
Minimizing the error is equivalent to minimizing the the squared error, which is $epsilon^2 = |t-Qp|^2 = t^Tt-2t^TQp+p^TQ^TQp$. The minimum is attained when its gradient, $2Q^TQp-2Q^Tt$, equals zero, which occurs when $p=(Q^TQ)^-1Q^Tt$. Thus the two approaches are equivalent.
â Rahul
Sep 7 at 9:20
add a comment |Â
active
oldest
votes
active
oldest
votes
active
oldest
votes
active
oldest
votes
active
oldest
votes
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f2908425%2fgradient-descent-vs-system-of-equations%23new-answer', 'question_page');
);
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Minimizing the error is equivalent to minimizing the the squared error, which is $epsilon^2 = |t-Qp|^2 = t^Tt-2t^TQp+p^TQ^TQp$. The minimum is attained when its gradient, $2Q^TQp-2Q^Tt$, equals zero, which occurs when $p=(Q^TQ)^-1Q^Tt$. Thus the two approaches are equivalent.
â Rahul
Sep 7 at 9:20