Linear algebra with matrices not only for $mathbbR^n$?
Clash Royale CLAN TAG#URR8PPP
up vote
1
down vote
favorite
I haven't taken a course on Linear Algebra, but I kind of used the concepts of row/column vectors, matrices a lot. But somehow only with respect with the $mathbbR^n$ vector space. When talking about other vector spaces, like a function space, do matrices (and implicitly column/row vectors) have that great impact?
My intuition says yes, because:
In $mathbbR^3$:
Let $B = e_1=(1,0,0),e_2=(0,1,0),e_3=(0,0,1)$ be a basis of $mathbbR^3$, where (a,b,c) is a 3-tuple
Then:
$[(1,2,3)]_B = beginbmatrix1\2\3endbmatrix$In the function space of polynomials:
Let $B = 1,x,x^2$ be a basis
Then:
$[1+2x+3x^2]_B = beginbmatrix1\2\3endbmatrix$.
Are these considerations useful or good to be kept in mind when working with linear algebra? I mean, is it important not to forget about matrices when turning from $mathbbR^n$ to another vector space?
linear-algebra matrices vector-spaces
add a comment |Â
up vote
1
down vote
favorite
I haven't taken a course on Linear Algebra, but I kind of used the concepts of row/column vectors, matrices a lot. But somehow only with respect with the $mathbbR^n$ vector space. When talking about other vector spaces, like a function space, do matrices (and implicitly column/row vectors) have that great impact?
My intuition says yes, because:
In $mathbbR^3$:
Let $B = e_1=(1,0,0),e_2=(0,1,0),e_3=(0,0,1)$ be a basis of $mathbbR^3$, where (a,b,c) is a 3-tuple
Then:
$[(1,2,3)]_B = beginbmatrix1\2\3endbmatrix$In the function space of polynomials:
Let $B = 1,x,x^2$ be a basis
Then:
$[1+2x+3x^2]_B = beginbmatrix1\2\3endbmatrix$.
Are these considerations useful or good to be kept in mind when working with linear algebra? I mean, is it important not to forget about matrices when turning from $mathbbR^n$ to another vector space?
linear-algebra matrices vector-spaces
2
Yes, it is very important. As you note, using bases and hence coordinate vectors in the general vector space, a lot of questions can be answered using the theory developed for $R^n $ or even $n $-tuples over a generic field, and matrices over that field
â AnyAD
Jun 15 at 15:22
6
Matrices can be used for any finite dimensional vector space over any field ($mathbf Q, mathbf C$ for instance), and even for any finitely generated free module over any commutative ring (modules are the structure corresponding to vector spaces when working over a ring instead of a field).
â Bernard
Jun 15 at 15:22
@Bernard Curiosity: you can actually drop the requirement that $R$ is commutative, but the opposite ring $R^rm op$ starts popping up. There are a couple of results in this direction in Adkins & Weintraub's Algebra: An Approach via Module Theory, if I recall correctly.
â Ivo Terek
Jun 15 at 15:31
Also, that importance to matrices is given because there is the concept of basis, right?
â user_anon
Jun 15 at 15:36
add a comment |Â
up vote
1
down vote
favorite
up vote
1
down vote
favorite
I haven't taken a course on Linear Algebra, but I kind of used the concepts of row/column vectors, matrices a lot. But somehow only with respect with the $mathbbR^n$ vector space. When talking about other vector spaces, like a function space, do matrices (and implicitly column/row vectors) have that great impact?
My intuition says yes, because:
In $mathbbR^3$:
Let $B = e_1=(1,0,0),e_2=(0,1,0),e_3=(0,0,1)$ be a basis of $mathbbR^3$, where (a,b,c) is a 3-tuple
Then:
$[(1,2,3)]_B = beginbmatrix1\2\3endbmatrix$In the function space of polynomials:
Let $B = 1,x,x^2$ be a basis
Then:
$[1+2x+3x^2]_B = beginbmatrix1\2\3endbmatrix$.
Are these considerations useful or good to be kept in mind when working with linear algebra? I mean, is it important not to forget about matrices when turning from $mathbbR^n$ to another vector space?
linear-algebra matrices vector-spaces
I haven't taken a course on Linear Algebra, but I kind of used the concepts of row/column vectors, matrices a lot. But somehow only with respect with the $mathbbR^n$ vector space. When talking about other vector spaces, like a function space, do matrices (and implicitly column/row vectors) have that great impact?
My intuition says yes, because:
In $mathbbR^3$:
Let $B = e_1=(1,0,0),e_2=(0,1,0),e_3=(0,0,1)$ be a basis of $mathbbR^3$, where (a,b,c) is a 3-tuple
Then:
$[(1,2,3)]_B = beginbmatrix1\2\3endbmatrix$In the function space of polynomials:
Let $B = 1,x,x^2$ be a basis
Then:
$[1+2x+3x^2]_B = beginbmatrix1\2\3endbmatrix$.
Are these considerations useful or good to be kept in mind when working with linear algebra? I mean, is it important not to forget about matrices when turning from $mathbbR^n$ to another vector space?
linear-algebra matrices vector-spaces
edited Aug 13 at 10:24
asked Jun 15 at 15:17
user_anon
508
508
2
Yes, it is very important. As you note, using bases and hence coordinate vectors in the general vector space, a lot of questions can be answered using the theory developed for $R^n $ or even $n $-tuples over a generic field, and matrices over that field
â AnyAD
Jun 15 at 15:22
6
Matrices can be used for any finite dimensional vector space over any field ($mathbf Q, mathbf C$ for instance), and even for any finitely generated free module over any commutative ring (modules are the structure corresponding to vector spaces when working over a ring instead of a field).
â Bernard
Jun 15 at 15:22
@Bernard Curiosity: you can actually drop the requirement that $R$ is commutative, but the opposite ring $R^rm op$ starts popping up. There are a couple of results in this direction in Adkins & Weintraub's Algebra: An Approach via Module Theory, if I recall correctly.
â Ivo Terek
Jun 15 at 15:31
Also, that importance to matrices is given because there is the concept of basis, right?
â user_anon
Jun 15 at 15:36
add a comment |Â
2
Yes, it is very important. As you note, using bases and hence coordinate vectors in the general vector space, a lot of questions can be answered using the theory developed for $R^n $ or even $n $-tuples over a generic field, and matrices over that field
â AnyAD
Jun 15 at 15:22
6
Matrices can be used for any finite dimensional vector space over any field ($mathbf Q, mathbf C$ for instance), and even for any finitely generated free module over any commutative ring (modules are the structure corresponding to vector spaces when working over a ring instead of a field).
â Bernard
Jun 15 at 15:22
@Bernard Curiosity: you can actually drop the requirement that $R$ is commutative, but the opposite ring $R^rm op$ starts popping up. There are a couple of results in this direction in Adkins & Weintraub's Algebra: An Approach via Module Theory, if I recall correctly.
â Ivo Terek
Jun 15 at 15:31
Also, that importance to matrices is given because there is the concept of basis, right?
â user_anon
Jun 15 at 15:36
2
2
Yes, it is very important. As you note, using bases and hence coordinate vectors in the general vector space, a lot of questions can be answered using the theory developed for $R^n $ or even $n $-tuples over a generic field, and matrices over that field
â AnyAD
Jun 15 at 15:22
Yes, it is very important. As you note, using bases and hence coordinate vectors in the general vector space, a lot of questions can be answered using the theory developed for $R^n $ or even $n $-tuples over a generic field, and matrices over that field
â AnyAD
Jun 15 at 15:22
6
6
Matrices can be used for any finite dimensional vector space over any field ($mathbf Q, mathbf C$ for instance), and even for any finitely generated free module over any commutative ring (modules are the structure corresponding to vector spaces when working over a ring instead of a field).
â Bernard
Jun 15 at 15:22
Matrices can be used for any finite dimensional vector space over any field ($mathbf Q, mathbf C$ for instance), and even for any finitely generated free module over any commutative ring (modules are the structure corresponding to vector spaces when working over a ring instead of a field).
â Bernard
Jun 15 at 15:22
@Bernard Curiosity: you can actually drop the requirement that $R$ is commutative, but the opposite ring $R^rm op$ starts popping up. There are a couple of results in this direction in Adkins & Weintraub's Algebra: An Approach via Module Theory, if I recall correctly.
â Ivo Terek
Jun 15 at 15:31
@Bernard Curiosity: you can actually drop the requirement that $R$ is commutative, but the opposite ring $R^rm op$ starts popping up. There are a couple of results in this direction in Adkins & Weintraub's Algebra: An Approach via Module Theory, if I recall correctly.
â Ivo Terek
Jun 15 at 15:31
Also, that importance to matrices is given because there is the concept of basis, right?
â user_anon
Jun 15 at 15:36
Also, that importance to matrices is given because there is the concept of basis, right?
â user_anon
Jun 15 at 15:36
add a comment |Â
active
oldest
votes
active
oldest
votes
active
oldest
votes
active
oldest
votes
active
oldest
votes
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f2820802%2flinear-algebra-with-matrices-not-only-for-mathbbrn%23new-answer', 'question_page');
);
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
2
Yes, it is very important. As you note, using bases and hence coordinate vectors in the general vector space, a lot of questions can be answered using the theory developed for $R^n $ or even $n $-tuples over a generic field, and matrices over that field
â AnyAD
Jun 15 at 15:22
6
Matrices can be used for any finite dimensional vector space over any field ($mathbf Q, mathbf C$ for instance), and even for any finitely generated free module over any commutative ring (modules are the structure corresponding to vector spaces when working over a ring instead of a field).
â Bernard
Jun 15 at 15:22
@Bernard Curiosity: you can actually drop the requirement that $R$ is commutative, but the opposite ring $R^rm op$ starts popping up. There are a couple of results in this direction in Adkins & Weintraub's Algebra: An Approach via Module Theory, if I recall correctly.
â Ivo Terek
Jun 15 at 15:31
Also, that importance to matrices is given because there is the concept of basis, right?
â user_anon
Jun 15 at 15:36