Solving for unknown 3x3 matrix
$begingroup$
Given the following I am trying to solve for $A$:
$$ v_1 = A v_2$$
where $v_1$ is a known $3times 2$ matrix, $A$ is an unknown $3times 3$ matrix and $v_2$ is a known $3times 2$ matrix
Additionally:
$A^{-1} v_1 = v_2$
I am not quite sure how to proceed—is there a way to get a solution for A since all of the given vectors are non square? Any leads would be appreciated!
linear-algebra matrices matrix-equations
New contributor
$endgroup$
add a comment |
$begingroup$
Given the following I am trying to solve for $A$:
$$ v_1 = A v_2$$
where $v_1$ is a known $3times 2$ matrix, $A$ is an unknown $3times 3$ matrix and $v_2$ is a known $3times 2$ matrix
Additionally:
$A^{-1} v_1 = v_2$
I am not quite sure how to proceed—is there a way to get a solution for A since all of the given vectors are non square? Any leads would be appreciated!
linear-algebra matrices matrix-equations
New contributor
$endgroup$
1
$begingroup$
There's a problem - the matrix product of a $3times 3$ and a $2times 3$ in that order doesn't exist. Would you prefer to fix that by reversing the order of multiplication, or by making them $3times 2$ matrices instead? (Formatting note: times is the symbol I used there)
$endgroup$
– jmerry
3 hours ago
$begingroup$
@jmerry whoops! Fixed to 3 $times$ 2.
$endgroup$
– user5826447
2 hours ago
add a comment |
$begingroup$
Given the following I am trying to solve for $A$:
$$ v_1 = A v_2$$
where $v_1$ is a known $3times 2$ matrix, $A$ is an unknown $3times 3$ matrix and $v_2$ is a known $3times 2$ matrix
Additionally:
$A^{-1} v_1 = v_2$
I am not quite sure how to proceed—is there a way to get a solution for A since all of the given vectors are non square? Any leads would be appreciated!
linear-algebra matrices matrix-equations
New contributor
$endgroup$
Given the following I am trying to solve for $A$:
$$ v_1 = A v_2$$
where $v_1$ is a known $3times 2$ matrix, $A$ is an unknown $3times 3$ matrix and $v_2$ is a known $3times 2$ matrix
Additionally:
$A^{-1} v_1 = v_2$
I am not quite sure how to proceed—is there a way to get a solution for A since all of the given vectors are non square? Any leads would be appreciated!
linear-algebra matrices matrix-equations
linear-algebra matrices matrix-equations
New contributor
New contributor
edited 1 hour ago
El Pasta
48815
48815
New contributor
asked 3 hours ago
user5826447user5826447
163
163
New contributor
New contributor
1
$begingroup$
There's a problem - the matrix product of a $3times 3$ and a $2times 3$ in that order doesn't exist. Would you prefer to fix that by reversing the order of multiplication, or by making them $3times 2$ matrices instead? (Formatting note: times is the symbol I used there)
$endgroup$
– jmerry
3 hours ago
$begingroup$
@jmerry whoops! Fixed to 3 $times$ 2.
$endgroup$
– user5826447
2 hours ago
add a comment |
1
$begingroup$
There's a problem - the matrix product of a $3times 3$ and a $2times 3$ in that order doesn't exist. Would you prefer to fix that by reversing the order of multiplication, or by making them $3times 2$ matrices instead? (Formatting note: times is the symbol I used there)
$endgroup$
– jmerry
3 hours ago
$begingroup$
@jmerry whoops! Fixed to 3 $times$ 2.
$endgroup$
– user5826447
2 hours ago
1
1
$begingroup$
There's a problem - the matrix product of a $3times 3$ and a $2times 3$ in that order doesn't exist. Would you prefer to fix that by reversing the order of multiplication, or by making them $3times 2$ matrices instead? (Formatting note: times is the symbol I used there)
$endgroup$
– jmerry
3 hours ago
$begingroup$
There's a problem - the matrix product of a $3times 3$ and a $2times 3$ in that order doesn't exist. Would you prefer to fix that by reversing the order of multiplication, or by making them $3times 2$ matrices instead? (Formatting note: times is the symbol I used there)
$endgroup$
– jmerry
3 hours ago
$begingroup$
@jmerry whoops! Fixed to 3 $times$ 2.
$endgroup$
– user5826447
2 hours ago
$begingroup$
@jmerry whoops! Fixed to 3 $times$ 2.
$endgroup$
– user5826447
2 hours ago
add a comment |
3 Answers
3
active
oldest
votes
$begingroup$
Basically the best we can do is:
$$A=v_1 v_2^+$$
where $v_2^+$ is the Moore-Penrose pseudoinverse.
Note that A will have at most rank 2, although we can make it full rank with an arbitrary vector.
$endgroup$
add a comment |
$begingroup$
Assuming $v_1$ and $v_2$ both have rank $2$, take any $3 times 1$ vector $x_1$ not in the column space of $v_1$, and any vector $3 times 1$ vector $x_2$ not in the column space of $v_2$; let $w_1$ be the $3 times 3$ matrix whose first two columns are $v_1$ and third column is $x_1$, and let $w_2$ be the $3 times 3$ matrix whose first two columns are $v_2$ and third column is $x_2$. Then $A = w_1 w_2^{-1}$ satisfies $A w_2 = w_1$ (and thus $A v_2 = v_1$ and $A x_2 = x_1$), $A^{-1} w_1 = w_2$ (and thus $A^{-1} v_1 = v_2$ and $A^{-1} x_1 = x_2$).
$endgroup$
add a comment |
$begingroup$
Now, the dimensionality says that we should get multiple solutions in general - the problem is finding one. A concept that will get us there? The pseudoinverse. We can't have a full two-sided inverse for the non-square matrix $v_2$, so we find a one-sided inverse $v_2^+$ so that $v_2^+v_2 = I_2$, the $2times 2$ identity. This $v_2^+$, of course, will be a $2times 3$ matrix. Then $A = v_1v_2^+$ will be a solution: $(v_1v_2^+)v_2=v_1(v_2^+v_2) = v_1$.
How do we construct such a matrix? The reference I linked has a bunch of methods, which probably went way over your head. Also, the standard version there has more properties than we need; we're looking for a solution, not the "best" solution. Don't sweat it if we don't end up with exactly the same thing as in the link. So then, let's go elementary. Switch over to the linear transform viewpoint; we have a linear map $T$ from a vector space $U$ of dimension $n$ to another vector space $W$ of dimension $m$. Choose a basis $w_1,w_2,dots$ for the image of $T$ in $W$. For each $w_i$, choose some $u_iin U$ so that $T(u_i)=w_i$. The $u_i$ are a linearly independent set in $U$. Extend both the $u_i$ and the $w_i$ to bases of their respective spaces $U$ and $W$; usually, we will only have to do this for the larger space.
Now, we define $T^+$ as follows: for $ile min(m,n)$, $T(u_i)=w_i$. For $i>min(m,n)$, $T(u_i)$ can be anything - we don't care, but we'll make them zero just to standardize. Then, since $(T^+T)w_i = T^+(Tw_i)=T^+u_i=w_i$ for each $ile m$, $T^+T$ is the identity on $W$.
All right, back to our setup with $3times 2$ matrices. We have $m=2 <3=n$, so we won't have any "extra" vectors to send to zero. There are two length-3 columns of $v_2$, which will be our $w_1$ and $w_2$ (unless they're linearly dependent). What $u_i$ can we use to make $v_2 u_i = w_i$? Well, the standard basis vectors $e_1=begin{pmatrix}1\0end{pmatrix}$ and $e_2=begin{pmatrix}0\1end{pmatrix}$ will do the trick. Then $w_3$ needs to be something linearly independent from $w_1$ and $w_2$ - we make an arbitrary choice here. For something easily constructed, take the cross product.
Now, to construct a matrix $v_2^+$ with $v_2^+ w_1=e_1$, $v_2^+w_2=e_2$, $v_2^+w_3=e_3$. In matrix form, that's
$$v_2^+begin{pmatrix}w_1&w_2&w_3end{pmatrix}=begin{pmatrix}e_1&e_2&0
end{pmatrix}=begin{pmatrix}1&0&0\0&1&0end{pmatrix}$$
$$v_2^+ = begin{pmatrix}1&0&0\0&1&0end{pmatrix}begin{pmatrix}w_1&w_2&w_3end{pmatrix}^{-1}$$
Right-multiplying by that $2times 3$ truncated identity can be summarized simply: throw away the third row.
All right, what if the columns of $v_2$ are linearly dependent? The construction is still there, but now only $w_1$ is actually in the column space of $v_2$. It still satisfies the official pseudoinverse property $v_2v_2^+v_2=v_2$, but we no longer have $v_2^+v_2=I_2$. As such, we can expect not to get solutions, and we won't unless the column space of $v_1$ is contained in the column space of $v_2$. If it is, the pseudoinverse construction will give one.
Now, this choice $A=v_1v_2^+$ is not invertible - it has rank $2$. Can we find an invertible matrix that will also be a solution, and give $A^{-1}v_2=v_1$ as well? We'll need to add some $B$ such that $Bv_2=0$ and $A+B$ is invertible. The rows in $B$ must be orthogonal to the columns of $v_2$, so they'll be multiples of that cross product $w_3^T$. In order to make $A'=A+B$ invertible, we need to avoid the rows of $B$ satisfying the same linear combination as the rows of $A$ - just add to the non-pivot row to make this happen. Then $A'$ is a solution, and its inverse works on the other side automatically.
$endgroup$
add a comment |
Your Answer
StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "69"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
user5826447 is a new contributor. Be nice, and check out our Code of Conduct.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3082731%2fsolving-for-unknown-3x3-matrix%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
3 Answers
3
active
oldest
votes
3 Answers
3
active
oldest
votes
active
oldest
votes
active
oldest
votes
$begingroup$
Basically the best we can do is:
$$A=v_1 v_2^+$$
where $v_2^+$ is the Moore-Penrose pseudoinverse.
Note that A will have at most rank 2, although we can make it full rank with an arbitrary vector.
$endgroup$
add a comment |
$begingroup$
Basically the best we can do is:
$$A=v_1 v_2^+$$
where $v_2^+$ is the Moore-Penrose pseudoinverse.
Note that A will have at most rank 2, although we can make it full rank with an arbitrary vector.
$endgroup$
add a comment |
$begingroup$
Basically the best we can do is:
$$A=v_1 v_2^+$$
where $v_2^+$ is the Moore-Penrose pseudoinverse.
Note that A will have at most rank 2, although we can make it full rank with an arbitrary vector.
$endgroup$
Basically the best we can do is:
$$A=v_1 v_2^+$$
where $v_2^+$ is the Moore-Penrose pseudoinverse.
Note that A will have at most rank 2, although we can make it full rank with an arbitrary vector.
edited 2 hours ago
answered 2 hours ago
I like SerenaI like Serena
4,0871721
4,0871721
add a comment |
add a comment |
$begingroup$
Assuming $v_1$ and $v_2$ both have rank $2$, take any $3 times 1$ vector $x_1$ not in the column space of $v_1$, and any vector $3 times 1$ vector $x_2$ not in the column space of $v_2$; let $w_1$ be the $3 times 3$ matrix whose first two columns are $v_1$ and third column is $x_1$, and let $w_2$ be the $3 times 3$ matrix whose first two columns are $v_2$ and third column is $x_2$. Then $A = w_1 w_2^{-1}$ satisfies $A w_2 = w_1$ (and thus $A v_2 = v_1$ and $A x_2 = x_1$), $A^{-1} w_1 = w_2$ (and thus $A^{-1} v_1 = v_2$ and $A^{-1} x_1 = x_2$).
$endgroup$
add a comment |
$begingroup$
Assuming $v_1$ and $v_2$ both have rank $2$, take any $3 times 1$ vector $x_1$ not in the column space of $v_1$, and any vector $3 times 1$ vector $x_2$ not in the column space of $v_2$; let $w_1$ be the $3 times 3$ matrix whose first two columns are $v_1$ and third column is $x_1$, and let $w_2$ be the $3 times 3$ matrix whose first two columns are $v_2$ and third column is $x_2$. Then $A = w_1 w_2^{-1}$ satisfies $A w_2 = w_1$ (and thus $A v_2 = v_1$ and $A x_2 = x_1$), $A^{-1} w_1 = w_2$ (and thus $A^{-1} v_1 = v_2$ and $A^{-1} x_1 = x_2$).
$endgroup$
add a comment |
$begingroup$
Assuming $v_1$ and $v_2$ both have rank $2$, take any $3 times 1$ vector $x_1$ not in the column space of $v_1$, and any vector $3 times 1$ vector $x_2$ not in the column space of $v_2$; let $w_1$ be the $3 times 3$ matrix whose first two columns are $v_1$ and third column is $x_1$, and let $w_2$ be the $3 times 3$ matrix whose first two columns are $v_2$ and third column is $x_2$. Then $A = w_1 w_2^{-1}$ satisfies $A w_2 = w_1$ (and thus $A v_2 = v_1$ and $A x_2 = x_1$), $A^{-1} w_1 = w_2$ (and thus $A^{-1} v_1 = v_2$ and $A^{-1} x_1 = x_2$).
$endgroup$
Assuming $v_1$ and $v_2$ both have rank $2$, take any $3 times 1$ vector $x_1$ not in the column space of $v_1$, and any vector $3 times 1$ vector $x_2$ not in the column space of $v_2$; let $w_1$ be the $3 times 3$ matrix whose first two columns are $v_1$ and third column is $x_1$, and let $w_2$ be the $3 times 3$ matrix whose first two columns are $v_2$ and third column is $x_2$. Then $A = w_1 w_2^{-1}$ satisfies $A w_2 = w_1$ (and thus $A v_2 = v_1$ and $A x_2 = x_1$), $A^{-1} w_1 = w_2$ (and thus $A^{-1} v_1 = v_2$ and $A^{-1} x_1 = x_2$).
answered 1 hour ago
Robert IsraelRobert Israel
320k23210462
320k23210462
add a comment |
add a comment |
$begingroup$
Now, the dimensionality says that we should get multiple solutions in general - the problem is finding one. A concept that will get us there? The pseudoinverse. We can't have a full two-sided inverse for the non-square matrix $v_2$, so we find a one-sided inverse $v_2^+$ so that $v_2^+v_2 = I_2$, the $2times 2$ identity. This $v_2^+$, of course, will be a $2times 3$ matrix. Then $A = v_1v_2^+$ will be a solution: $(v_1v_2^+)v_2=v_1(v_2^+v_2) = v_1$.
How do we construct such a matrix? The reference I linked has a bunch of methods, which probably went way over your head. Also, the standard version there has more properties than we need; we're looking for a solution, not the "best" solution. Don't sweat it if we don't end up with exactly the same thing as in the link. So then, let's go elementary. Switch over to the linear transform viewpoint; we have a linear map $T$ from a vector space $U$ of dimension $n$ to another vector space $W$ of dimension $m$. Choose a basis $w_1,w_2,dots$ for the image of $T$ in $W$. For each $w_i$, choose some $u_iin U$ so that $T(u_i)=w_i$. The $u_i$ are a linearly independent set in $U$. Extend both the $u_i$ and the $w_i$ to bases of their respective spaces $U$ and $W$; usually, we will only have to do this for the larger space.
Now, we define $T^+$ as follows: for $ile min(m,n)$, $T(u_i)=w_i$. For $i>min(m,n)$, $T(u_i)$ can be anything - we don't care, but we'll make them zero just to standardize. Then, since $(T^+T)w_i = T^+(Tw_i)=T^+u_i=w_i$ for each $ile m$, $T^+T$ is the identity on $W$.
All right, back to our setup with $3times 2$ matrices. We have $m=2 <3=n$, so we won't have any "extra" vectors to send to zero. There are two length-3 columns of $v_2$, which will be our $w_1$ and $w_2$ (unless they're linearly dependent). What $u_i$ can we use to make $v_2 u_i = w_i$? Well, the standard basis vectors $e_1=begin{pmatrix}1\0end{pmatrix}$ and $e_2=begin{pmatrix}0\1end{pmatrix}$ will do the trick. Then $w_3$ needs to be something linearly independent from $w_1$ and $w_2$ - we make an arbitrary choice here. For something easily constructed, take the cross product.
Now, to construct a matrix $v_2^+$ with $v_2^+ w_1=e_1$, $v_2^+w_2=e_2$, $v_2^+w_3=e_3$. In matrix form, that's
$$v_2^+begin{pmatrix}w_1&w_2&w_3end{pmatrix}=begin{pmatrix}e_1&e_2&0
end{pmatrix}=begin{pmatrix}1&0&0\0&1&0end{pmatrix}$$
$$v_2^+ = begin{pmatrix}1&0&0\0&1&0end{pmatrix}begin{pmatrix}w_1&w_2&w_3end{pmatrix}^{-1}$$
Right-multiplying by that $2times 3$ truncated identity can be summarized simply: throw away the third row.
All right, what if the columns of $v_2$ are linearly dependent? The construction is still there, but now only $w_1$ is actually in the column space of $v_2$. It still satisfies the official pseudoinverse property $v_2v_2^+v_2=v_2$, but we no longer have $v_2^+v_2=I_2$. As such, we can expect not to get solutions, and we won't unless the column space of $v_1$ is contained in the column space of $v_2$. If it is, the pseudoinverse construction will give one.
Now, this choice $A=v_1v_2^+$ is not invertible - it has rank $2$. Can we find an invertible matrix that will also be a solution, and give $A^{-1}v_2=v_1$ as well? We'll need to add some $B$ such that $Bv_2=0$ and $A+B$ is invertible. The rows in $B$ must be orthogonal to the columns of $v_2$, so they'll be multiples of that cross product $w_3^T$. In order to make $A'=A+B$ invertible, we need to avoid the rows of $B$ satisfying the same linear combination as the rows of $A$ - just add to the non-pivot row to make this happen. Then $A'$ is a solution, and its inverse works on the other side automatically.
$endgroup$
add a comment |
$begingroup$
Now, the dimensionality says that we should get multiple solutions in general - the problem is finding one. A concept that will get us there? The pseudoinverse. We can't have a full two-sided inverse for the non-square matrix $v_2$, so we find a one-sided inverse $v_2^+$ so that $v_2^+v_2 = I_2$, the $2times 2$ identity. This $v_2^+$, of course, will be a $2times 3$ matrix. Then $A = v_1v_2^+$ will be a solution: $(v_1v_2^+)v_2=v_1(v_2^+v_2) = v_1$.
How do we construct such a matrix? The reference I linked has a bunch of methods, which probably went way over your head. Also, the standard version there has more properties than we need; we're looking for a solution, not the "best" solution. Don't sweat it if we don't end up with exactly the same thing as in the link. So then, let's go elementary. Switch over to the linear transform viewpoint; we have a linear map $T$ from a vector space $U$ of dimension $n$ to another vector space $W$ of dimension $m$. Choose a basis $w_1,w_2,dots$ for the image of $T$ in $W$. For each $w_i$, choose some $u_iin U$ so that $T(u_i)=w_i$. The $u_i$ are a linearly independent set in $U$. Extend both the $u_i$ and the $w_i$ to bases of their respective spaces $U$ and $W$; usually, we will only have to do this for the larger space.
Now, we define $T^+$ as follows: for $ile min(m,n)$, $T(u_i)=w_i$. For $i>min(m,n)$, $T(u_i)$ can be anything - we don't care, but we'll make them zero just to standardize. Then, since $(T^+T)w_i = T^+(Tw_i)=T^+u_i=w_i$ for each $ile m$, $T^+T$ is the identity on $W$.
All right, back to our setup with $3times 2$ matrices. We have $m=2 <3=n$, so we won't have any "extra" vectors to send to zero. There are two length-3 columns of $v_2$, which will be our $w_1$ and $w_2$ (unless they're linearly dependent). What $u_i$ can we use to make $v_2 u_i = w_i$? Well, the standard basis vectors $e_1=begin{pmatrix}1\0end{pmatrix}$ and $e_2=begin{pmatrix}0\1end{pmatrix}$ will do the trick. Then $w_3$ needs to be something linearly independent from $w_1$ and $w_2$ - we make an arbitrary choice here. For something easily constructed, take the cross product.
Now, to construct a matrix $v_2^+$ with $v_2^+ w_1=e_1$, $v_2^+w_2=e_2$, $v_2^+w_3=e_3$. In matrix form, that's
$$v_2^+begin{pmatrix}w_1&w_2&w_3end{pmatrix}=begin{pmatrix}e_1&e_2&0
end{pmatrix}=begin{pmatrix}1&0&0\0&1&0end{pmatrix}$$
$$v_2^+ = begin{pmatrix}1&0&0\0&1&0end{pmatrix}begin{pmatrix}w_1&w_2&w_3end{pmatrix}^{-1}$$
Right-multiplying by that $2times 3$ truncated identity can be summarized simply: throw away the third row.
All right, what if the columns of $v_2$ are linearly dependent? The construction is still there, but now only $w_1$ is actually in the column space of $v_2$. It still satisfies the official pseudoinverse property $v_2v_2^+v_2=v_2$, but we no longer have $v_2^+v_2=I_2$. As such, we can expect not to get solutions, and we won't unless the column space of $v_1$ is contained in the column space of $v_2$. If it is, the pseudoinverse construction will give one.
Now, this choice $A=v_1v_2^+$ is not invertible - it has rank $2$. Can we find an invertible matrix that will also be a solution, and give $A^{-1}v_2=v_1$ as well? We'll need to add some $B$ such that $Bv_2=0$ and $A+B$ is invertible. The rows in $B$ must be orthogonal to the columns of $v_2$, so they'll be multiples of that cross product $w_3^T$. In order to make $A'=A+B$ invertible, we need to avoid the rows of $B$ satisfying the same linear combination as the rows of $A$ - just add to the non-pivot row to make this happen. Then $A'$ is a solution, and its inverse works on the other side automatically.
$endgroup$
add a comment |
$begingroup$
Now, the dimensionality says that we should get multiple solutions in general - the problem is finding one. A concept that will get us there? The pseudoinverse. We can't have a full two-sided inverse for the non-square matrix $v_2$, so we find a one-sided inverse $v_2^+$ so that $v_2^+v_2 = I_2$, the $2times 2$ identity. This $v_2^+$, of course, will be a $2times 3$ matrix. Then $A = v_1v_2^+$ will be a solution: $(v_1v_2^+)v_2=v_1(v_2^+v_2) = v_1$.
How do we construct such a matrix? The reference I linked has a bunch of methods, which probably went way over your head. Also, the standard version there has more properties than we need; we're looking for a solution, not the "best" solution. Don't sweat it if we don't end up with exactly the same thing as in the link. So then, let's go elementary. Switch over to the linear transform viewpoint; we have a linear map $T$ from a vector space $U$ of dimension $n$ to another vector space $W$ of dimension $m$. Choose a basis $w_1,w_2,dots$ for the image of $T$ in $W$. For each $w_i$, choose some $u_iin U$ so that $T(u_i)=w_i$. The $u_i$ are a linearly independent set in $U$. Extend both the $u_i$ and the $w_i$ to bases of their respective spaces $U$ and $W$; usually, we will only have to do this for the larger space.
Now, we define $T^+$ as follows: for $ile min(m,n)$, $T(u_i)=w_i$. For $i>min(m,n)$, $T(u_i)$ can be anything - we don't care, but we'll make them zero just to standardize. Then, since $(T^+T)w_i = T^+(Tw_i)=T^+u_i=w_i$ for each $ile m$, $T^+T$ is the identity on $W$.
All right, back to our setup with $3times 2$ matrices. We have $m=2 <3=n$, so we won't have any "extra" vectors to send to zero. There are two length-3 columns of $v_2$, which will be our $w_1$ and $w_2$ (unless they're linearly dependent). What $u_i$ can we use to make $v_2 u_i = w_i$? Well, the standard basis vectors $e_1=begin{pmatrix}1\0end{pmatrix}$ and $e_2=begin{pmatrix}0\1end{pmatrix}$ will do the trick. Then $w_3$ needs to be something linearly independent from $w_1$ and $w_2$ - we make an arbitrary choice here. For something easily constructed, take the cross product.
Now, to construct a matrix $v_2^+$ with $v_2^+ w_1=e_1$, $v_2^+w_2=e_2$, $v_2^+w_3=e_3$. In matrix form, that's
$$v_2^+begin{pmatrix}w_1&w_2&w_3end{pmatrix}=begin{pmatrix}e_1&e_2&0
end{pmatrix}=begin{pmatrix}1&0&0\0&1&0end{pmatrix}$$
$$v_2^+ = begin{pmatrix}1&0&0\0&1&0end{pmatrix}begin{pmatrix}w_1&w_2&w_3end{pmatrix}^{-1}$$
Right-multiplying by that $2times 3$ truncated identity can be summarized simply: throw away the third row.
All right, what if the columns of $v_2$ are linearly dependent? The construction is still there, but now only $w_1$ is actually in the column space of $v_2$. It still satisfies the official pseudoinverse property $v_2v_2^+v_2=v_2$, but we no longer have $v_2^+v_2=I_2$. As such, we can expect not to get solutions, and we won't unless the column space of $v_1$ is contained in the column space of $v_2$. If it is, the pseudoinverse construction will give one.
Now, this choice $A=v_1v_2^+$ is not invertible - it has rank $2$. Can we find an invertible matrix that will also be a solution, and give $A^{-1}v_2=v_1$ as well? We'll need to add some $B$ such that $Bv_2=0$ and $A+B$ is invertible. The rows in $B$ must be orthogonal to the columns of $v_2$, so they'll be multiples of that cross product $w_3^T$. In order to make $A'=A+B$ invertible, we need to avoid the rows of $B$ satisfying the same linear combination as the rows of $A$ - just add to the non-pivot row to make this happen. Then $A'$ is a solution, and its inverse works on the other side automatically.
$endgroup$
Now, the dimensionality says that we should get multiple solutions in general - the problem is finding one. A concept that will get us there? The pseudoinverse. We can't have a full two-sided inverse for the non-square matrix $v_2$, so we find a one-sided inverse $v_2^+$ so that $v_2^+v_2 = I_2$, the $2times 2$ identity. This $v_2^+$, of course, will be a $2times 3$ matrix. Then $A = v_1v_2^+$ will be a solution: $(v_1v_2^+)v_2=v_1(v_2^+v_2) = v_1$.
How do we construct such a matrix? The reference I linked has a bunch of methods, which probably went way over your head. Also, the standard version there has more properties than we need; we're looking for a solution, not the "best" solution. Don't sweat it if we don't end up with exactly the same thing as in the link. So then, let's go elementary. Switch over to the linear transform viewpoint; we have a linear map $T$ from a vector space $U$ of dimension $n$ to another vector space $W$ of dimension $m$. Choose a basis $w_1,w_2,dots$ for the image of $T$ in $W$. For each $w_i$, choose some $u_iin U$ so that $T(u_i)=w_i$. The $u_i$ are a linearly independent set in $U$. Extend both the $u_i$ and the $w_i$ to bases of their respective spaces $U$ and $W$; usually, we will only have to do this for the larger space.
Now, we define $T^+$ as follows: for $ile min(m,n)$, $T(u_i)=w_i$. For $i>min(m,n)$, $T(u_i)$ can be anything - we don't care, but we'll make them zero just to standardize. Then, since $(T^+T)w_i = T^+(Tw_i)=T^+u_i=w_i$ for each $ile m$, $T^+T$ is the identity on $W$.
All right, back to our setup with $3times 2$ matrices. We have $m=2 <3=n$, so we won't have any "extra" vectors to send to zero. There are two length-3 columns of $v_2$, which will be our $w_1$ and $w_2$ (unless they're linearly dependent). What $u_i$ can we use to make $v_2 u_i = w_i$? Well, the standard basis vectors $e_1=begin{pmatrix}1\0end{pmatrix}$ and $e_2=begin{pmatrix}0\1end{pmatrix}$ will do the trick. Then $w_3$ needs to be something linearly independent from $w_1$ and $w_2$ - we make an arbitrary choice here. For something easily constructed, take the cross product.
Now, to construct a matrix $v_2^+$ with $v_2^+ w_1=e_1$, $v_2^+w_2=e_2$, $v_2^+w_3=e_3$. In matrix form, that's
$$v_2^+begin{pmatrix}w_1&w_2&w_3end{pmatrix}=begin{pmatrix}e_1&e_2&0
end{pmatrix}=begin{pmatrix}1&0&0\0&1&0end{pmatrix}$$
$$v_2^+ = begin{pmatrix}1&0&0\0&1&0end{pmatrix}begin{pmatrix}w_1&w_2&w_3end{pmatrix}^{-1}$$
Right-multiplying by that $2times 3$ truncated identity can be summarized simply: throw away the third row.
All right, what if the columns of $v_2$ are linearly dependent? The construction is still there, but now only $w_1$ is actually in the column space of $v_2$. It still satisfies the official pseudoinverse property $v_2v_2^+v_2=v_2$, but we no longer have $v_2^+v_2=I_2$. As such, we can expect not to get solutions, and we won't unless the column space of $v_1$ is contained in the column space of $v_2$. If it is, the pseudoinverse construction will give one.
Now, this choice $A=v_1v_2^+$ is not invertible - it has rank $2$. Can we find an invertible matrix that will also be a solution, and give $A^{-1}v_2=v_1$ as well? We'll need to add some $B$ such that $Bv_2=0$ and $A+B$ is invertible. The rows in $B$ must be orthogonal to the columns of $v_2$, so they'll be multiples of that cross product $w_3^T$. In order to make $A'=A+B$ invertible, we need to avoid the rows of $B$ satisfying the same linear combination as the rows of $A$ - just add to the non-pivot row to make this happen. Then $A'$ is a solution, and its inverse works on the other side automatically.
answered 1 hour ago
jmerryjmerry
4,404514
4,404514
add a comment |
add a comment |
user5826447 is a new contributor. Be nice, and check out our Code of Conduct.
user5826447 is a new contributor. Be nice, and check out our Code of Conduct.
user5826447 is a new contributor. Be nice, and check out our Code of Conduct.
user5826447 is a new contributor. Be nice, and check out our Code of Conduct.
Thanks for contributing an answer to Mathematics Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3082731%2fsolving-for-unknown-3x3-matrix%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
1
$begingroup$
There's a problem - the matrix product of a $3times 3$ and a $2times 3$ in that order doesn't exist. Would you prefer to fix that by reversing the order of multiplication, or by making them $3times 2$ matrices instead? (Formatting note: times is the symbol I used there)
$endgroup$
– jmerry
3 hours ago
$begingroup$
@jmerry whoops! Fixed to 3 $times$ 2.
$endgroup$
– user5826447
2 hours ago