Talk:Transpose

Note
Another note -- should mention that the generalization of the notion of transpose to complex matrices is to make element (i,j) equal to the conjugate of the (j,i) element. I agree with Chris W that the notation A' should be mentioned as an alternative. Wile E. Heresiarch 08:39, 11 Mar 2004 (UTC)

Differentiation of transposed matrices
I was trying to understand what the derivative of a transposed matrix is with respect to that matrix? So something like $$ \frac{\mathrm{d}A \Theta}{\mathrm{d}\Theta}$$ where both are matrices.

yanneman 13:26, 20 November 2006 (UTC)

Transpose of linear maps
In the Transpose of linear maps section, the article had read:
 * If f : V&rarr;W is a linear map between vector spaces V and W with dual spaces W* and V*, we define the transpose of f to be the linear map tf : W*&rarr;V* with

{}^t f (\phi) = \phi\circ f\,$$  for every $$\ \phi$$ in W*.

An anonymous user changed that to
 * If f : V&rarr;W is a linear map between vector spaces V and W with dual spaces W* and V*, we define the transpose of f to be the linear map fT : W*&rarr;V* with

f (\phi)^{\mathrm{T}} = \phi\circ f\,$$  for every $$\ \phi$$ in W*.

Now, I think that's confusing (it's f that's transposed, not f(&phi;)), and it would be better to write

f^{\mathrm{T}}(\phi) = \phi\circ f\,$$  for every $$\ \phi$$ in W*. But even that seems ambiguous; is this "transpose f" or "f to the T power"? But I'm unfamiliar with this notation, and I certainly don't have the same objection to T used with matrices. So, are there any experts who could weigh in with the most common usage in this area? (I'm not proposing to change the T notation used in the earlier sections, just to keep the t notation in this one section.) --Quuxplusone 21:14, 8 August 2005 (UTC)


 * Putting the T in front sounds like a worthy experiment. In general though, you have to realise that mathematics notation is never anywhere near as rigorous as people would like to think. A classic example is -1 designates both the inverse and the reciprocal. Cesiumfrog (talk) 04:52, 11 May 2010 (UTC)


 * What about using $$f^\top(\phi)$$? —Ben FrantzDale (talk) 12:40, 11 May 2010 (UTC)

hermitian transpose?
hermitian transpose = conjugate transpose? --Moala 09:23, 20 December 2005 (UTC)

Transpose on tensors...
I'm confused. It seems like much of linear algebra glosses over the meaning of transposition and simply uses it as a mechanism for manipulating numbers, for example, defining the norm of v as $$\|v\| = \sqrt{v^\top v}$$.

In some linear-algebra topics, however, it appears that column and row vectors have different meanings (that appear to have something to do with covariance and contravariance of vectors). The transpose of a column vector, c, gives you a row vector -- a vector in the dual space of c. I think the idea is that column vectors would be indexed with raised indices and row vectors with lowered indices with tensors.

Here's my confusion: If row vectors and column vectors are in distinct spaces (and they certainly are in that you can't just add them), then taking the transpose of a vector isn't just some notational convenience, it is an application of a nontrivial function, $$f(v)=v^\top:V\rightarrow V^*$$. To do something like this in general, we can use any bilinear form, but that involves more structure than just

So:
 * 1) Is it correct that there are are two things going on here: (1) using transpose for numerical convenience and (2) using rows versus columns to for indicateng co- versus contravariance?
 * 2) Isn't the conventional Euclidean metric defined with a contravariant metric tensor: $$\|v\|=\sqrt{g_{ij}v^iv^j}$$? Doesn't that not involve any transposition in that both vs have raised indices?

Thanks. —Ben FrantzDale (talk) 05:00, 11 November 2009 (UTC)

Transpose and tensors
I posed a question on Talk:Transpose that didn't get any responses there. Perhaps this is a better audience since it's a bit of an essoteric question for such an elementary topic; Here's the question again:


 * I'm confused. It seems like much of linear algebra glosses over the meaning of transposition and simply uses it as a mechanism for manipulating numbers, for example, defining the norm of v as $$\|v\| = \sqrt{v^\top v}$$.


 * In some linear-algebra topics, however, it appears that column and row vectors have different meanings (that appear to have something to do with covariance and contravariance of vectors). In that context, the transpose of a column vector, c, gives you a row vector -- a vector in the dual space of c. I think the idea is that column vectors would be indexed with raised indices and row vectors with lowered indices with tensors.


 * Here's my confusion: If row vectors and column vectors are in distinct spaces (and they certainly, even in elementary linear algebra in that you can't just add a column to a row vector because they have different shapes), then taking the transpose of a vector isn't just some notational convenience, it is an application of a nontrivial function, $$f(v)=v^\top:V\rightarrow V^*$$. To do something like this in general, we can use any bilinear form, but that involves more structure than just


 * So:
 * Is it correct that there are are two things going on here: (1) using transpose for numerical convenience and (2) using rows versus columns to for indicateng co- versus contravariance?
 * Isn't the conventional Euclidean metric defined with a contravariant metric tensor: $$\|v\|=\sqrt{g_{ij}v^iv^j}$$? Doesn't that not involve any transposition in that both vs have raised indices?


 * Thanks. —Ben FrantzDale (talk) 14:16, 23 November 2009 (UTC)


 * I guess it depends on how we define vectors. If we consider a vector as just being an n by m matrix with either n=1 or m=1, then transposition is just what it is with any other matrix - a map from the space of n by m matrices to the space of m by n matrices. --Tango (talk) 14:38, 23 November 2009 (UTC)
 * Sure. I'm asking because I get the sense that there are some unwritten ruels going on. At one extreme is the purely-mechanical notion of tranpose that you describe, which I'm happy with. In that context, transpose is just used along with matrix operations to simplify the expression of some operations. At the other extreme, rows and columns correspond to co- and contra-variant vectors, in which case transpose is completely non-trivial.
 * My hunch is that the co- and contravariance convention is useful for some limited cases in which all transformations are mixex of type (1,1) and all (co-) vectors are either of type (0,1) or (1,0). But that usage doesn't extend to problems involving things like type-(0,2) or type-(2,0) tensors since usual linear algebra doesn't allow for a row vector of row vectors. My hunch is that in this case, transpose is used a kludge to allow expressions like $$g_{ij} v^i v^j$$ to be represented with matrices as $$v^\top g v$$. Does that sound right, or am I jumping to conclusions? If this is right, it could do with some explanation somewhere. —Ben FrantzDale (talk) 15:13, 23 November 2009 (UTC)

Using an orthonormal basis, $$\scriptstyle g_{ij}=\delta_{ij}\,$$, and $$\scriptstyle g_{ij} v^i v^j=(g_{ij} v^i) v^j=(\delta_{ij} v^i) v^j=v^\top_j v^j=v^\top v.$$ That "usual linear algebra doesn't allow for a row vector of row vectors" is the reason why tensor notation is used when a row vector of row vectors, such as $$\scriptstyle g_{ij}\,$$, is needed. Bo Jacoby (talk) 16:53, 23 November 2009 (UTC).


 * Also note that there is no canonical isomorphism between V and V* if V is a plain real vector space of finite dimension >1, with no additional structure. What is of course canonical is the pairing VxV* → R. Fixing a base on V is the same as fixing an isomorphism with Rn, hence produces a special isomorphism V→V*, because Rn does possess a preferred isomorphism with its dual, that is the transpose, if we represent n-vectors with columns and forms with rows. Fixing an isomorphism V→V* is the same as giving V a scalar product (check the equivalence), which is a tensor of type (0,2), that eats pairs of vectors and defecates scalars. --pma (talk) 18:46, 23 November 2009 (UTC)


 * Those are great answers! That clarifies some things that have been nagging me for a long time! I feel like It is particularly helpful to think that conventional matrix notation doesn't provide notation for a row of row vectors or the like. I will probably copy the above discussion to Talk:Transpose for postarity and will probably add explanation along these lines to appropriate articles.
 * I haven't worked much with complex tensors, but your use of conjugate transpose reminds me that I've also long been suspiscious of its "meaning" (and simply that of complex conjugate) for the same reasons. Could you comment on that? In some sense $$c^*c$$ on a complex number, $$c=a+bi$$ is the same operation as $$v^\top v$$ on a vector, using conjugate transpose as a mechanism to compute $$c^*c=a^2+b^2$$. For a complex number, I'm not sure what would generalize to "row vector" or "column vector"... I'm not sure what I'm asking, but I feel like there's a little more that could be said connecting the above great explanations to conjugate transpose. :-) —Ben FrantzDale (talk) 19:19, 23 November 2009 (UTC)
 * A complex number (just as a real number) is a 1-D vector, so rows and columns are the same thing. The modulus on $$\Complex$$ can be thought of as a special case of the norm on $$\Complex^n$$ (ie. for n=1). From an algebraic point of view, complex conjugation is the unique (non-trivial) automorphism on $$\Complex$$ that keeps $$\R$$ fixed. Such automorphisms are central to Galois theory. I'm not really sure what the importance and meaning is from a geometrical or analytic point of view... --Tango (talk) 19:41, 23 November 2009 (UTC)

Let V and W be two vector spaces, and &fnof; : V &rarr; W be a linear map. Let F be the matrix representation of &fnof; with respect to some bases {vi} and {wj}. I seem to recall, please do correct me if I'm wrong, that F : V &rarr; W and FT : W* &rarr; V* where V* and W* are the dual spaces of V and W respectively. In this setting vT is dual to v. So the quantity vTv is the evaluation of the vector v by the covector vT. Dr Dec  (Talk)   23:26, 23 November 2009 (UTC)

Orthogonal Matrices
In the "Special transpose matrices" section, the writing implies that an orthogonal matrix G is defined as one for which GT=G-1. Thus I was going to change the "if" in "...that is, G is orthogonal if..." to "iff" but I was unsure if this was really a fundamental definition. The "Orthogonal Matrix" page does the same as this one.

It seems like a decent definition of an orthogonal matrix could be a G such that GGT and GTG are (one or both) diagonal or something. Not necessarily that one, but it's enough to make me suspect there's a more general definition some people might use.

Hopefully someone better versed in (multi-)linear algebra literature comes along and knows if there's a more general definition. If there isn't, or if it's still fully compatible, let's change the "if" to "iff" here and possibly in the "Orthogonal Matrix" page too. --Horn.imh (talk) 19:20, 16 June 2011 (UTC)


 * I'm pretty sure you are right and that it is iff. Suppose G is not orthogonal. Then two columns of G aren't orthogonal (or a column doesn't have norm of one). Then $$G^\top G$$ will not be diagonal (because off-diagonal terms are inner products of columns with different columns) in the case that the columns aren't orthogonal, and it will have something other than one on the diagonal in the case that any columns don't have norm of one (because the diagonals are the inner products of columns with themselves). QED. —Ben FrantzDale (talk) 20:05, 16 June 2011 (UTC)

Notation in 'Transpose of Linear Maps'
$$o$$ is a terrible notation for anything because it looks like a zero. We should change this to $$w$$ or something. Is the author trying to suggest an $$\omega$$? Because then they should just use the $$\omega$$...

Also, we need to be consistent for our transpose notation. Should it be $$\textstyle{^t}v$$ $$\textstyle{^T}v$$ $$v^T$$ or $$v^t$$? We use three of the four possibilities here.

129.32.11.206 (talk) 19:16, 10 October 2012 (UTC)

Transpose of linear maps: why defined in terms of a bilinear form?
In the section Transpose of linear maps, the abstract definition of a transpose is in principle independent of any bilinear form. This was stated in this way until changed by this edit (which may have been taken from Linear Algebra Quick Study Guide for Smartphones and Mobile Devices). This fundamentally changes the definition of a transpose in the abstract context. It would make more sense to me if it were defined primarily in the metric-free context, and (if desired) related to the concept defined in the section at present when suitable bilinear forms are available. I suggest reverting this section to the earlier form, with the approach using bilinear forms omitted. Does anyone with more familiarity of the area know what the most generally accepted definition is? — Quondum 14:02, 1 June 2013 (UTC)


 * I would call what is described in that section of the article the adjoint rather than the transpose, although I'm not sure whether there is a universally accepted definition. It would make sense to me to define the transpose in a metric free setting and define the adjoint as a generalization.  I'm a little surprised that we don't already have an article on the adjoint (except for the special case hermitian adjoint.)  Sławomir Biały  (talk) 14:59, 1 June 2013 (UTC)


 * Thanks. It is a pity that the definitions seem to be a little variable (gauging from the few references I've browsed).  I'll make a change along these lines in the next week or so, any comment from other editors being welcome.  — Quondum 11:07, 2 June 2013 (UTC)
 * I've made some comprehensive changes to the section, criticism welcome. I also removed a misguided association of the transpose of a coordinate vector and the more abstract concept of a transpose from Dual basis.  — Quondum 02:00, 5 June 2013 (UTC)

Error in definition of Hermitian matrix?
States $$\mathbf{A}^{\mathrm{T}} = \mathbf{A}^{*}$$ but I think it should be $$\mathbf{A}^{\mathrm{T}} = \overline{\mathbf{A}}$$. I think this is also what user Moala pointed out below way back in 2005. — Preceding unsigned comment added by 2601:9:2C80:464:809A:FAF:9F56:4439 (talk) 03:42, 19 May 2014 (UTC)


 * I think this is a matter of notation: sometimes the star is used to denote complex conjugate. I've changed this in the interest of reducing ambiguity and misunderstanding. —Quondum 04:41, 19 May 2014 (UTC)

Which Prime character?
> (also written A′, Atr, tA or At)

Should that be ′ or ʹ? Source? JDAWiseman (talk) 08:18, 9 November 2017 (UTC)

Yodalearning
The following external link was removed: Clicking on this link calls for enrollment in a course. If someone enrolls and finds good information on Transpose, then it might be used. For now it is just linkspam. — Rgdboer (talk) 22:59, 24 July 2018 (UTC)
 * How to create Transpose, yodalearning.com

Matrix to the Ath power ...?
In the text, should "a matrix raised to the Ath power" read instead "a matrix raised to the T-th power"? — Preceding unsigned comment added by Lehnekbn (talk • contribs) 21:42, 28 January 2019 (UTC)

Adjugate
The following was removed:
 * The reason transpose of a matrix is used is to get the sin(θ) between two vectors in a matrix. The dot product gives the cos(θ) of two vectors, and if we want to get the sin(θ), we would have to do a cos inverse operation to get the angle or use sin^2 + cos^2 = 1. Transpose of a matrix rotates the matrix angle to its complementary angle of pi/2 changing cos(θ) to sin(θ), allowing us to use simpler equations especially in inverse(A) = transpose(adjoint(A)) / det(A). An equation with sin(θ) and cos(θ) is vastly simpler than one with just cos(θ) to compute all other mathematical equations using tan(θ), tanh(θ), etc.

Perhaps the contributor means adjugate matrix. The business of sin(θ) being found this way is unclear and is given with no reference. — Rgdboer (talk) 22:39, 22 February 2019 (UTC)

New WP:User Devssh has taken an interest in this article, made the above contribution, thanked me for the correction, and today entered more unhelpful edits into the article. Only contributes here and has not taken up a user page. Devssh is encouraged to communicate in this Talk space before contributing further. — Rgdboer (talk) 01:01, 7 March 2019 (UTC)

Pertransposed
In some articles about numerical methods for control systems there is also the notion of the pertransposed matrix.

An example for such an article is:.

The definition there is: "transposed with respect to the main antidiagonal". (Just search for pertransposed within that article.)

If I take this literally for a square matrix $$A\in\mathbb{R}^{n\times n}$$ I get the pertransposed as $$A^P = A({\rm end}:-1:1,{\rm end}:-1:1)'$$ with Octave notation.

This is also the way it is defined in.

But, in it is used as $$A^P = A({\rm end}:-1:1,{\rm end}:-1:1)$$ without the transposition. Maybe that is just a typo. But, it caused me to search for pertranspose here on Wikipedia.

I needed the pertransposed in the context of the generation of the Quasi-Kronecker form of a matrix pencil (see e.g., Varga's article).

Wouldn't it be great to mention the pertranspose with the definition from articles as one of the generalizations of the transposed matrix here in this article? --TN (talk) 08:14, 8 January 2021 (UTC)

Serious omission
For the transpose Mt of a real square matrix considered as a linear mapping M : Rn —> Rn, with the standard dot product ⟨v,w⟩, we have the standard fact that

⟨Mv,w⟩ = ⟨v,Mtw⟩

for all vectors v, w ∊ Rn.

Yet in the article this fact is buried deep in the article, in the section Adjoint, only in the greatest generality.

The section Adjoint is entirely appropriate. But the basic fact above, in its most common manifestation, ought to be mentioned much, much earlier in the article. Especially because that is often how the transpose is defined. 2601:200:C000:1A0:291B:4FAF:4C47:67FE (talk) 18:18, 24 September 2021 (UTC)

Overwhelming details without the right intuition
This Wikipedia article can easily overwhelm a beginner. there is no point in giving all the unstructured & unstitched details. The article requires a section on motivation for transposes and the link to least squares and the four fundamental sub spaces. Can we work towards a well directed article? With a couple of volunteers I can take up this responsibility. 103.118.50.5 (talk) 05:46, 18 September 2022 (UTC)
 * Hello Rupnagar. Using this Talk space to start, please explain the 4 fundamental subspaces. And what is the connection to  least squares? Note that transpose refers to binary relations as well as linear transformations, so various details are required. As for motivation, the relation (mathematics) context is common. Rgdboer (talk) 04:02, 19 September 2022 (UTC)