tinytheorems - about groups - rings - modules - models atom/rss

Exploration 4: Character theory

.

Questions:


There are two problems when it comes to working with matrix representations:

Let’s start by defining what it means for two (general) representations ρ:GAut(M)\rho:G\to\Aut(M) and ρ:GAut(M)\rho':G\to\Aut(M') to be the same representation. Say that these two representations are equivalent if there is an isomorphism ϕ\phi between MM and MM' that preserves the group action, i.e. ϕρ(g)=ρ(g)ϕ\phi\circ\rho(g)=\rho'(g)\circ\phi for all gGg\in G, i.e. the following diagram commutes:

Mρ(g)Mϕϕ where ϕ:MM is an isomorphismMρ(g)M\begin{aligned} M&\xrightarrow{\rho(g)}&M \\ \downarrow{\phi}&&\downarrow{\phi}&\quad\text{ where }\phi:M\to M'\text{ is an isomorphism}\\ M'&\xrightarrow{\rho'(g)}&M' \end{aligned}

Let’s explore what happens in the case of matrix representations. When M,MM,M' are free modules RnR^n, then ρ,ρ\rho,\rho' become matrix representations, and thus Aut(Rn)Aut(R^n) is composed of n×nn\times n matrices. Then the problem of checking whether M,MM,M' are isomorphic comes down to comparing properties of the representative matrices.

Theorem: Conjugate elements g,hg,h in GG are represented by similar matrices A,BA,B in any matrix representation ρ:GAut(Rn)\rho:G\to\Aut(R^n).
  • g,hg,h being conjugate elements means kgk1=hkgk^{-1}=h for some kGk\in G. Applying ρ\rho gives CAC1=BCAC^{-1}=B, which is the same as saying AA and BB are similar.

Corollary: Conjugacy classes in GG are represented by similarity classes in Aut(Rn)\Aut(R^n).

Corollary: Equivalent matrix representations differ only in which similarity class they assign to each conjugacy class.

In this section, we explore some facts about class functions.

Since representations of elements of GG in the same conjugacy class are similar matrices, we’d like to classify the similarity classes of matrices by defining a function on matrices that is unchanged under conjugation. So we’re interested in class functions on GG, functions GRG\to R that are invariant under conjugation in GG.

Theorem: Given RR a commutative ring, the class functions GRG\to R form an RR-module.
  • Most of this comes from RR itself. Given f1,f2:GRf_1,f_2:G\to R are two arbitrary class functions on GG, we can show that the class functions form an additive abelian group:
    • Closure: (f1+f2)(g)=f1(g)+f2(g)(f_1+f_2)(g)=f_1(g)+f_2(g)
    • Identity: 0(g)=00(g)=0
    • Inverse: (f1)(g)=f1(g)(-f_1)(g)=-f_1(g)
    • Commutativity: inherited from RR
    • Associativity: inherited from RR
  • To show that they form an RR-module, we have closure under scalar multiplication: (rf)(g)=rf(g)(rf)(g)=r\cdot f(g)

Corollary: Given GG a finite group and RR a commutative ring, the class functions GRG\to R form a free RR-module.
  • This is the same as the previous theorem, but now we need to prove that class functions on finite groups form a free RR-module. This requires showing a basis.
  • The standard basis (in the form of indicator functions) will do. Since GG is finite, index the conjugacy classes from 11 to nn, and let uiu_i be the class function whose value is 11 on elements from the iith class and 00 otherwise.
  • Then the uiu_i form a basis, meaning that the RR-module of class functions is free.

Ideally we can decompose every class function GRG\to R into a matrix representation GAut(Rn)G\to\Aut(R^n) followed by a linear class function Aut(Rn)R\Aut(R^n)\to R. This works out nicely because of the following theorem:

Theorem: For matrices MM over an integral domain RR, the trace tr(M)\tr(M) is the unique nontrivial linear class function, up to scalar multiplication.
  • Invariance under conjugation means t(A)=t(BAB1)t(A)=t(BAB^{-1}). Due to linearity of tt, this is equivalent to t(AB)=t(BA)t(AB)=t(BA).
  • We decompose the input matrix MM into its matrix units EijE_{ij}, where EijE_{ij} has a one on the iith row and jjth column and zero elsewhere. Then t(AB)=t(BA)t(AB)=t(BA) implies t(EijEkl)=t(EklEij)t(E_{ij}E_{kl})=t(E_{kl}E_{ij}).
  • Assume that both sides are nonzero, since otherwise the integral domain lets us factor out t(I)=0t(I)=0 which implies t(anything)=0t(\text{anything})=0 by linearity, and we’re not interested in trivial tt.
  • Since t(0)=0t(0)=0 by linearity, t(M)t(M) is only nonzero when MM is nonzero. Over an integral domain, this means EijEklE_{ij}E_{kl} and EklEijE_{kl}E_{ij} must be nonzero as well.
  • But EijEklE_{ij}E_{kl} is only nonzero when j=kj=k, and EklEijE_{kl}E_{ij} is only nonzero when i=li=l. Then EijEkl=Eil=EiiE_{ij}E_{kl}=E_{il}=E_{ii} and EklEij=Ekj=EjjE_{kl}E_{ij}=E_{kj}=E_{jj} must be nonzero. Further, we have t(Eii)=t(Ejj)t(E_{ii})=t(E_{jj}) for all i,ji,j implying that all t(Eii)t(E_{ii}) is a constant λ\lambda.
  • Therefore t(Eij)=λt(E_{ij})=\lambda if i=ji=j and is zero otherwise. Then we can decompose the original matrix and apply linearity of tt to get: t(M)=t(i,jmijEij)=i,jmijt(Eij) by linearity of t=imiit(Eii) since t(Eij)=0 for ichar "338=j=λimii since t(Eij)=λ for i=j\begin{aligned} t(M)&=t\left(\sum_{i,j} m_{ij}E_{ij}\right)\\ &=\sum_{i,j} m_{ij}t(E_{ij})&\text{ by linearity of }t\\ &=\sum_i m_{ii}t(E_{ii})&\text{ since }t(E_{ij})=0\text{ for }i\ne j\\ &=\lambda\sum_i m_{ii}&\text{ since }t(E_{ij})=\lambda\text{ for }i=j\\ \end{aligned}
  • The trace of a matrix tr(M)\tr(M) is exactly imii\sum_i m_{ii}, therefore t(M)=λtr(M)t(M)=\lambda\tr(M), and so tt must be some scalar multiple of the trace operator.

That is, the only linear class function is the trace, and therefore we can decompose every class function GRG\to R into a matrix representation ρ:GAut(Rn)\rho:G\to\Aut(R^n) followed by the trace (multiplied by some scalar).

G G R R G->R χ Aut Aut (R n ) G->Aut ρ Aut->R tr

In this section, we discover how to classify equivalent matrix representations.

Given a matrix representation ρ:GAut(Rn)\rho:G\to\Aut(R^n), define the character of ρ\rho as χρ=trρ\chi_\rho=\tr\circ\rho, essentially taking the trace of each representative matrix for gGg\in G. By construction, characters are one-dimensional representations GRG\to R that are invariant under conjugation in GG. Because equivalent matrix representations only differ by conjugation in GG, representations that are invariant under conjugation are perfect for classifying matrix representations. In particular, each distinct value that χρ(g)\chi_\rho(g) takes on is representative of a distinct conjugacy class in GG.

Assume we’re in an integral domain RR whose characteristic is 00, so that char RG\char R\nmid |G| is true for all finite GG. Then Maschke’s theorem lets us describe any group representation in terms of its irreps, and so we shift our focus of study towards irreps. Characters of irreps are known as irreducible characters.

Characters have a number of other properties in an integral domain, and much more in a field. Let’s go over the integral domain properties first.

In this section, we view a strategy for classifying group representations over fields.

Recall Schur’s lemma:

Theorem: Any homomorphism σ:MN\sigma:M\to N between simple modules MM and NN is either trivial or an isomorphism.

A corollary (often also called Schur’s lemma) appears when RR is actually an algebraically closed field FF:

Schur’s Lemma: If the field FF is algebraically closed, then any finite irrep ρ:GAut(F)\rho:G\to\Aut(F) must be a multiplication by a scalar.
  • Every automorphism ϕAut(F)\phi\in\Aut(F) over a vector field is a matrix, and the characteristic polynomial of every matrix over an algebraically closed field FF has at least one root λ\lambda.
  • Since λ\lambda is an eigenvalue it is associated with an eigenspace, a subspace of FF. Since ρ\rho is an irrep, FF is simple, and by our previous form for Schur’s Lemma this means ϕ\phi is either trivial or an isomorphism.
  • It’s not possible for ϕ\phi to be trivial since we assume ϕ\phi being an automorphism is not the zero map. So the image of ϕ\phi must be all of FF, implying that ϕλI\phi-\lambda I is the zero map, thus ϕ\phi is multiplication by a scalar.
  • Since ϕ\phi is an automorphism by definition, it must be an isomorphism and not trivial. Since ϕ(x)=λx\phi(x)=\lambda x, this means ϕ=λI\phi=\lambda I, which is multiplication by a scalar.

Character theory is most often used over the field of complex numbers C\CC. This is mostly because C\CC is an algebraically closed field, which gives us the above.

TODO either expand on character theory theorems like orthogonality in C OR go back and highlight the importance of

TODO: three proofs that link together to get schur’s lemma


Let ϕ : ρ_1 → ρ_2 be a nonzero homomorphism.

prove it in 2 parts:

If ρ_2 is irreducible then ϕ is surjective.
If ρ_1 is irreducible then ϕ is injective. 

The inner product ,\langle,\rangle is a function of two arguments Rn×RnRR^n\times R^n\to R that satisfy the following:

Conjugate Symmetry Linearity in the first argument Positive-definiteness

Irreducible characters are orthogonal

Orthogonal means ⟨φ,ψ⟩=1 if they’re the same representation up to isomorphism, and =0 otherwise.

Irreducible characters are orthogonal, proof is in lecture 22

We call this property character orthogonality, as in “by character orthogonality…”

Computing character tables over ℂ

A character table has irreducible characters χ as rows and conjugacy classes of G as columns. Then each entry is χ((g)) where g is in that conjugacy class of G. For example, let ζ = a primitive third root of unity. Then use the degree 1 irreducible χ⁽ᵏ⁾ = g ↦ ζᵏ. Then the cyclic group C₃ = ⟨g⟩ has the character table:

       1   1   1      ← size of the conjugacy class of C₃
      (1) (g) (g²)    ← the conjugacy classes of C₃
     ------------
χ⁽⁰⁾ | 1   1   1      ← trivial representation g ↦ [1]
χ⁽¹⁾ | 1   ζ   ζ²     ← dim 1 representation g ↦ [ζ]
χ⁽²⁾ | 1   ζ²  ζ      ← dim 1 representation g ↦ [ζ²]

If we define χ⁽ᵏ⁾ = g ↦ iᵏ (a dim 1 representation), then the cyclic group C₄ = ⟨g⟩ has the character table:

       1   1   1   1      ← size of the conjugacy class of C₃
      (1) (g) (g²)(g³)    ← the conjugacy classes of C₃
     ----------------
χ⁽⁰⁾ | 1   1   1   1      ← trivial representation g ↦ [1]
χ⁽¹⁾ | 1   i  -1  -i      ← dim 1 representation g ↦ [ζ]
χ⁽²⁾ | 1  -1   1  -1      ← dim 1 representation g ↦ [ζ²]
χ⁽³⁾ | 1  -i  -1   i      ← dim 1 representation g ↦ [ζ³]

The symmetric group S₃ = ⟨(1 2),(1 2 3)⟩ has the character table:

       1    1    1      ← size of the conjugacy class of C₃
      (1) (12) (123)    ← the conjugacy classes of C₃
     ---------------
1    | 1    1    1      ← trivial representation π ↦ [1]
χˢ   | 1   -1    1      ← dim 1 sign representation π ↦ sign(π)
χᵁ   | 2    0   -1      ← dim 2 representation π ↦ (number of fixed points of π) - 1

The dihedral group D₄ = ⟨r,s | r⁴=s²=e, srs = r⁻¹⟩ has the character table:

       1    1    1       1       1       ← size of the conjugacy class of C₃
      {e} {r²} {r,r³} {s,sr²} {sr,sr³}   ← the conjugacy classes of C₃
     ---------------------------------
χ₁   | 1    1    1       1       1       ← trivial representation g ↦ [1]
χ₂   | 1    1   -1       1      -1       ← dim 1 representation g ↦ ?
χ₃   | 1    1    1      -1      -1       ← dim 1 representation g ↦ -1 if reflection
χ₄   | 1    1   -1      -1       1       ← dim 1 representation g ↦ ?
χ₅   | 2   -2    0       0       0       ← dim 2 representation g ↦ ?

Things to note:

This website gives all the character tables for all groups with at most 10 irreps: https://people.maths.bris.ac.uk/~matyd/GroupNames/characters.html

TODO

Recall that a field FF is algebraically closed iff every nonconstant polynomial in F[x]F[x] has a root in FF.

Theorem: Any linear operator has at least one eigenvalue in an algebraically closed field FF.
  • When we do a simple extension by an element F(α)F(\alpha), we adjoin the element α\alpha to FF and take the field of fractions. By definition, this is the smallest field containing FF and α\alpha.
  • Then F(α)(β)F(\alpha)(\beta) is the smallest field containing F(α)F(\alpha) and β\beta, i.e. it is the smallest field containing FF, α\alpha, and β\beta.
  • Likewise, F(β)(α)F(\beta)(\alpha) is the smallest field containing F(β)F(\beta) and α\alpha, i.e. it is the smallest field containing FF, β\beta, and α\alpha.
  • Therefore these are the exact same field.

Characters completely determine representations up to isomorphism

The original motivation for characters is to identify representations without regard to a basis or its matrix at all. There is also an amazing theorem that:

Theorem: ⟨χ,χ’⟩ = (1/N)∑_g (̅χ(g))χ’(g), which is just a normalized dot product in the complex vector space.

Corollary: In particular, ⟨χ,χ⟩=1 iff χ is irreducible. (This must be equal to ∑xᵢ², which can only be the dot product of a elementary basis vector with itself. Therefore irreducible)

Theorem: TODO
  • TODO


Cayley-Hamilton Theorem: Every (square) RR-matrix (for a commutative ring RR) satisfies its own characteristic equation

TODO

Characteristic polynomials

Every n×nn\times n matrix AA over FF has a characteristic polynomial χA\chi_A equal to det(xIA)\det(xI-A).

Motivation: To find the eigenvectors. χA(λ)=0\chi_A(\lambda)=0 iff λ\lambda is an eigenvalue of AA. So just factor χA\chi_A to get the eigenvalues.

Minimal polynomials

Recall you can evaluate a polynomial mm at a matrix AA: m(A)m(A).

Every n×nn\times n matrix AA over FF has a unique minimal polynomial mAm_A that is monic, lowest degree, and mA(A)=0m_A(A)=0.

Motivation: the roots of mAm_A are the eigenvalues. Proof: WTS mA(λ)=0m_A(\lambda)=0 iff λ\lambda is an eigenvalue of A.A. Since mAm_A is lowest degree, there should be no other factors in mAm_A other than the eigenvalues, if this is true. (->) mAχAm_A\mid\chi_A, so χ(λ)=0\chi(\lambda)=0, which means λ\lambda is an eigenvalue of AA. (<-) We know mA(A)=0m_A(A)=0 so we know mA(λ)v=0m_A(\lambda)\cdot v=0. If you factor mAm_A in some field where it splits into xμix-\mu_i, then we must have (xμi)v=0(x-\mu_i)\cdot v=0 for some μi\mu_i.

Another motivation: to detect diagonalizability of AA. If mAm_A factors into distinct factors it’s diagonalizable. Proof involves finding the Jordan block matrix conjugate to AA

Given ff is the minimal polynomial of α\alpha over FF, we can make an isomorphism: φ:F[x]/(f)F(α)\varphi:F[x]/(f)\iso F(\alpha) as φ=g+(f)g(α)\varphi=g+(f)\mapsto g(\alpha)

Conjugation doesn’t change these polynomials

Recall a conjugate matrix of AA is a matrix PAP1PAP^{-1} where PP is any matrix of the same size in the same field as AA. Note:

So conjugation doesn’t change either of these polynomials.

Cayley-Hamilton theorem: the characteristic polynomial χ_A of T is det(tI-T)=f₁f₂…fᵣ ∈ F[t]. Then χ_T(T) = O.



Character theory: - We care about the conjugacy classes of groups, since they help us find the irreducible characters of groups

< Back to category Exploration 4: Character theory (permalink)
Exploration 3: Representation theory Exploration 1: Modules and vector spaces