Contravariant: Linear Combination of Vectors 푥1 = 3 푛 푥1 퐛 퐯 = ෍ 푥푖퐛푖 → 퐯 ≡ ⋮ 1 푥2 = 2 푥푛 푖=1 퐛2 ▪ Covariant

Contravariant: Linear Combination of Vectors 푥1 = 3 푛 푥1 퐛 퐯 = ෍ 푥푖퐛푖 → 퐯 ≡ ⋮ 1 푥2 = 2 푥푛 푖=1 퐛2 ▪ Covariant

Modelling 1 SUMMER TERM 2020 ′ 푓 퐱 푥1 = 퐯, 퐛1 = 2 퐯, 퐛1 푥1 = 1.5 ′ 퐛1 퐛1 푥2 = 1 ′ 푥2 = 퐯, 퐛ퟐ 퐛2 ′ 퐛2 = 2 퐯, 퐛ퟐ ADDENDUM Co- and Contravariance Michael Wand · Institut für Informatik · . Covariance & Contravariance Representing Vectors Two operations in linear algebra 푓 퐱 ▪ Contravariant: Linear combination of vectors 푥1 = 3 푛 푥1 퐛 퐯 = ෍ 푥푖퐛푖 → 퐯 ≡ ⋮ 1 푥2 = 2 푥푛 푖=1 퐛2 ▪ Covariant: Projection on vectors (w/scalar product) 푥1 = 퐯, 퐛1 퐯, 퐛1 퐯 ≡ ⋮ 퐯, 퐛푛 퐛1 푥2 = 퐯, 퐛ퟐ 퐛2 Where is the difference? Change of basis 푓 퐱 ▪ Contravariant: 푥1 = 1.5 ▪ Keep same output vector: 퐛1 −1 푥2 = 1 퐛푖 → 퐓퐛푖 requires 퐱 → 퐓 퐱 퐛2 ▪ Covariant: ′ 퐱, 퐛1 푥1 = 퐯, 퐛1 푓 퐱 = ⋮ = 2 퐯, 퐛1 퐱, 퐛푛 ▪ Keep same output vector: ′ 퐛1 ′ 퐛 → 퐓퐛 requires 퐱 → 퐓퐱 푥2 = 퐯, 퐛ퟐ 푖 푖 ′ 퐛2 = 2 퐯, 퐛ퟐ Awesome Video Tensors, Co-/Contra-Variance ▪ „Tensors Explained Intuitively: Covariant, Contravariant, Rank“ Physics Videos by Eugene Khutoryansky https://www.youtube.com/watch?v=CliW7kSxxWU Covariance & Contravariance Linear map 퐟: 푉1 → 푉2 Matrix representation (standard basis) 퐌 ∈ ℝ푑1×푑2 Change of basis | | | | 퐁 = 퐛 1 ⋯ 퐛 1 , 퐁 = 퐛 2 ⋯ 퐛 2 1 1 푑1 2 1 푑2 | | | | New matrix representation (bases 퐁1, 퐁2) −1 푑1×푑2 퐁2 퐌퐁1 ∈ ℝ Covariance & Contravariance Situation covariant 퐟: 푉1 → 푉2 퐱 퐈 퐌 퐈 mapping M −1 퐁1 퐁2 퐌퐁1 퐁2 퐲 contravariant Transformation law ▪ Input vectors 퐱 (퐌퐱): 퐱 퐁1 = 퐁1퐱 퐈 (covariant) −1 ▪ Output vectors 퐲 = 퐌퐱: 퐲 퐁2 = 퐁2 퐲 퐈 (contravariant) Covariance & Contravariance Situation covariant 퐟: 푉1 → 푉2 퐱 퐈 퐌 퐈 mapping M −1 퐁1 퐁2 퐌퐁1 퐁2 퐲 contravariant Transformation law ▪ Input vectors 퐱 (퐌퐱): 퐱 퐁1 = 퐁1퐱 퐈 (covariant) −1 ▪ Output vectors 퐲 = 퐌퐱: 퐲 퐁2 = 퐁2 퐲 퐈 (contravariant) Covariance & Contravariance −1 푓 퐱 ← 퐁2 퐌퐁1 ← 퐱 −1 퐁2 퐌퐁1 transforms row-vectors −1 퐁2 퐌 퐁1 transforms column-vectors Scalar Product covariant 퐱 퐲 General scalar product 푑 퐱, 퐲 ∈ ℝ mapping M 퐱, 퐲 = 퐱T 퐐 퐲, T 퐐 = 퐐 , 퐐 > 0 ⟨퐱, 퐲⟩ scalar (no coordinate system) Change of basis 퐈 퐁 퐱, 퐲 = 퐱T 퐐 퐲 퐱 → 퐁퐱, 퐈 퐲 → 퐁퐲 T 퐓 퐱, 퐲 퐁 = 퐱 ⋅ 퐁 ⋅ 퐐 ⋅ 퐁 ⋅ 퐲 Three shades of dual PCA, SVD, MDS Inputs and Outputs Input (“covariant”) side of the matrix Output (“contravariant”) side of the matrix Squaring a Matrix ▪ Possibility 1: 퐀 ⋅ 퐀푇 ⋅ = ▪ Possibility 2: 퐀푇 ⋅ 퐀 ⋅ = A Story about Dual Spaces VT X U D 1 0 0 0 0 0 = 0 2 0 0 0 0 0 0 3 0 0 0 SVD 0 0 0 4 0 0 XT X U D2 UT 1² 0 0 0 0 2² 0 0 = PCA 0 0 3² 0 0 0 0 4² T X V D2 VT X 1² 0 0 0 0 0 0 2² 0 0 0 0 0 0 3² 0 0 0 = MDS 0 0 0 4² 0 0 0 0 0 0 0 0 0 0 0 0 0 0 Tensors: Multi-Linear Maps Tensors General notion: Tensor ▪ Tensor: multi-linear form*) with 푟 ∈ ℕ input vectors 퐓: 푉1, … , 푉푛, 푉푛+1, … , 푉푟 → 퐹 (usually: field 퐹 = ℝ) ▪ “Rank 푟” tensor ▪ Linear in each input (when keeping the rest constant) ▪ Each input can be covariant or contravariant ▪ (푛, 푚) tensor ▪ 푟 = 푛 + 푚 ▪ 푛 – contravariant inputs ▪ 푚 – covariant inputs *) i.e.: multi-linear mapping / function Tensors Representation ▪ Represented as r-dimensional array 푡푖1,푖2,…,푖푛 푗1,푗2,…,푗푚 ▪ n – contravariant inputs (“indices”) ▪ m – covariant inputs (“indices”) ▪ Mapping rule 퐓 퐯 1 , … , 퐯 푛 , 퐰 1 , … , 퐰 푚 ≔ ෍ ⋯ ෍ ෍ ⋯ ෍ 푣 1 ⋯ 푣 푛 푤 1 ⋯ 푤 푚 푡푖1,푖2,…,푖푛 푖1 푖푛 푗1 푗푚 푗1,푗2,…,푗푚 푖1=0,…,푛푖1 푖푛=0,…,푛푖푛 푗1=0,…,푛푗1 푗푚=0,…,푛푗푚 (Note: writing the application of 퐓 as multi-linear mapping here) Tensors Remarks ▪ No difference between co-/contravariant dimensions in terms of numerical representation ▪ Generalization of matrix Example 푧1 푥1 푦1 퐓 , , 푧2 = 42푥1푦1푧1 + 23푥1푦1푧2 + ⋯ + 16푥2푦2푧3 푥2 푦2 푧3 ▪ Purely linear polynomial in each input parameter when all others remain constant. ▪ 3D array - 2 × 2 × 3 combinations of coefficients Einstein Notation Example: Quadratic polynomial ℝ3 → ℝ3 푝푗 퐱 = 퐱T퐀퐱 + 퐛퐱 + c 3 3 3 푗 푗 푗 푗 푝 = ෍ ෍ 푥푘푥푙 푎푘푙 + ෍ 푥푘푏푘 + c 푘=1 푙=1 푘=1 Tensor notation ▪ Input: 푥푖, 푖 = 1. 3 ▪ Quadratic form (Matrix) 퐀: 푎푘푙 푗 ▪ Output: 푝 , 푗 = 1. 3 ▪ Linear form (Co-Vector) 퐛: 푏푘 ▪ Constant 푐 Einstein Notation Example: Quadratic polynomial ℝ3 → ℝ3 푝푗 퐱 = 퐱T퐀퐱 + 퐛퐱 + c Einstein notation (implicit sums over common indices) 푗 푗 푗 푗 푝 = 푥푘푥푙푎푘푙 + 푥푘푏푘 + c Tensor notation ▪ Input: 푥푖, 푖 = 1. 3 ▪ Quadratic form (Matrix) 퐀: 푎푘푙 푗 ▪ Output: 푝 , 푗 = 1. 3 ▪ Linear form (Co-Vector) 퐛: 푏푘 ▪ Constant 푐 Further Examples Examples ▪ (푛, 푚)-tensor ▪ n contravariant “indices” ▪ m covariant “indices” ▪ Matrix: (1,1)-tensor ▪ Scalar product: (0,2)-tensor ▪ Vector: (1,0)-tensor ▪ Co-vector: (0,1)-tensor ▪ Geometric vectors: (1,0)-tensors Covariant Derivatives? Examples ▪ Geometric vectors: (1,0) tensors ▪ Derivatives*) / gradients / normal vectors: (0,1) tensors *) to be precise: ▪ Spatial derivatives co-vary for changes of the basis of the space – 푓: ℝ푛 → ℝ, 푓 퐱 = 퐲, ⇒ 훻푓 is covariant (0,1). – Examples: Gradient vector ▪ Derivatives of vector functions by unrelated dimensions remain contravariant d – 푓: ℝ → ℝ푛, 푓 푡 = 퐲, ⇒ 푓 remains contravariant (1,0). d푡 – Examples: velocity, acceleration 푛 푚 ▪ Mixed case: 푓: ℝ → ℝ , 훻푓 = 퐽푓 is a (1,1)-tensor (1,1) Example: Plane Equation Plane equation(s) ▪ Parametric: 퐱 = 휆1퐫1 + 휆2퐫2 + 퐨 ▪ Implicit: 퐧, 퐱 − 푑 = 0 Transformation 퐱 → 퐓퐱 ▪ Parametric: 퐓퐱 = 퐓 휆1퐫1 + 휆2퐫2 + 퐨 = 휆1퐓퐫1 + 휆2퐓퐫2 + 퐓퐨 ▪ Implicit: 퐧, 퐓퐱 − 푑 = 퐧T퐓 퐱 − 푑0 More Structure? Connecting ▪ Integrals ▪ Derivatives ▪ In higher dimensions ▪ And their transformation rules “Exterior Calculus” ▪ Unified framework ▪ Beyond this lecture (take a real math course :-) ) Vectors & Covectors in Function Spaces Remark: Function Spaces Discrete vector spaces ▪ Picking entries by index is a linear operation ▪ Can be represented by projection to vector (multiplication with “co-vector”) Example ▪ 퐱 = (푥1, 푥2, 푥3, 푥4, 푥5) ▪ 퐱 ↦ 푥4 is a linear maps ▪ Represented by 0,0,0,1,0 , 퐱 ▪ “Linear form”: 퐱 ↦ 0,0,0,1,0 , 퐱 , ퟎ T ퟎ in short, ⋅, 0,0,0,1,0 , shorter: 0,0,0,1,0 = ퟎ ퟏ ퟎ Linear Forms in Function Spaces In function spaces ▪ Picking entries by x-axis is a linear operation ▪ Cannot be represented by projection to another function (multiplication with “co-vector”) 푔 1, if 푥 = 4 푔 푥 = ቊ 푓 0, elsewhere Example න 푓 푥 푔 푥 푑푥 = 0 ℝ ▪ 푓: ℝ → ℝ, 푓 푥 = sin 푥 ▪ 퐿: ℝ → ℝ → ℝ, 퐿: 푓 ↦ 푓 4.0 is a linear map ▪ A function 푔 with 푔, 푓 = 푓 4.0 does not exist Dirac’s “Delta Function” f(x) f(x) x x f(x) f(x) න 푓 푥 푑푥 = 1 x x Ω Dirac Delta “Function” ▪ ׬ℝ 훿 푥 푑푥 = 1, zero everywhere but at 푥 = 0 ▪ Idealization (“distribution”) – think of very sharp peak Distributions Distributions ▪ Adding all linear forms to the vector space ▪ All linear mappings from the function space to ℝ ▪ This makes the situation symmetric ▪ 훿 is a distribution, not a (traditional) function Formalization ▪ Different approaches (details beyond our course) ▪ Limits of “bumps” ▪ Space of linear-forms (“co-vectors”, “dual functions”) ▪ Difference of complex functions on Riemann sphere (exotic).

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    28 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us