Formal Definition of Dot ProductSpecial relativity: how to prove that $g = L^t g L$?More Vector Product Possibilities?Extension of Lami's theoremHow the Poisson bracket transform when we change coordinates?Definition of vector cross productWhat exactly is the Parity transformation? Parity in spherical coordinatesSimple question about change of coordinatesDefinition of velocity in classical mechanicsDefinition of inner product as in the case of workConfusion about Change in Integration Variable

Why do academics prefer Mac/Linux?

Are there any crystals that are theoretically possible, but haven't yet been made?

Does the usage of mathematical symbols work differently in books than in theses?

Alternative classical explanation of the Stern-Gerlach Experiment?

Prints each letter of a string in different colors

Managing heat dissipation in a magic wand

In Dutch history two people are referred to as "William III"; are there any more cases where this happens?

What's is the easiest way to purchase a stock and hold it

How to laser-level close to a surface

What should I wear to go and sign an employment contract?

Why does the setuid bit work inconsistently?

Why are there five extra turns in tournament Magic?

How do you cope with rejection?

Lock out of Oracle based on Windows username

At what point can a confirmation be established between words of similar meaning in context?

How to get all possible paths in 0/1 matrix better way?

Who is frowning in the sentence "Daisy looked at Tom frowning"?

Shortest amud or daf in Shas?

How was the blinking terminal cursor invented?

Largest memory peripheral for Sinclair ZX81?

Taylor series leads to two different functions - why?

Pedaling at different gear ratios on flat terrain: what's the point?

Physically unpleasant work environment

Save my secrets!



Formal Definition of Dot Product


Special relativity: how to prove that $g = L^t g L$?More Vector Product Possibilities?Extension of Lami's theoremHow the Poisson bracket transform when we change coordinates?Definition of vector cross productWhat exactly is the Parity transformation? Parity in spherical coordinatesSimple question about change of coordinatesDefinition of velocity in classical mechanicsDefinition of inner product as in the case of workConfusion about Change in Integration Variable













16












$begingroup$


In most textbooks, dot product between two vectors is defined as:



$$langle x_1,x_2,x_3rangle cdot langle y_1,y_2,y_3rangle = x_1 y_1 + x_2 y_2 + x_3 y _3$$



I understand how this definition works most of the time. However, in this definition, there is no reference to coordinate system (i.e. no basis is included for the vector components). So, if I had two vectors in two different coordinate systems:



$$x_1 vece_1 + x_2 vece_2 + x_3 vece_3$$
$$y_1 vece_1' + y_2 vece_2' + y_3 vece_3'$$



How, would I compute their dot product? In particular, is there a more formal/abstract/generalized definition of the dot product (that would allow me to compute $vece_1 cdot vece_1'$ without converting the vectors to the same coordinate system)? Even if I did convert the vectors to the same coordinate system, why do we know that the result will be the same if I multiply the components in the primed system versus in the unprimed system?










share|cite|improve this question











$endgroup$







  • 1




    $begingroup$
    However, in this definition, there is no reference to coordinate system (i.e. no basis is included for the vector components). But I think that it IS always strongly IMPLIED that the 2 vector component sets are obtained with respect to the same orthonormal basis.
    $endgroup$
    – Trunk
    May 13 at 16:03










  • $begingroup$
    What you'd need is the change of basis matrix for the relationship between $hate_i$ and $hate_j'$, you should be able to then go from there. But as it stands technically this question could be even better answered on mathematics stack exchange as it's purely mathematical in nature
    $endgroup$
    – Triatticus
    May 13 at 18:20







  • 1




    $begingroup$
    @Evpok: In hindsight, I'm wondering how I got the cross product and dot product mixed up, especially given the definition in the question itself. Let's blame mondays.
    $endgroup$
    – MSalters
    2 days ago















16












$begingroup$


In most textbooks, dot product between two vectors is defined as:



$$langle x_1,x_2,x_3rangle cdot langle y_1,y_2,y_3rangle = x_1 y_1 + x_2 y_2 + x_3 y _3$$



I understand how this definition works most of the time. However, in this definition, there is no reference to coordinate system (i.e. no basis is included for the vector components). So, if I had two vectors in two different coordinate systems:



$$x_1 vece_1 + x_2 vece_2 + x_3 vece_3$$
$$y_1 vece_1' + y_2 vece_2' + y_3 vece_3'$$



How, would I compute their dot product? In particular, is there a more formal/abstract/generalized definition of the dot product (that would allow me to compute $vece_1 cdot vece_1'$ without converting the vectors to the same coordinate system)? Even if I did convert the vectors to the same coordinate system, why do we know that the result will be the same if I multiply the components in the primed system versus in the unprimed system?










share|cite|improve this question











$endgroup$







  • 1




    $begingroup$
    However, in this definition, there is no reference to coordinate system (i.e. no basis is included for the vector components). But I think that it IS always strongly IMPLIED that the 2 vector component sets are obtained with respect to the same orthonormal basis.
    $endgroup$
    – Trunk
    May 13 at 16:03










  • $begingroup$
    What you'd need is the change of basis matrix for the relationship between $hate_i$ and $hate_j'$, you should be able to then go from there. But as it stands technically this question could be even better answered on mathematics stack exchange as it's purely mathematical in nature
    $endgroup$
    – Triatticus
    May 13 at 18:20







  • 1




    $begingroup$
    @Evpok: In hindsight, I'm wondering how I got the cross product and dot product mixed up, especially given the definition in the question itself. Let's blame mondays.
    $endgroup$
    – MSalters
    2 days ago













16












16








16


3



$begingroup$


In most textbooks, dot product between two vectors is defined as:



$$langle x_1,x_2,x_3rangle cdot langle y_1,y_2,y_3rangle = x_1 y_1 + x_2 y_2 + x_3 y _3$$



I understand how this definition works most of the time. However, in this definition, there is no reference to coordinate system (i.e. no basis is included for the vector components). So, if I had two vectors in two different coordinate systems:



$$x_1 vece_1 + x_2 vece_2 + x_3 vece_3$$
$$y_1 vece_1' + y_2 vece_2' + y_3 vece_3'$$



How, would I compute their dot product? In particular, is there a more formal/abstract/generalized definition of the dot product (that would allow me to compute $vece_1 cdot vece_1'$ without converting the vectors to the same coordinate system)? Even if I did convert the vectors to the same coordinate system, why do we know that the result will be the same if I multiply the components in the primed system versus in the unprimed system?










share|cite|improve this question











$endgroup$




In most textbooks, dot product between two vectors is defined as:



$$langle x_1,x_2,x_3rangle cdot langle y_1,y_2,y_3rangle = x_1 y_1 + x_2 y_2 + x_3 y _3$$



I understand how this definition works most of the time. However, in this definition, there is no reference to coordinate system (i.e. no basis is included for the vector components). So, if I had two vectors in two different coordinate systems:



$$x_1 vece_1 + x_2 vece_2 + x_3 vece_3$$
$$y_1 vece_1' + y_2 vece_2' + y_3 vece_3'$$



How, would I compute their dot product? In particular, is there a more formal/abstract/generalized definition of the dot product (that would allow me to compute $vece_1 cdot vece_1'$ without converting the vectors to the same coordinate system)? Even if I did convert the vectors to the same coordinate system, why do we know that the result will be the same if I multiply the components in the primed system versus in the unprimed system?







vectors coordinate-systems linear-algebra






share|cite|improve this question















share|cite|improve this question













share|cite|improve this question




share|cite|improve this question








edited May 13 at 1:56









Gilbert

5,215919




5,215919










asked May 13 at 0:08









dtsdts

411615




411615







  • 1




    $begingroup$
    However, in this definition, there is no reference to coordinate system (i.e. no basis is included for the vector components). But I think that it IS always strongly IMPLIED that the 2 vector component sets are obtained with respect to the same orthonormal basis.
    $endgroup$
    – Trunk
    May 13 at 16:03










  • $begingroup$
    What you'd need is the change of basis matrix for the relationship between $hate_i$ and $hate_j'$, you should be able to then go from there. But as it stands technically this question could be even better answered on mathematics stack exchange as it's purely mathematical in nature
    $endgroup$
    – Triatticus
    May 13 at 18:20







  • 1




    $begingroup$
    @Evpok: In hindsight, I'm wondering how I got the cross product and dot product mixed up, especially given the definition in the question itself. Let's blame mondays.
    $endgroup$
    – MSalters
    2 days ago












  • 1




    $begingroup$
    However, in this definition, there is no reference to coordinate system (i.e. no basis is included for the vector components). But I think that it IS always strongly IMPLIED that the 2 vector component sets are obtained with respect to the same orthonormal basis.
    $endgroup$
    – Trunk
    May 13 at 16:03










  • $begingroup$
    What you'd need is the change of basis matrix for the relationship between $hate_i$ and $hate_j'$, you should be able to then go from there. But as it stands technically this question could be even better answered on mathematics stack exchange as it's purely mathematical in nature
    $endgroup$
    – Triatticus
    May 13 at 18:20







  • 1




    $begingroup$
    @Evpok: In hindsight, I'm wondering how I got the cross product and dot product mixed up, especially given the definition in the question itself. Let's blame mondays.
    $endgroup$
    – MSalters
    2 days ago







1




1




$begingroup$
However, in this definition, there is no reference to coordinate system (i.e. no basis is included for the vector components). But I think that it IS always strongly IMPLIED that the 2 vector component sets are obtained with respect to the same orthonormal basis.
$endgroup$
– Trunk
May 13 at 16:03




$begingroup$
However, in this definition, there is no reference to coordinate system (i.e. no basis is included for the vector components). But I think that it IS always strongly IMPLIED that the 2 vector component sets are obtained with respect to the same orthonormal basis.
$endgroup$
– Trunk
May 13 at 16:03












$begingroup$
What you'd need is the change of basis matrix for the relationship between $hate_i$ and $hate_j'$, you should be able to then go from there. But as it stands technically this question could be even better answered on mathematics stack exchange as it's purely mathematical in nature
$endgroup$
– Triatticus
May 13 at 18:20





$begingroup$
What you'd need is the change of basis matrix for the relationship between $hate_i$ and $hate_j'$, you should be able to then go from there. But as it stands technically this question could be even better answered on mathematics stack exchange as it's purely mathematical in nature
$endgroup$
– Triatticus
May 13 at 18:20





1




1




$begingroup$
@Evpok: In hindsight, I'm wondering how I got the cross product and dot product mixed up, especially given the definition in the question itself. Let's blame mondays.
$endgroup$
– MSalters
2 days ago




$begingroup$
@Evpok: In hindsight, I'm wondering how I got the cross product and dot product mixed up, especially given the definition in the question itself. Let's blame mondays.
$endgroup$
– MSalters
2 days ago










8 Answers
8






active

oldest

votes


















16












$begingroup$

Your top-line question can be answered at many levels. Setting aside issues of forms and covariant/contravariant, the answer is:




The dot product is the product of the magnitudes of the two vectors, times the cosine of the angle between them.




No matter what basis you compute that in, you have to get the same answer because it's a physical quantity.



The usual "sum of products of orthonormal components" is then a convenient computational approach, but as you've seen it's not the only way to compute them.



The dot product's properties includes linear, commutative, distributive, etc. So when you expand the dot product



$$(a_x hatx+a_y haty + a_z hatz) cdot (b_x hatX+b_y hatY + b_z hatZ)$$



you get nine terms like $( a_x b_x hatxcdothatX) + (a_x b_y hatxcdothatY)+$ etc. In the usual orthonormal basis, the same-axis $hatxcdothatX$ factors just become 1, while the different-axis $hatxcdothatY$ et al factors are zero. That reduces to the formula you know.



In a non-orthonormal basis, you have to figure out what those basis products are. To do that, you refer back to the definition: The product of the size of each, times the cosine of the angle between. Once you have all of those, you're again all set to compute. It just looks a bit more complicated...






share|cite|improve this answer











$endgroup$








  • 13




    $begingroup$
    I don't think the dot product is associative.
    $endgroup$
    – eyeballfrog
    May 13 at 1:18






  • 4




    $begingroup$
    "No matter what basis you compute that in, you have to get the same answer because it's a physical quantity." Only if you have a physical vector. If we're speaking mathematically, vectors can be abstract objects, and the "angle" is not defined. In fact, generally speaking, if "angle" is defined, it's defined in terms of the dot product, making your definition circular.
    $endgroup$
    – Acccumulation
    May 13 at 17:14






  • 3




    $begingroup$
    @Acccumulation This is Physics Stack Exchange.
    $endgroup$
    – Bob Jacobsen
    May 13 at 17:18






  • 4




    $begingroup$
    @Bob Jacobsen Yes, but physics also has abstract Hilbert spaces. Consider, for example, Quantum Mechanics.
    $endgroup$
    – scaphys
    May 13 at 18:36






  • 2




    $begingroup$
    "No matter what basis you compute that in, you have to get the same answer because it's a physical quantity." What about two bases which are not related by an orthogonal transformation? For example $hat e_i = 2 hat f_i$.
    $endgroup$
    – Display Name
    May 13 at 18:54


















18












$begingroup$

Dot products, or inner products are defined axiomatically, or abstractly. An inner product on a vector space $V$ over $R$ is a pairing $Vtimes Vto R$, denoted by $ langle u,vrangle$, with properties $langle u,vrangle=langle v,urangle$, $langle u+cw,vrangle= langle u,vrangle+clangle w,vrangle$, and $ langle u,uranglegt0$ if $une0$. In general, a vector space can be endowed with an inner product in many ways. Notice here there is no reference to a basis/coordinate system.



Using what is called the Gram-Schmidt process, one can then construct a basis $e_1,cdots e_n$ for $V$ in which the inner product takes the computational form which you stated in your question.



In your question, you are actually starting with what is called an orthonormal basis for an inner product. The coordinate-free approach is to state the postulates that an inner product should obey, then after being given an explicit inner product, construct an orthonormal basis in which to do computations.



In general, an orthonormal basis $e_1,e_2,e_3$ for one inner product on $V$ will not be an orthonormal basis for another inner product on $V$.






share|cite|improve this answer









$endgroup$




















    9












    $begingroup$

    The dot product can be defined in a coordinate-independent way as



    $$vecacdotvecb=|veca||vecb|costheta$$



    where $theta$ is the angle between the two vectors. This involves only lengths and angles, not coordinates.



    To use your first formula, the coordinates must be in the same basis.



    You can convert between bases using a rotation matrix, and the fact that a rotation matrix preserves vector lengths is sufficient to show that it preserves the dot product. This is because



    $$vecacdotvecb=frac12left(|veca+vecb|^2-|veca|^2-|vecb|^2right).$$



    This formula is another purely-geometric, coordinate-free definition of the dot product.






    share|cite|improve this answer











    $endgroup$












    • $begingroup$
      Thank you! That makes sense. But what happens if you are dealing with a non-orthonormal system? Is the dot product's value preserved in making the coordinate transformation?
      $endgroup$
      – dts
      May 13 at 0:21










    • $begingroup$
      Yes, the value is preserved, but the coordinate-based formula in a non-orthonormal basis is more complicated than your first formula.
      $endgroup$
      – G. Smith
      May 13 at 0:31






    • 1




      $begingroup$
      "You can convert between bases using a rotation matrix", I strongly disagree. Only if the base vectors are normalised, but that needn't be the case. However there exists a Matrix $A$ such that $e_i^prime = A e_i$ where $e_i$ is to be understood at the ith basic vector (not the component).
      $endgroup$
      – infinitezero
      May 13 at 16:33



















    7












    $begingroup$

    The coordinate free definition of a dot product is:



    $$ vec a cdot vec b = frac 1 4 [(vec a + vec b)^2 - (vec a - vec b)^2]$$



    It's up to you to figure out what the norm is:



    $$ ||vec a|| = sqrt(vec a)^2$$



    Here is a reference for this viewpoint:
    http://www.pmaweb.caltech.edu/Courses/ph136/yr2012/1202.1.K.pdf
    Section 2.3






    share|cite|improve this answer











    $endgroup$








    • 3




      $begingroup$
      This is a circular definition as the norm is defined via the dot product.
      $endgroup$
      – Winther
      May 13 at 9:01










    • $begingroup$
      @Winther You've got to input something: the dot product cannot be derived only from the underlying vector space structure. The norm seems a reasonable choice here, for geometric intuition.
      $endgroup$
      – Denis Nardin
      May 13 at 11:06






    • 2




      $begingroup$
      This will only define an inner product iff the norm satisfies the parallelogram identity $2||x||^2+2||y||^2=||x+y||^2+||x-y||^2$
      $endgroup$
      – Jannik Pitt
      May 13 at 11:50










    • $begingroup$
      Yes you have to input something: either define a norm or define an inner product and have the norm be induced by this. However my point was that you seem to define the norm via $|a|=sqrtacdot a$ which is why I said it was circular. On second reading it does look like you say you need to specify the norm externally so then this would be fine. However doesn't then the definition of the norm require you to specify a coordinate system so it's not really coordinate free?
      $endgroup$
      – Winther
      May 13 at 13:48










    • $begingroup$
      @Winther Well, it depends on how your vector space is given to you. If your vectors are a bunch of coordinates (like in the usual description of $mathbbR^n$), of course every definition you give will be coordinate dependent (coordinates are all you have!), but if your vector space is composed by something more exotic (e.g. the space of solutions of a certain ODE) then you can hope to write down a definition of the norm using something else. (and yes, indeed a Banach space is Hilbert iff the norm satisfies the parallelogram identity, plus some added condition if over $mathbbC$).
      $endgroup$
      – Denis Nardin
      May 13 at 18:10


















    3












    $begingroup$

    On computing the following matrix will give you the dot product $$beginbmatrix x_1 & x_2& x_3 endbmatrix.beginbmatrix e_1.e'_1 & e_1.e'_2 & e_1.e'_3 \ e_2.e'_1 & e_2.e'_2 & e_2.e'_3 \ e_3.e'_1 & e_3.e'_2 & e_3.e'_3endbmatrix.beginbmatrixy_1\y_2\y_3endbmatrix$$ If we transform the cordinate of the a vector, only the components and basis of vector changes. The vector remains unchanged. Thus the dot product remain unchanged even if we compute dot product between primed and unprimed vectors.






    share|cite|improve this answer











    $endgroup$








    • 1




      $begingroup$
      I like this because it provides a prior motivation for representing inner products with a metric tensor in relativity.
      $endgroup$
      – dmckee
      14 hours ago


















    1












    $begingroup$

    A vector space (or linear space) is a set and two operations, which are vector addition and scalar multiplication, and some rules (spelled out in the Definition section of this Wikipedia article). The net result of this definition is that vectors behave like little arrows or ordered tuples under addition and scalar multiplication.



    This is good, but often more structure is needed. (See the Vector Spaces with Additional Structure section of the link above.)



    For example, a norm can be defined on a vector space. This defines a magnitude or length for each vector. Again there are some rules. No magnitude can be negative. Only the $vec0$ vector can have a magnitude of $0$. The triangle inequality: $lvert(a+b)rvert <= lvert arvert + lvert brvert$



    Likewise an inner product can be defined on a vector space. It adds enough structure to support the ideas of orthogonality and projection. For spaces where it makes sense, this leads to the idea of angle.



    The formal definition or a norm is that an inner product is a function that associates two vectors with a number, with some rules. See this for the details.




    These are general definitions which work on all vector spaces. The links above give examples of vector spaces that may not be familiar. E.G. The set of all functions of the form $y = ax^2 + bx + c$ is a 3 dimensional vector space.



    The most familiar vector spaces are N dimensional Euclidian spaces. These are normed vector spaces, where the norm matches the everyday definition of distance.



    The dot product is the inner product on these spaces that matches the everyday definition of orthogonality and angle. See this Wikipedia article.






    share|cite|improve this answer









    $endgroup$




















      1












      $begingroup$


      How, would I compute their dot product?




      You pretty much have to convert them to the same basis system. You can multiply them out and get nine different terms, and then find the dot product in terms of the nine dot products of the basis vectors, but the math is pretty much the same as converting to the same coordinate system.




      In particular, is there a more formal/abstract/generalized definition of the dot product (that would allow me to compute e1→⋅e′1→ without converting the vectors to the same coordinate system)?




      The value of $vece_1 cdot vece_1'$ is an empirical value. You can't calculate it simply from a definition.




      Even if I did convert the vectors to the same coordinate system, why do we know that the result will be the same if I multiply the components in the primed system versus in the unprimed system?




      Given a physical system in which "length" and "angle" are defined, the dot product is invariant under rotations and reflections, i.e. orthonormal transformations. So given two coordinate systems, as long the axes are orthogonal to each other within each coordinate system, and the two coordinate systems have the same origin and the same scale (one unit is the same length, regardless of which direction or coordinate system), dot products will be the same.



      In that case, the change of basis can be represented with a matrix $U$ such that $(U^*)U=I$ (For real numbers, $U^*$ is just the transpose, so I'll be using that for the rest, since presumably you're asking about vectors over the real numbers). The dot product of two vectors $x$ and $y$ is $x^Ty$. If $x'=UX$ and $y'=Uy$, then the dot product of $x'$ and $y'$ is $x'^Ty'=(Ux)^TUy=x^TU^TUy=x^TIy=x^Ty$






      share|cite|improve this answer









      $endgroup$




















        1












        $begingroup$

        The formula



        $$langle x_1,x_2,x_3rangle cdot langle y_1,y_2,y_3rangle = x_1 y_1 + x_2 y_2 + x_3 y _3$$



        is just a start and, as you go further down in physics, will need quite a few generalizations. The assumptions here are that your vectors are (a) real (b) three-dimensional (c) tuples (d) written in a "standard basis". There are points at which either of these are broken: for example, one of the first things you learn in special theory of relativity(*) is how to work with (b') four-dimensional vectors that (d') don't even allow a standard basis at all, so you get a different formula (of which this is a special case). Similarly, in quantum mechanics, depending on the text, you need to grasp in quantum mechanics are (a') complex vector spaces of (b'') infinite-dimensional things that (c') may not be tuples at all (although often can be written so, again allowing a formula of which this is a special case).



        You just yourself figured out that (d) will not always be the case, and that's a splendid job on your part.



        Before any of those generalizations take place, the assumptions (a - d) are taken for granted. That is, we are working in a basis
        $$e_1 equiv langle 1,0,0 rangle \
        e_2 equiv langle 0,1,0 rangle \
        e_3 equiv langle 0,0,1 rangle$$

        and
        $$e_1 cdot e_1 = 1, e_1 cdot e_2 = 0, e_1 cdot e_3 = 0 textetc.$$
        If a triple of numbers is written it is in this basis. While there are other bases, they just represent concrete triples which you have to multiply by the corresponding coefficients and sum up, effectively transforming to $(e_1, e_2, e_3)$, if you insist on applying the scalar product formula above.



        The generalization to taking vectors not as triples of numbers, but as combinations of some abstract $e'_1$, $e'_2$, $e'_3$, then requires specifying what $e'_i cdot e'_j$ is for all $i$, $j$, as other answers have already said in a plenty of ways. If $(e_i)$ and $(e'_i)$ are two different bases, and you know the scalar product in one, the scalar product in the other can be computed from the relations between the basis vectors. And so can a formula for taking scalar products of two vectors, one in each of the two bases.



        The basic idea remains, though, and it is a good idea to get oneself familiarized with all the aspects of the above as deeply as possible: to understand the relation between scalar product and norm, orthogonality, expression of geometrical properties and relations (length, angle, distance), etc., before things get too abstract. That's why many texts just hold on to the simplest formula as long as they can.




        To actually answer your question: let



        $$vecx = x_1 vece_1 + x_2 vece_2 + x_3 vece_3$$
        $$vecy = y_1 vece_1' + y_2 vece_2' + y_3 vece_3'$$



        such that $(vece_1, vece_2, vece_3)$ is the standard basis. Let further



        $$vece_i' = sum_j=1^3 E_i,j vece_j,$$



        so using distributivity and linearity it holds that



        $$vece_i' cdot vece_k
        = left( sum_j=1^3 E_i,j vece_j right) cdot vece_k
        = sum_j=1^3 E_i,j left( vece_j cdot vece_k right)
        = sum_j=1^3 E_i,j delta_jk (**)
        = E_i,k,$$



        (also $vece_k cdot vece_i' = E_i,k$), so



        $$vecx cdot vecy
        = left( sum_i=1^3 x_i vece_i right) cdot left( sum_j=1^3 y_j vece_j' right)
        = sum_i=1^3 sum_j=1^3 x_i y_j left( vece_i cdot vece_j' right)
        = sum_i=1^3 sum_j=1^3 x_i y_j E_j,i.$$



        You can use this formula for taking dot products of two vertices in different bases.
        I'm not sure if this counts as not converting to the same basis or not: you will need the conversion matrix $(E_i,j)$ anyway. You won't need to explicitly write $vecy$ in the $(vece_i)$ basis beforehand, though.




        (*) Mathematically speaking, special relativity does not use an actual 'scalar product'. But for my example this suffices without further details.



        (**) $delta_jk$ is shorthand for "one when $j=k$ and zero otherwise".






        share|cite|improve this answer











        $endgroup$













          Your Answer








          StackExchange.ready(function()
          var channelOptions =
          tags: "".split(" "),
          id: "151"
          ;
          initTagRenderer("".split(" "), "".split(" "), channelOptions);

          StackExchange.using("externalEditor", function()
          // Have to fire editor after snippets, if snippets enabled
          if (StackExchange.settings.snippets.snippetsEnabled)
          StackExchange.using("snippets", function()
          createEditor();
          );

          else
          createEditor();

          );

          function createEditor()
          StackExchange.prepareEditor(
          heartbeatType: 'answer',
          autoActivateHeartbeat: false,
          convertImagesToLinks: false,
          noModals: true,
          showLowRepImageUploadWarning: true,
          reputationToPostImages: null,
          bindNavPrevention: true,
          postfix: "",
          imageUploader:
          brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
          contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
          allowUrls: true
          ,
          noCode: true, onDemand: true,
          discardSelector: ".discard-answer"
          ,immediatelyShowMarkdownHelp:true
          );



          );













          draft saved

          draft discarded


















          StackExchange.ready(
          function ()
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fphysics.stackexchange.com%2fquestions%2f479656%2fformal-definition-of-dot-product%23new-answer', 'question_page');

          );

          Post as a guest















          Required, but never shown

























          8 Answers
          8






          active

          oldest

          votes








          8 Answers
          8






          active

          oldest

          votes









          active

          oldest

          votes






          active

          oldest

          votes









          16












          $begingroup$

          Your top-line question can be answered at many levels. Setting aside issues of forms and covariant/contravariant, the answer is:




          The dot product is the product of the magnitudes of the two vectors, times the cosine of the angle between them.




          No matter what basis you compute that in, you have to get the same answer because it's a physical quantity.



          The usual "sum of products of orthonormal components" is then a convenient computational approach, but as you've seen it's not the only way to compute them.



          The dot product's properties includes linear, commutative, distributive, etc. So when you expand the dot product



          $$(a_x hatx+a_y haty + a_z hatz) cdot (b_x hatX+b_y hatY + b_z hatZ)$$



          you get nine terms like $( a_x b_x hatxcdothatX) + (a_x b_y hatxcdothatY)+$ etc. In the usual orthonormal basis, the same-axis $hatxcdothatX$ factors just become 1, while the different-axis $hatxcdothatY$ et al factors are zero. That reduces to the formula you know.



          In a non-orthonormal basis, you have to figure out what those basis products are. To do that, you refer back to the definition: The product of the size of each, times the cosine of the angle between. Once you have all of those, you're again all set to compute. It just looks a bit more complicated...






          share|cite|improve this answer











          $endgroup$








          • 13




            $begingroup$
            I don't think the dot product is associative.
            $endgroup$
            – eyeballfrog
            May 13 at 1:18






          • 4




            $begingroup$
            "No matter what basis you compute that in, you have to get the same answer because it's a physical quantity." Only if you have a physical vector. If we're speaking mathematically, vectors can be abstract objects, and the "angle" is not defined. In fact, generally speaking, if "angle" is defined, it's defined in terms of the dot product, making your definition circular.
            $endgroup$
            – Acccumulation
            May 13 at 17:14






          • 3




            $begingroup$
            @Acccumulation This is Physics Stack Exchange.
            $endgroup$
            – Bob Jacobsen
            May 13 at 17:18






          • 4




            $begingroup$
            @Bob Jacobsen Yes, but physics also has abstract Hilbert spaces. Consider, for example, Quantum Mechanics.
            $endgroup$
            – scaphys
            May 13 at 18:36






          • 2




            $begingroup$
            "No matter what basis you compute that in, you have to get the same answer because it's a physical quantity." What about two bases which are not related by an orthogonal transformation? For example $hat e_i = 2 hat f_i$.
            $endgroup$
            – Display Name
            May 13 at 18:54















          16












          $begingroup$

          Your top-line question can be answered at many levels. Setting aside issues of forms and covariant/contravariant, the answer is:




          The dot product is the product of the magnitudes of the two vectors, times the cosine of the angle between them.




          No matter what basis you compute that in, you have to get the same answer because it's a physical quantity.



          The usual "sum of products of orthonormal components" is then a convenient computational approach, but as you've seen it's not the only way to compute them.



          The dot product's properties includes linear, commutative, distributive, etc. So when you expand the dot product



          $$(a_x hatx+a_y haty + a_z hatz) cdot (b_x hatX+b_y hatY + b_z hatZ)$$



          you get nine terms like $( a_x b_x hatxcdothatX) + (a_x b_y hatxcdothatY)+$ etc. In the usual orthonormal basis, the same-axis $hatxcdothatX$ factors just become 1, while the different-axis $hatxcdothatY$ et al factors are zero. That reduces to the formula you know.



          In a non-orthonormal basis, you have to figure out what those basis products are. To do that, you refer back to the definition: The product of the size of each, times the cosine of the angle between. Once you have all of those, you're again all set to compute. It just looks a bit more complicated...






          share|cite|improve this answer











          $endgroup$








          • 13




            $begingroup$
            I don't think the dot product is associative.
            $endgroup$
            – eyeballfrog
            May 13 at 1:18






          • 4




            $begingroup$
            "No matter what basis you compute that in, you have to get the same answer because it's a physical quantity." Only if you have a physical vector. If we're speaking mathematically, vectors can be abstract objects, and the "angle" is not defined. In fact, generally speaking, if "angle" is defined, it's defined in terms of the dot product, making your definition circular.
            $endgroup$
            – Acccumulation
            May 13 at 17:14






          • 3




            $begingroup$
            @Acccumulation This is Physics Stack Exchange.
            $endgroup$
            – Bob Jacobsen
            May 13 at 17:18






          • 4




            $begingroup$
            @Bob Jacobsen Yes, but physics also has abstract Hilbert spaces. Consider, for example, Quantum Mechanics.
            $endgroup$
            – scaphys
            May 13 at 18:36






          • 2




            $begingroup$
            "No matter what basis you compute that in, you have to get the same answer because it's a physical quantity." What about two bases which are not related by an orthogonal transformation? For example $hat e_i = 2 hat f_i$.
            $endgroup$
            – Display Name
            May 13 at 18:54













          16












          16








          16





          $begingroup$

          Your top-line question can be answered at many levels. Setting aside issues of forms and covariant/contravariant, the answer is:




          The dot product is the product of the magnitudes of the two vectors, times the cosine of the angle between them.




          No matter what basis you compute that in, you have to get the same answer because it's a physical quantity.



          The usual "sum of products of orthonormal components" is then a convenient computational approach, but as you've seen it's not the only way to compute them.



          The dot product's properties includes linear, commutative, distributive, etc. So when you expand the dot product



          $$(a_x hatx+a_y haty + a_z hatz) cdot (b_x hatX+b_y hatY + b_z hatZ)$$



          you get nine terms like $( a_x b_x hatxcdothatX) + (a_x b_y hatxcdothatY)+$ etc. In the usual orthonormal basis, the same-axis $hatxcdothatX$ factors just become 1, while the different-axis $hatxcdothatY$ et al factors are zero. That reduces to the formula you know.



          In a non-orthonormal basis, you have to figure out what those basis products are. To do that, you refer back to the definition: The product of the size of each, times the cosine of the angle between. Once you have all of those, you're again all set to compute. It just looks a bit more complicated...






          share|cite|improve this answer











          $endgroup$



          Your top-line question can be answered at many levels. Setting aside issues of forms and covariant/contravariant, the answer is:




          The dot product is the product of the magnitudes of the two vectors, times the cosine of the angle between them.




          No matter what basis you compute that in, you have to get the same answer because it's a physical quantity.



          The usual "sum of products of orthonormal components" is then a convenient computational approach, but as you've seen it's not the only way to compute them.



          The dot product's properties includes linear, commutative, distributive, etc. So when you expand the dot product



          $$(a_x hatx+a_y haty + a_z hatz) cdot (b_x hatX+b_y hatY + b_z hatZ)$$



          you get nine terms like $( a_x b_x hatxcdothatX) + (a_x b_y hatxcdothatY)+$ etc. In the usual orthonormal basis, the same-axis $hatxcdothatX$ factors just become 1, while the different-axis $hatxcdothatY$ et al factors are zero. That reduces to the formula you know.



          In a non-orthonormal basis, you have to figure out what those basis products are. To do that, you refer back to the definition: The product of the size of each, times the cosine of the angle between. Once you have all of those, you're again all set to compute. It just looks a bit more complicated...







          share|cite|improve this answer














          share|cite|improve this answer



          share|cite|improve this answer








          edited May 13 at 1:36

























          answered May 13 at 0:26









          Bob JacobsenBob Jacobsen

          6,0511020




          6,0511020







          • 13




            $begingroup$
            I don't think the dot product is associative.
            $endgroup$
            – eyeballfrog
            May 13 at 1:18






          • 4




            $begingroup$
            "No matter what basis you compute that in, you have to get the same answer because it's a physical quantity." Only if you have a physical vector. If we're speaking mathematically, vectors can be abstract objects, and the "angle" is not defined. In fact, generally speaking, if "angle" is defined, it's defined in terms of the dot product, making your definition circular.
            $endgroup$
            – Acccumulation
            May 13 at 17:14






          • 3




            $begingroup$
            @Acccumulation This is Physics Stack Exchange.
            $endgroup$
            – Bob Jacobsen
            May 13 at 17:18






          • 4




            $begingroup$
            @Bob Jacobsen Yes, but physics also has abstract Hilbert spaces. Consider, for example, Quantum Mechanics.
            $endgroup$
            – scaphys
            May 13 at 18:36






          • 2




            $begingroup$
            "No matter what basis you compute that in, you have to get the same answer because it's a physical quantity." What about two bases which are not related by an orthogonal transformation? For example $hat e_i = 2 hat f_i$.
            $endgroup$
            – Display Name
            May 13 at 18:54












          • 13




            $begingroup$
            I don't think the dot product is associative.
            $endgroup$
            – eyeballfrog
            May 13 at 1:18






          • 4




            $begingroup$
            "No matter what basis you compute that in, you have to get the same answer because it's a physical quantity." Only if you have a physical vector. If we're speaking mathematically, vectors can be abstract objects, and the "angle" is not defined. In fact, generally speaking, if "angle" is defined, it's defined in terms of the dot product, making your definition circular.
            $endgroup$
            – Acccumulation
            May 13 at 17:14






          • 3




            $begingroup$
            @Acccumulation This is Physics Stack Exchange.
            $endgroup$
            – Bob Jacobsen
            May 13 at 17:18






          • 4




            $begingroup$
            @Bob Jacobsen Yes, but physics also has abstract Hilbert spaces. Consider, for example, Quantum Mechanics.
            $endgroup$
            – scaphys
            May 13 at 18:36






          • 2




            $begingroup$
            "No matter what basis you compute that in, you have to get the same answer because it's a physical quantity." What about two bases which are not related by an orthogonal transformation? For example $hat e_i = 2 hat f_i$.
            $endgroup$
            – Display Name
            May 13 at 18:54







          13




          13




          $begingroup$
          I don't think the dot product is associative.
          $endgroup$
          – eyeballfrog
          May 13 at 1:18




          $begingroup$
          I don't think the dot product is associative.
          $endgroup$
          – eyeballfrog
          May 13 at 1:18




          4




          4




          $begingroup$
          "No matter what basis you compute that in, you have to get the same answer because it's a physical quantity." Only if you have a physical vector. If we're speaking mathematically, vectors can be abstract objects, and the "angle" is not defined. In fact, generally speaking, if "angle" is defined, it's defined in terms of the dot product, making your definition circular.
          $endgroup$
          – Acccumulation
          May 13 at 17:14




          $begingroup$
          "No matter what basis you compute that in, you have to get the same answer because it's a physical quantity." Only if you have a physical vector. If we're speaking mathematically, vectors can be abstract objects, and the "angle" is not defined. In fact, generally speaking, if "angle" is defined, it's defined in terms of the dot product, making your definition circular.
          $endgroup$
          – Acccumulation
          May 13 at 17:14




          3




          3




          $begingroup$
          @Acccumulation This is Physics Stack Exchange.
          $endgroup$
          – Bob Jacobsen
          May 13 at 17:18




          $begingroup$
          @Acccumulation This is Physics Stack Exchange.
          $endgroup$
          – Bob Jacobsen
          May 13 at 17:18




          4




          4




          $begingroup$
          @Bob Jacobsen Yes, but physics also has abstract Hilbert spaces. Consider, for example, Quantum Mechanics.
          $endgroup$
          – scaphys
          May 13 at 18:36




          $begingroup$
          @Bob Jacobsen Yes, but physics also has abstract Hilbert spaces. Consider, for example, Quantum Mechanics.
          $endgroup$
          – scaphys
          May 13 at 18:36




          2




          2




          $begingroup$
          "No matter what basis you compute that in, you have to get the same answer because it's a physical quantity." What about two bases which are not related by an orthogonal transformation? For example $hat e_i = 2 hat f_i$.
          $endgroup$
          – Display Name
          May 13 at 18:54




          $begingroup$
          "No matter what basis you compute that in, you have to get the same answer because it's a physical quantity." What about two bases which are not related by an orthogonal transformation? For example $hat e_i = 2 hat f_i$.
          $endgroup$
          – Display Name
          May 13 at 18:54











          18












          $begingroup$

          Dot products, or inner products are defined axiomatically, or abstractly. An inner product on a vector space $V$ over $R$ is a pairing $Vtimes Vto R$, denoted by $ langle u,vrangle$, with properties $langle u,vrangle=langle v,urangle$, $langle u+cw,vrangle= langle u,vrangle+clangle w,vrangle$, and $ langle u,uranglegt0$ if $une0$. In general, a vector space can be endowed with an inner product in many ways. Notice here there is no reference to a basis/coordinate system.



          Using what is called the Gram-Schmidt process, one can then construct a basis $e_1,cdots e_n$ for $V$ in which the inner product takes the computational form which you stated in your question.



          In your question, you are actually starting with what is called an orthonormal basis for an inner product. The coordinate-free approach is to state the postulates that an inner product should obey, then after being given an explicit inner product, construct an orthonormal basis in which to do computations.



          In general, an orthonormal basis $e_1,e_2,e_3$ for one inner product on $V$ will not be an orthonormal basis for another inner product on $V$.






          share|cite|improve this answer









          $endgroup$

















            18












            $begingroup$

            Dot products, or inner products are defined axiomatically, or abstractly. An inner product on a vector space $V$ over $R$ is a pairing $Vtimes Vto R$, denoted by $ langle u,vrangle$, with properties $langle u,vrangle=langle v,urangle$, $langle u+cw,vrangle= langle u,vrangle+clangle w,vrangle$, and $ langle u,uranglegt0$ if $une0$. In general, a vector space can be endowed with an inner product in many ways. Notice here there is no reference to a basis/coordinate system.



            Using what is called the Gram-Schmidt process, one can then construct a basis $e_1,cdots e_n$ for $V$ in which the inner product takes the computational form which you stated in your question.



            In your question, you are actually starting with what is called an orthonormal basis for an inner product. The coordinate-free approach is to state the postulates that an inner product should obey, then after being given an explicit inner product, construct an orthonormal basis in which to do computations.



            In general, an orthonormal basis $e_1,e_2,e_3$ for one inner product on $V$ will not be an orthonormal basis for another inner product on $V$.






            share|cite|improve this answer









            $endgroup$















              18












              18








              18





              $begingroup$

              Dot products, or inner products are defined axiomatically, or abstractly. An inner product on a vector space $V$ over $R$ is a pairing $Vtimes Vto R$, denoted by $ langle u,vrangle$, with properties $langle u,vrangle=langle v,urangle$, $langle u+cw,vrangle= langle u,vrangle+clangle w,vrangle$, and $ langle u,uranglegt0$ if $une0$. In general, a vector space can be endowed with an inner product in many ways. Notice here there is no reference to a basis/coordinate system.



              Using what is called the Gram-Schmidt process, one can then construct a basis $e_1,cdots e_n$ for $V$ in which the inner product takes the computational form which you stated in your question.



              In your question, you are actually starting with what is called an orthonormal basis for an inner product. The coordinate-free approach is to state the postulates that an inner product should obey, then after being given an explicit inner product, construct an orthonormal basis in which to do computations.



              In general, an orthonormal basis $e_1,e_2,e_3$ for one inner product on $V$ will not be an orthonormal basis for another inner product on $V$.






              share|cite|improve this answer









              $endgroup$



              Dot products, or inner products are defined axiomatically, or abstractly. An inner product on a vector space $V$ over $R$ is a pairing $Vtimes Vto R$, denoted by $ langle u,vrangle$, with properties $langle u,vrangle=langle v,urangle$, $langle u+cw,vrangle= langle u,vrangle+clangle w,vrangle$, and $ langle u,uranglegt0$ if $une0$. In general, a vector space can be endowed with an inner product in many ways. Notice here there is no reference to a basis/coordinate system.



              Using what is called the Gram-Schmidt process, one can then construct a basis $e_1,cdots e_n$ for $V$ in which the inner product takes the computational form which you stated in your question.



              In your question, you are actually starting with what is called an orthonormal basis for an inner product. The coordinate-free approach is to state the postulates that an inner product should obey, then after being given an explicit inner product, construct an orthonormal basis in which to do computations.



              In general, an orthonormal basis $e_1,e_2,e_3$ for one inner product on $V$ will not be an orthonormal basis for another inner product on $V$.







              share|cite|improve this answer












              share|cite|improve this answer



              share|cite|improve this answer










              answered May 13 at 2:32









              user52817user52817

              3312




              3312





















                  9












                  $begingroup$

                  The dot product can be defined in a coordinate-independent way as



                  $$vecacdotvecb=|veca||vecb|costheta$$



                  where $theta$ is the angle between the two vectors. This involves only lengths and angles, not coordinates.



                  To use your first formula, the coordinates must be in the same basis.



                  You can convert between bases using a rotation matrix, and the fact that a rotation matrix preserves vector lengths is sufficient to show that it preserves the dot product. This is because



                  $$vecacdotvecb=frac12left(|veca+vecb|^2-|veca|^2-|vecb|^2right).$$



                  This formula is another purely-geometric, coordinate-free definition of the dot product.






                  share|cite|improve this answer











                  $endgroup$












                  • $begingroup$
                    Thank you! That makes sense. But what happens if you are dealing with a non-orthonormal system? Is the dot product's value preserved in making the coordinate transformation?
                    $endgroup$
                    – dts
                    May 13 at 0:21










                  • $begingroup$
                    Yes, the value is preserved, but the coordinate-based formula in a non-orthonormal basis is more complicated than your first formula.
                    $endgroup$
                    – G. Smith
                    May 13 at 0:31






                  • 1




                    $begingroup$
                    "You can convert between bases using a rotation matrix", I strongly disagree. Only if the base vectors are normalised, but that needn't be the case. However there exists a Matrix $A$ such that $e_i^prime = A e_i$ where $e_i$ is to be understood at the ith basic vector (not the component).
                    $endgroup$
                    – infinitezero
                    May 13 at 16:33
















                  9












                  $begingroup$

                  The dot product can be defined in a coordinate-independent way as



                  $$vecacdotvecb=|veca||vecb|costheta$$



                  where $theta$ is the angle between the two vectors. This involves only lengths and angles, not coordinates.



                  To use your first formula, the coordinates must be in the same basis.



                  You can convert between bases using a rotation matrix, and the fact that a rotation matrix preserves vector lengths is sufficient to show that it preserves the dot product. This is because



                  $$vecacdotvecb=frac12left(|veca+vecb|^2-|veca|^2-|vecb|^2right).$$



                  This formula is another purely-geometric, coordinate-free definition of the dot product.






                  share|cite|improve this answer











                  $endgroup$












                  • $begingroup$
                    Thank you! That makes sense. But what happens if you are dealing with a non-orthonormal system? Is the dot product's value preserved in making the coordinate transformation?
                    $endgroup$
                    – dts
                    May 13 at 0:21










                  • $begingroup$
                    Yes, the value is preserved, but the coordinate-based formula in a non-orthonormal basis is more complicated than your first formula.
                    $endgroup$
                    – G. Smith
                    May 13 at 0:31






                  • 1




                    $begingroup$
                    "You can convert between bases using a rotation matrix", I strongly disagree. Only if the base vectors are normalised, but that needn't be the case. However there exists a Matrix $A$ such that $e_i^prime = A e_i$ where $e_i$ is to be understood at the ith basic vector (not the component).
                    $endgroup$
                    – infinitezero
                    May 13 at 16:33














                  9












                  9








                  9





                  $begingroup$

                  The dot product can be defined in a coordinate-independent way as



                  $$vecacdotvecb=|veca||vecb|costheta$$



                  where $theta$ is the angle between the two vectors. This involves only lengths and angles, not coordinates.



                  To use your first formula, the coordinates must be in the same basis.



                  You can convert between bases using a rotation matrix, and the fact that a rotation matrix preserves vector lengths is sufficient to show that it preserves the dot product. This is because



                  $$vecacdotvecb=frac12left(|veca+vecb|^2-|veca|^2-|vecb|^2right).$$



                  This formula is another purely-geometric, coordinate-free definition of the dot product.






                  share|cite|improve this answer











                  $endgroup$



                  The dot product can be defined in a coordinate-independent way as



                  $$vecacdotvecb=|veca||vecb|costheta$$



                  where $theta$ is the angle between the two vectors. This involves only lengths and angles, not coordinates.



                  To use your first formula, the coordinates must be in the same basis.



                  You can convert between bases using a rotation matrix, and the fact that a rotation matrix preserves vector lengths is sufficient to show that it preserves the dot product. This is because



                  $$vecacdotvecb=frac12left(|veca+vecb|^2-|veca|^2-|vecb|^2right).$$



                  This formula is another purely-geometric, coordinate-free definition of the dot product.







                  share|cite|improve this answer














                  share|cite|improve this answer



                  share|cite|improve this answer








                  edited May 13 at 0:23

























                  answered May 13 at 0:14









                  G. SmithG. Smith

                  13.1k12145




                  13.1k12145











                  • $begingroup$
                    Thank you! That makes sense. But what happens if you are dealing with a non-orthonormal system? Is the dot product's value preserved in making the coordinate transformation?
                    $endgroup$
                    – dts
                    May 13 at 0:21










                  • $begingroup$
                    Yes, the value is preserved, but the coordinate-based formula in a non-orthonormal basis is more complicated than your first formula.
                    $endgroup$
                    – G. Smith
                    May 13 at 0:31






                  • 1




                    $begingroup$
                    "You can convert between bases using a rotation matrix", I strongly disagree. Only if the base vectors are normalised, but that needn't be the case. However there exists a Matrix $A$ such that $e_i^prime = A e_i$ where $e_i$ is to be understood at the ith basic vector (not the component).
                    $endgroup$
                    – infinitezero
                    May 13 at 16:33

















                  • $begingroup$
                    Thank you! That makes sense. But what happens if you are dealing with a non-orthonormal system? Is the dot product's value preserved in making the coordinate transformation?
                    $endgroup$
                    – dts
                    May 13 at 0:21










                  • $begingroup$
                    Yes, the value is preserved, but the coordinate-based formula in a non-orthonormal basis is more complicated than your first formula.
                    $endgroup$
                    – G. Smith
                    May 13 at 0:31






                  • 1




                    $begingroup$
                    "You can convert between bases using a rotation matrix", I strongly disagree. Only if the base vectors are normalised, but that needn't be the case. However there exists a Matrix $A$ such that $e_i^prime = A e_i$ where $e_i$ is to be understood at the ith basic vector (not the component).
                    $endgroup$
                    – infinitezero
                    May 13 at 16:33
















                  $begingroup$
                  Thank you! That makes sense. But what happens if you are dealing with a non-orthonormal system? Is the dot product's value preserved in making the coordinate transformation?
                  $endgroup$
                  – dts
                  May 13 at 0:21




                  $begingroup$
                  Thank you! That makes sense. But what happens if you are dealing with a non-orthonormal system? Is the dot product's value preserved in making the coordinate transformation?
                  $endgroup$
                  – dts
                  May 13 at 0:21












                  $begingroup$
                  Yes, the value is preserved, but the coordinate-based formula in a non-orthonormal basis is more complicated than your first formula.
                  $endgroup$
                  – G. Smith
                  May 13 at 0:31




                  $begingroup$
                  Yes, the value is preserved, but the coordinate-based formula in a non-orthonormal basis is more complicated than your first formula.
                  $endgroup$
                  – G. Smith
                  May 13 at 0:31




                  1




                  1




                  $begingroup$
                  "You can convert between bases using a rotation matrix", I strongly disagree. Only if the base vectors are normalised, but that needn't be the case. However there exists a Matrix $A$ such that $e_i^prime = A e_i$ where $e_i$ is to be understood at the ith basic vector (not the component).
                  $endgroup$
                  – infinitezero
                  May 13 at 16:33





                  $begingroup$
                  "You can convert between bases using a rotation matrix", I strongly disagree. Only if the base vectors are normalised, but that needn't be the case. However there exists a Matrix $A$ such that $e_i^prime = A e_i$ where $e_i$ is to be understood at the ith basic vector (not the component).
                  $endgroup$
                  – infinitezero
                  May 13 at 16:33












                  7












                  $begingroup$

                  The coordinate free definition of a dot product is:



                  $$ vec a cdot vec b = frac 1 4 [(vec a + vec b)^2 - (vec a - vec b)^2]$$



                  It's up to you to figure out what the norm is:



                  $$ ||vec a|| = sqrt(vec a)^2$$



                  Here is a reference for this viewpoint:
                  http://www.pmaweb.caltech.edu/Courses/ph136/yr2012/1202.1.K.pdf
                  Section 2.3






                  share|cite|improve this answer











                  $endgroup$








                  • 3




                    $begingroup$
                    This is a circular definition as the norm is defined via the dot product.
                    $endgroup$
                    – Winther
                    May 13 at 9:01










                  • $begingroup$
                    @Winther You've got to input something: the dot product cannot be derived only from the underlying vector space structure. The norm seems a reasonable choice here, for geometric intuition.
                    $endgroup$
                    – Denis Nardin
                    May 13 at 11:06






                  • 2




                    $begingroup$
                    This will only define an inner product iff the norm satisfies the parallelogram identity $2||x||^2+2||y||^2=||x+y||^2+||x-y||^2$
                    $endgroup$
                    – Jannik Pitt
                    May 13 at 11:50










                  • $begingroup$
                    Yes you have to input something: either define a norm or define an inner product and have the norm be induced by this. However my point was that you seem to define the norm via $|a|=sqrtacdot a$ which is why I said it was circular. On second reading it does look like you say you need to specify the norm externally so then this would be fine. However doesn't then the definition of the norm require you to specify a coordinate system so it's not really coordinate free?
                    $endgroup$
                    – Winther
                    May 13 at 13:48










                  • $begingroup$
                    @Winther Well, it depends on how your vector space is given to you. If your vectors are a bunch of coordinates (like in the usual description of $mathbbR^n$), of course every definition you give will be coordinate dependent (coordinates are all you have!), but if your vector space is composed by something more exotic (e.g. the space of solutions of a certain ODE) then you can hope to write down a definition of the norm using something else. (and yes, indeed a Banach space is Hilbert iff the norm satisfies the parallelogram identity, plus some added condition if over $mathbbC$).
                    $endgroup$
                    – Denis Nardin
                    May 13 at 18:10















                  7












                  $begingroup$

                  The coordinate free definition of a dot product is:



                  $$ vec a cdot vec b = frac 1 4 [(vec a + vec b)^2 - (vec a - vec b)^2]$$



                  It's up to you to figure out what the norm is:



                  $$ ||vec a|| = sqrt(vec a)^2$$



                  Here is a reference for this viewpoint:
                  http://www.pmaweb.caltech.edu/Courses/ph136/yr2012/1202.1.K.pdf
                  Section 2.3






                  share|cite|improve this answer











                  $endgroup$








                  • 3




                    $begingroup$
                    This is a circular definition as the norm is defined via the dot product.
                    $endgroup$
                    – Winther
                    May 13 at 9:01










                  • $begingroup$
                    @Winther You've got to input something: the dot product cannot be derived only from the underlying vector space structure. The norm seems a reasonable choice here, for geometric intuition.
                    $endgroup$
                    – Denis Nardin
                    May 13 at 11:06






                  • 2




                    $begingroup$
                    This will only define an inner product iff the norm satisfies the parallelogram identity $2||x||^2+2||y||^2=||x+y||^2+||x-y||^2$
                    $endgroup$
                    – Jannik Pitt
                    May 13 at 11:50










                  • $begingroup$
                    Yes you have to input something: either define a norm or define an inner product and have the norm be induced by this. However my point was that you seem to define the norm via $|a|=sqrtacdot a$ which is why I said it was circular. On second reading it does look like you say you need to specify the norm externally so then this would be fine. However doesn't then the definition of the norm require you to specify a coordinate system so it's not really coordinate free?
                    $endgroup$
                    – Winther
                    May 13 at 13:48










                  • $begingroup$
                    @Winther Well, it depends on how your vector space is given to you. If your vectors are a bunch of coordinates (like in the usual description of $mathbbR^n$), of course every definition you give will be coordinate dependent (coordinates are all you have!), but if your vector space is composed by something more exotic (e.g. the space of solutions of a certain ODE) then you can hope to write down a definition of the norm using something else. (and yes, indeed a Banach space is Hilbert iff the norm satisfies the parallelogram identity, plus some added condition if over $mathbbC$).
                    $endgroup$
                    – Denis Nardin
                    May 13 at 18:10













                  7












                  7








                  7





                  $begingroup$

                  The coordinate free definition of a dot product is:



                  $$ vec a cdot vec b = frac 1 4 [(vec a + vec b)^2 - (vec a - vec b)^2]$$



                  It's up to you to figure out what the norm is:



                  $$ ||vec a|| = sqrt(vec a)^2$$



                  Here is a reference for this viewpoint:
                  http://www.pmaweb.caltech.edu/Courses/ph136/yr2012/1202.1.K.pdf
                  Section 2.3






                  share|cite|improve this answer











                  $endgroup$



                  The coordinate free definition of a dot product is:



                  $$ vec a cdot vec b = frac 1 4 [(vec a + vec b)^2 - (vec a - vec b)^2]$$



                  It's up to you to figure out what the norm is:



                  $$ ||vec a|| = sqrt(vec a)^2$$



                  Here is a reference for this viewpoint:
                  http://www.pmaweb.caltech.edu/Courses/ph136/yr2012/1202.1.K.pdf
                  Section 2.3







                  share|cite|improve this answer














                  share|cite|improve this answer



                  share|cite|improve this answer








                  edited May 13 at 3:50

























                  answered May 13 at 1:30









                  JEBJEB

                  7,0031819




                  7,0031819







                  • 3




                    $begingroup$
                    This is a circular definition as the norm is defined via the dot product.
                    $endgroup$
                    – Winther
                    May 13 at 9:01










                  • $begingroup$
                    @Winther You've got to input something: the dot product cannot be derived only from the underlying vector space structure. The norm seems a reasonable choice here, for geometric intuition.
                    $endgroup$
                    – Denis Nardin
                    May 13 at 11:06






                  • 2




                    $begingroup$
                    This will only define an inner product iff the norm satisfies the parallelogram identity $2||x||^2+2||y||^2=||x+y||^2+||x-y||^2$
                    $endgroup$
                    – Jannik Pitt
                    May 13 at 11:50










                  • $begingroup$
                    Yes you have to input something: either define a norm or define an inner product and have the norm be induced by this. However my point was that you seem to define the norm via $|a|=sqrtacdot a$ which is why I said it was circular. On second reading it does look like you say you need to specify the norm externally so then this would be fine. However doesn't then the definition of the norm require you to specify a coordinate system so it's not really coordinate free?
                    $endgroup$
                    – Winther
                    May 13 at 13:48










                  • $begingroup$
                    @Winther Well, it depends on how your vector space is given to you. If your vectors are a bunch of coordinates (like in the usual description of $mathbbR^n$), of course every definition you give will be coordinate dependent (coordinates are all you have!), but if your vector space is composed by something more exotic (e.g. the space of solutions of a certain ODE) then you can hope to write down a definition of the norm using something else. (and yes, indeed a Banach space is Hilbert iff the norm satisfies the parallelogram identity, plus some added condition if over $mathbbC$).
                    $endgroup$
                    – Denis Nardin
                    May 13 at 18:10












                  • 3




                    $begingroup$
                    This is a circular definition as the norm is defined via the dot product.
                    $endgroup$
                    – Winther
                    May 13 at 9:01










                  • $begingroup$
                    @Winther You've got to input something: the dot product cannot be derived only from the underlying vector space structure. The norm seems a reasonable choice here, for geometric intuition.
                    $endgroup$
                    – Denis Nardin
                    May 13 at 11:06






                  • 2




                    $begingroup$
                    This will only define an inner product iff the norm satisfies the parallelogram identity $2||x||^2+2||y||^2=||x+y||^2+||x-y||^2$
                    $endgroup$
                    – Jannik Pitt
                    May 13 at 11:50










                  • $begingroup$
                    Yes you have to input something: either define a norm or define an inner product and have the norm be induced by this. However my point was that you seem to define the norm via $|a|=sqrtacdot a$ which is why I said it was circular. On second reading it does look like you say you need to specify the norm externally so then this would be fine. However doesn't then the definition of the norm require you to specify a coordinate system so it's not really coordinate free?
                    $endgroup$
                    – Winther
                    May 13 at 13:48










                  • $begingroup$
                    @Winther Well, it depends on how your vector space is given to you. If your vectors are a bunch of coordinates (like in the usual description of $mathbbR^n$), of course every definition you give will be coordinate dependent (coordinates are all you have!), but if your vector space is composed by something more exotic (e.g. the space of solutions of a certain ODE) then you can hope to write down a definition of the norm using something else. (and yes, indeed a Banach space is Hilbert iff the norm satisfies the parallelogram identity, plus some added condition if over $mathbbC$).
                    $endgroup$
                    – Denis Nardin
                    May 13 at 18:10







                  3




                  3




                  $begingroup$
                  This is a circular definition as the norm is defined via the dot product.
                  $endgroup$
                  – Winther
                  May 13 at 9:01




                  $begingroup$
                  This is a circular definition as the norm is defined via the dot product.
                  $endgroup$
                  – Winther
                  May 13 at 9:01












                  $begingroup$
                  @Winther You've got to input something: the dot product cannot be derived only from the underlying vector space structure. The norm seems a reasonable choice here, for geometric intuition.
                  $endgroup$
                  – Denis Nardin
                  May 13 at 11:06




                  $begingroup$
                  @Winther You've got to input something: the dot product cannot be derived only from the underlying vector space structure. The norm seems a reasonable choice here, for geometric intuition.
                  $endgroup$
                  – Denis Nardin
                  May 13 at 11:06




                  2




                  2




                  $begingroup$
                  This will only define an inner product iff the norm satisfies the parallelogram identity $2||x||^2+2||y||^2=||x+y||^2+||x-y||^2$
                  $endgroup$
                  – Jannik Pitt
                  May 13 at 11:50




                  $begingroup$
                  This will only define an inner product iff the norm satisfies the parallelogram identity $2||x||^2+2||y||^2=||x+y||^2+||x-y||^2$
                  $endgroup$
                  – Jannik Pitt
                  May 13 at 11:50












                  $begingroup$
                  Yes you have to input something: either define a norm or define an inner product and have the norm be induced by this. However my point was that you seem to define the norm via $|a|=sqrtacdot a$ which is why I said it was circular. On second reading it does look like you say you need to specify the norm externally so then this would be fine. However doesn't then the definition of the norm require you to specify a coordinate system so it's not really coordinate free?
                  $endgroup$
                  – Winther
                  May 13 at 13:48




                  $begingroup$
                  Yes you have to input something: either define a norm or define an inner product and have the norm be induced by this. However my point was that you seem to define the norm via $|a|=sqrtacdot a$ which is why I said it was circular. On second reading it does look like you say you need to specify the norm externally so then this would be fine. However doesn't then the definition of the norm require you to specify a coordinate system so it's not really coordinate free?
                  $endgroup$
                  – Winther
                  May 13 at 13:48












                  $begingroup$
                  @Winther Well, it depends on how your vector space is given to you. If your vectors are a bunch of coordinates (like in the usual description of $mathbbR^n$), of course every definition you give will be coordinate dependent (coordinates are all you have!), but if your vector space is composed by something more exotic (e.g. the space of solutions of a certain ODE) then you can hope to write down a definition of the norm using something else. (and yes, indeed a Banach space is Hilbert iff the norm satisfies the parallelogram identity, plus some added condition if over $mathbbC$).
                  $endgroup$
                  – Denis Nardin
                  May 13 at 18:10




                  $begingroup$
                  @Winther Well, it depends on how your vector space is given to you. If your vectors are a bunch of coordinates (like in the usual description of $mathbbR^n$), of course every definition you give will be coordinate dependent (coordinates are all you have!), but if your vector space is composed by something more exotic (e.g. the space of solutions of a certain ODE) then you can hope to write down a definition of the norm using something else. (and yes, indeed a Banach space is Hilbert iff the norm satisfies the parallelogram identity, plus some added condition if over $mathbbC$).
                  $endgroup$
                  – Denis Nardin
                  May 13 at 18:10











                  3












                  $begingroup$

                  On computing the following matrix will give you the dot product $$beginbmatrix x_1 & x_2& x_3 endbmatrix.beginbmatrix e_1.e'_1 & e_1.e'_2 & e_1.e'_3 \ e_2.e'_1 & e_2.e'_2 & e_2.e'_3 \ e_3.e'_1 & e_3.e'_2 & e_3.e'_3endbmatrix.beginbmatrixy_1\y_2\y_3endbmatrix$$ If we transform the cordinate of the a vector, only the components and basis of vector changes. The vector remains unchanged. Thus the dot product remain unchanged even if we compute dot product between primed and unprimed vectors.






                  share|cite|improve this answer











                  $endgroup$








                  • 1




                    $begingroup$
                    I like this because it provides a prior motivation for representing inner products with a metric tensor in relativity.
                    $endgroup$
                    – dmckee
                    14 hours ago















                  3












                  $begingroup$

                  On computing the following matrix will give you the dot product $$beginbmatrix x_1 & x_2& x_3 endbmatrix.beginbmatrix e_1.e'_1 & e_1.e'_2 & e_1.e'_3 \ e_2.e'_1 & e_2.e'_2 & e_2.e'_3 \ e_3.e'_1 & e_3.e'_2 & e_3.e'_3endbmatrix.beginbmatrixy_1\y_2\y_3endbmatrix$$ If we transform the cordinate of the a vector, only the components and basis of vector changes. The vector remains unchanged. Thus the dot product remain unchanged even if we compute dot product between primed and unprimed vectors.






                  share|cite|improve this answer











                  $endgroup$








                  • 1




                    $begingroup$
                    I like this because it provides a prior motivation for representing inner products with a metric tensor in relativity.
                    $endgroup$
                    – dmckee
                    14 hours ago













                  3












                  3








                  3





                  $begingroup$

                  On computing the following matrix will give you the dot product $$beginbmatrix x_1 & x_2& x_3 endbmatrix.beginbmatrix e_1.e'_1 & e_1.e'_2 & e_1.e'_3 \ e_2.e'_1 & e_2.e'_2 & e_2.e'_3 \ e_3.e'_1 & e_3.e'_2 & e_3.e'_3endbmatrix.beginbmatrixy_1\y_2\y_3endbmatrix$$ If we transform the cordinate of the a vector, only the components and basis of vector changes. The vector remains unchanged. Thus the dot product remain unchanged even if we compute dot product between primed and unprimed vectors.






                  share|cite|improve this answer











                  $endgroup$



                  On computing the following matrix will give you the dot product $$beginbmatrix x_1 & x_2& x_3 endbmatrix.beginbmatrix e_1.e'_1 & e_1.e'_2 & e_1.e'_3 \ e_2.e'_1 & e_2.e'_2 & e_2.e'_3 \ e_3.e'_1 & e_3.e'_2 & e_3.e'_3endbmatrix.beginbmatrixy_1\y_2\y_3endbmatrix$$ If we transform the cordinate of the a vector, only the components and basis of vector changes. The vector remains unchanged. Thus the dot product remain unchanged even if we compute dot product between primed and unprimed vectors.







                  share|cite|improve this answer














                  share|cite|improve this answer



                  share|cite|improve this answer








                  edited 4 hours ago

























                  answered May 13 at 1:13









                  walber97walber97

                  403110




                  403110







                  • 1




                    $begingroup$
                    I like this because it provides a prior motivation for representing inner products with a metric tensor in relativity.
                    $endgroup$
                    – dmckee
                    14 hours ago












                  • 1




                    $begingroup$
                    I like this because it provides a prior motivation for representing inner products with a metric tensor in relativity.
                    $endgroup$
                    – dmckee
                    14 hours ago







                  1




                  1




                  $begingroup$
                  I like this because it provides a prior motivation for representing inner products with a metric tensor in relativity.
                  $endgroup$
                  – dmckee
                  14 hours ago




                  $begingroup$
                  I like this because it provides a prior motivation for representing inner products with a metric tensor in relativity.
                  $endgroup$
                  – dmckee
                  14 hours ago











                  1












                  $begingroup$

                  A vector space (or linear space) is a set and two operations, which are vector addition and scalar multiplication, and some rules (spelled out in the Definition section of this Wikipedia article). The net result of this definition is that vectors behave like little arrows or ordered tuples under addition and scalar multiplication.



                  This is good, but often more structure is needed. (See the Vector Spaces with Additional Structure section of the link above.)



                  For example, a norm can be defined on a vector space. This defines a magnitude or length for each vector. Again there are some rules. No magnitude can be negative. Only the $vec0$ vector can have a magnitude of $0$. The triangle inequality: $lvert(a+b)rvert <= lvert arvert + lvert brvert$



                  Likewise an inner product can be defined on a vector space. It adds enough structure to support the ideas of orthogonality and projection. For spaces where it makes sense, this leads to the idea of angle.



                  The formal definition or a norm is that an inner product is a function that associates two vectors with a number, with some rules. See this for the details.




                  These are general definitions which work on all vector spaces. The links above give examples of vector spaces that may not be familiar. E.G. The set of all functions of the form $y = ax^2 + bx + c$ is a 3 dimensional vector space.



                  The most familiar vector spaces are N dimensional Euclidian spaces. These are normed vector spaces, where the norm matches the everyday definition of distance.



                  The dot product is the inner product on these spaces that matches the everyday definition of orthogonality and angle. See this Wikipedia article.






                  share|cite|improve this answer









                  $endgroup$

















                    1












                    $begingroup$

                    A vector space (or linear space) is a set and two operations, which are vector addition and scalar multiplication, and some rules (spelled out in the Definition section of this Wikipedia article). The net result of this definition is that vectors behave like little arrows or ordered tuples under addition and scalar multiplication.



                    This is good, but often more structure is needed. (See the Vector Spaces with Additional Structure section of the link above.)



                    For example, a norm can be defined on a vector space. This defines a magnitude or length for each vector. Again there are some rules. No magnitude can be negative. Only the $vec0$ vector can have a magnitude of $0$. The triangle inequality: $lvert(a+b)rvert <= lvert arvert + lvert brvert$



                    Likewise an inner product can be defined on a vector space. It adds enough structure to support the ideas of orthogonality and projection. For spaces where it makes sense, this leads to the idea of angle.



                    The formal definition or a norm is that an inner product is a function that associates two vectors with a number, with some rules. See this for the details.




                    These are general definitions which work on all vector spaces. The links above give examples of vector spaces that may not be familiar. E.G. The set of all functions of the form $y = ax^2 + bx + c$ is a 3 dimensional vector space.



                    The most familiar vector spaces are N dimensional Euclidian spaces. These are normed vector spaces, where the norm matches the everyday definition of distance.



                    The dot product is the inner product on these spaces that matches the everyday definition of orthogonality and angle. See this Wikipedia article.






                    share|cite|improve this answer









                    $endgroup$















                      1












                      1








                      1





                      $begingroup$

                      A vector space (or linear space) is a set and two operations, which are vector addition and scalar multiplication, and some rules (spelled out in the Definition section of this Wikipedia article). The net result of this definition is that vectors behave like little arrows or ordered tuples under addition and scalar multiplication.



                      This is good, but often more structure is needed. (See the Vector Spaces with Additional Structure section of the link above.)



                      For example, a norm can be defined on a vector space. This defines a magnitude or length for each vector. Again there are some rules. No magnitude can be negative. Only the $vec0$ vector can have a magnitude of $0$. The triangle inequality: $lvert(a+b)rvert <= lvert arvert + lvert brvert$



                      Likewise an inner product can be defined on a vector space. It adds enough structure to support the ideas of orthogonality and projection. For spaces where it makes sense, this leads to the idea of angle.



                      The formal definition or a norm is that an inner product is a function that associates two vectors with a number, with some rules. See this for the details.




                      These are general definitions which work on all vector spaces. The links above give examples of vector spaces that may not be familiar. E.G. The set of all functions of the form $y = ax^2 + bx + c$ is a 3 dimensional vector space.



                      The most familiar vector spaces are N dimensional Euclidian spaces. These are normed vector spaces, where the norm matches the everyday definition of distance.



                      The dot product is the inner product on these spaces that matches the everyday definition of orthogonality and angle. See this Wikipedia article.






                      share|cite|improve this answer









                      $endgroup$



                      A vector space (or linear space) is a set and two operations, which are vector addition and scalar multiplication, and some rules (spelled out in the Definition section of this Wikipedia article). The net result of this definition is that vectors behave like little arrows or ordered tuples under addition and scalar multiplication.



                      This is good, but often more structure is needed. (See the Vector Spaces with Additional Structure section of the link above.)



                      For example, a norm can be defined on a vector space. This defines a magnitude or length for each vector. Again there are some rules. No magnitude can be negative. Only the $vec0$ vector can have a magnitude of $0$. The triangle inequality: $lvert(a+b)rvert <= lvert arvert + lvert brvert$



                      Likewise an inner product can be defined on a vector space. It adds enough structure to support the ideas of orthogonality and projection. For spaces where it makes sense, this leads to the idea of angle.



                      The formal definition or a norm is that an inner product is a function that associates two vectors with a number, with some rules. See this for the details.




                      These are general definitions which work on all vector spaces. The links above give examples of vector spaces that may not be familiar. E.G. The set of all functions of the form $y = ax^2 + bx + c$ is a 3 dimensional vector space.



                      The most familiar vector spaces are N dimensional Euclidian spaces. These are normed vector spaces, where the norm matches the everyday definition of distance.



                      The dot product is the inner product on these spaces that matches the everyday definition of orthogonality and angle. See this Wikipedia article.







                      share|cite|improve this answer












                      share|cite|improve this answer



                      share|cite|improve this answer










                      answered May 13 at 3:03









                      mmesser314mmesser314

                      9,66521834




                      9,66521834





















                          1












                          $begingroup$


                          How, would I compute their dot product?




                          You pretty much have to convert them to the same basis system. You can multiply them out and get nine different terms, and then find the dot product in terms of the nine dot products of the basis vectors, but the math is pretty much the same as converting to the same coordinate system.




                          In particular, is there a more formal/abstract/generalized definition of the dot product (that would allow me to compute e1→⋅e′1→ without converting the vectors to the same coordinate system)?




                          The value of $vece_1 cdot vece_1'$ is an empirical value. You can't calculate it simply from a definition.




                          Even if I did convert the vectors to the same coordinate system, why do we know that the result will be the same if I multiply the components in the primed system versus in the unprimed system?




                          Given a physical system in which "length" and "angle" are defined, the dot product is invariant under rotations and reflections, i.e. orthonormal transformations. So given two coordinate systems, as long the axes are orthogonal to each other within each coordinate system, and the two coordinate systems have the same origin and the same scale (one unit is the same length, regardless of which direction or coordinate system), dot products will be the same.



                          In that case, the change of basis can be represented with a matrix $U$ such that $(U^*)U=I$ (For real numbers, $U^*$ is just the transpose, so I'll be using that for the rest, since presumably you're asking about vectors over the real numbers). The dot product of two vectors $x$ and $y$ is $x^Ty$. If $x'=UX$ and $y'=Uy$, then the dot product of $x'$ and $y'$ is $x'^Ty'=(Ux)^TUy=x^TU^TUy=x^TIy=x^Ty$






                          share|cite|improve this answer









                          $endgroup$

















                            1












                            $begingroup$


                            How, would I compute their dot product?




                            You pretty much have to convert them to the same basis system. You can multiply them out and get nine different terms, and then find the dot product in terms of the nine dot products of the basis vectors, but the math is pretty much the same as converting to the same coordinate system.




                            In particular, is there a more formal/abstract/generalized definition of the dot product (that would allow me to compute e1→⋅e′1→ without converting the vectors to the same coordinate system)?




                            The value of $vece_1 cdot vece_1'$ is an empirical value. You can't calculate it simply from a definition.




                            Even if I did convert the vectors to the same coordinate system, why do we know that the result will be the same if I multiply the components in the primed system versus in the unprimed system?




                            Given a physical system in which "length" and "angle" are defined, the dot product is invariant under rotations and reflections, i.e. orthonormal transformations. So given two coordinate systems, as long the axes are orthogonal to each other within each coordinate system, and the two coordinate systems have the same origin and the same scale (one unit is the same length, regardless of which direction or coordinate system), dot products will be the same.



                            In that case, the change of basis can be represented with a matrix $U$ such that $(U^*)U=I$ (For real numbers, $U^*$ is just the transpose, so I'll be using that for the rest, since presumably you're asking about vectors over the real numbers). The dot product of two vectors $x$ and $y$ is $x^Ty$. If $x'=UX$ and $y'=Uy$, then the dot product of $x'$ and $y'$ is $x'^Ty'=(Ux)^TUy=x^TU^TUy=x^TIy=x^Ty$






                            share|cite|improve this answer









                            $endgroup$















                              1












                              1








                              1





                              $begingroup$


                              How, would I compute their dot product?




                              You pretty much have to convert them to the same basis system. You can multiply them out and get nine different terms, and then find the dot product in terms of the nine dot products of the basis vectors, but the math is pretty much the same as converting to the same coordinate system.




                              In particular, is there a more formal/abstract/generalized definition of the dot product (that would allow me to compute e1→⋅e′1→ without converting the vectors to the same coordinate system)?




                              The value of $vece_1 cdot vece_1'$ is an empirical value. You can't calculate it simply from a definition.




                              Even if I did convert the vectors to the same coordinate system, why do we know that the result will be the same if I multiply the components in the primed system versus in the unprimed system?




                              Given a physical system in which "length" and "angle" are defined, the dot product is invariant under rotations and reflections, i.e. orthonormal transformations. So given two coordinate systems, as long the axes are orthogonal to each other within each coordinate system, and the two coordinate systems have the same origin and the same scale (one unit is the same length, regardless of which direction or coordinate system), dot products will be the same.



                              In that case, the change of basis can be represented with a matrix $U$ such that $(U^*)U=I$ (For real numbers, $U^*$ is just the transpose, so I'll be using that for the rest, since presumably you're asking about vectors over the real numbers). The dot product of two vectors $x$ and $y$ is $x^Ty$. If $x'=UX$ and $y'=Uy$, then the dot product of $x'$ and $y'$ is $x'^Ty'=(Ux)^TUy=x^TU^TUy=x^TIy=x^Ty$






                              share|cite|improve this answer









                              $endgroup$




                              How, would I compute their dot product?




                              You pretty much have to convert them to the same basis system. You can multiply them out and get nine different terms, and then find the dot product in terms of the nine dot products of the basis vectors, but the math is pretty much the same as converting to the same coordinate system.




                              In particular, is there a more formal/abstract/generalized definition of the dot product (that would allow me to compute e1→⋅e′1→ without converting the vectors to the same coordinate system)?




                              The value of $vece_1 cdot vece_1'$ is an empirical value. You can't calculate it simply from a definition.




                              Even if I did convert the vectors to the same coordinate system, why do we know that the result will be the same if I multiply the components in the primed system versus in the unprimed system?




                              Given a physical system in which "length" and "angle" are defined, the dot product is invariant under rotations and reflections, i.e. orthonormal transformations. So given two coordinate systems, as long the axes are orthogonal to each other within each coordinate system, and the two coordinate systems have the same origin and the same scale (one unit is the same length, regardless of which direction or coordinate system), dot products will be the same.



                              In that case, the change of basis can be represented with a matrix $U$ such that $(U^*)U=I$ (For real numbers, $U^*$ is just the transpose, so I'll be using that for the rest, since presumably you're asking about vectors over the real numbers). The dot product of two vectors $x$ and $y$ is $x^Ty$. If $x'=UX$ and $y'=Uy$, then the dot product of $x'$ and $y'$ is $x'^Ty'=(Ux)^TUy=x^TU^TUy=x^TIy=x^Ty$







                              share|cite|improve this answer












                              share|cite|improve this answer



                              share|cite|improve this answer










                              answered May 13 at 18:09









                              AcccumulationAcccumulation

                              3,459615




                              3,459615





















                                  1












                                  $begingroup$

                                  The formula



                                  $$langle x_1,x_2,x_3rangle cdot langle y_1,y_2,y_3rangle = x_1 y_1 + x_2 y_2 + x_3 y _3$$



                                  is just a start and, as you go further down in physics, will need quite a few generalizations. The assumptions here are that your vectors are (a) real (b) three-dimensional (c) tuples (d) written in a "standard basis". There are points at which either of these are broken: for example, one of the first things you learn in special theory of relativity(*) is how to work with (b') four-dimensional vectors that (d') don't even allow a standard basis at all, so you get a different formula (of which this is a special case). Similarly, in quantum mechanics, depending on the text, you need to grasp in quantum mechanics are (a') complex vector spaces of (b'') infinite-dimensional things that (c') may not be tuples at all (although often can be written so, again allowing a formula of which this is a special case).



                                  You just yourself figured out that (d) will not always be the case, and that's a splendid job on your part.



                                  Before any of those generalizations take place, the assumptions (a - d) are taken for granted. That is, we are working in a basis
                                  $$e_1 equiv langle 1,0,0 rangle \
                                  e_2 equiv langle 0,1,0 rangle \
                                  e_3 equiv langle 0,0,1 rangle$$

                                  and
                                  $$e_1 cdot e_1 = 1, e_1 cdot e_2 = 0, e_1 cdot e_3 = 0 textetc.$$
                                  If a triple of numbers is written it is in this basis. While there are other bases, they just represent concrete triples which you have to multiply by the corresponding coefficients and sum up, effectively transforming to $(e_1, e_2, e_3)$, if you insist on applying the scalar product formula above.



                                  The generalization to taking vectors not as triples of numbers, but as combinations of some abstract $e'_1$, $e'_2$, $e'_3$, then requires specifying what $e'_i cdot e'_j$ is for all $i$, $j$, as other answers have already said in a plenty of ways. If $(e_i)$ and $(e'_i)$ are two different bases, and you know the scalar product in one, the scalar product in the other can be computed from the relations between the basis vectors. And so can a formula for taking scalar products of two vectors, one in each of the two bases.



                                  The basic idea remains, though, and it is a good idea to get oneself familiarized with all the aspects of the above as deeply as possible: to understand the relation between scalar product and norm, orthogonality, expression of geometrical properties and relations (length, angle, distance), etc., before things get too abstract. That's why many texts just hold on to the simplest formula as long as they can.




                                  To actually answer your question: let



                                  $$vecx = x_1 vece_1 + x_2 vece_2 + x_3 vece_3$$
                                  $$vecy = y_1 vece_1' + y_2 vece_2' + y_3 vece_3'$$



                                  such that $(vece_1, vece_2, vece_3)$ is the standard basis. Let further



                                  $$vece_i' = sum_j=1^3 E_i,j vece_j,$$



                                  so using distributivity and linearity it holds that



                                  $$vece_i' cdot vece_k
                                  = left( sum_j=1^3 E_i,j vece_j right) cdot vece_k
                                  = sum_j=1^3 E_i,j left( vece_j cdot vece_k right)
                                  = sum_j=1^3 E_i,j delta_jk (**)
                                  = E_i,k,$$



                                  (also $vece_k cdot vece_i' = E_i,k$), so



                                  $$vecx cdot vecy
                                  = left( sum_i=1^3 x_i vece_i right) cdot left( sum_j=1^3 y_j vece_j' right)
                                  = sum_i=1^3 sum_j=1^3 x_i y_j left( vece_i cdot vece_j' right)
                                  = sum_i=1^3 sum_j=1^3 x_i y_j E_j,i.$$



                                  You can use this formula for taking dot products of two vertices in different bases.
                                  I'm not sure if this counts as not converting to the same basis or not: you will need the conversion matrix $(E_i,j)$ anyway. You won't need to explicitly write $vecy$ in the $(vece_i)$ basis beforehand, though.




                                  (*) Mathematically speaking, special relativity does not use an actual 'scalar product'. But for my example this suffices without further details.



                                  (**) $delta_jk$ is shorthand for "one when $j=k$ and zero otherwise".






                                  share|cite|improve this answer











                                  $endgroup$

















                                    1












                                    $begingroup$

                                    The formula



                                    $$langle x_1,x_2,x_3rangle cdot langle y_1,y_2,y_3rangle = x_1 y_1 + x_2 y_2 + x_3 y _3$$



                                    is just a start and, as you go further down in physics, will need quite a few generalizations. The assumptions here are that your vectors are (a) real (b) three-dimensional (c) tuples (d) written in a "standard basis". There are points at which either of these are broken: for example, one of the first things you learn in special theory of relativity(*) is how to work with (b') four-dimensional vectors that (d') don't even allow a standard basis at all, so you get a different formula (of which this is a special case). Similarly, in quantum mechanics, depending on the text, you need to grasp in quantum mechanics are (a') complex vector spaces of (b'') infinite-dimensional things that (c') may not be tuples at all (although often can be written so, again allowing a formula of which this is a special case).



                                    You just yourself figured out that (d) will not always be the case, and that's a splendid job on your part.



                                    Before any of those generalizations take place, the assumptions (a - d) are taken for granted. That is, we are working in a basis
                                    $$e_1 equiv langle 1,0,0 rangle \
                                    e_2 equiv langle 0,1,0 rangle \
                                    e_3 equiv langle 0,0,1 rangle$$

                                    and
                                    $$e_1 cdot e_1 = 1, e_1 cdot e_2 = 0, e_1 cdot e_3 = 0 textetc.$$
                                    If a triple of numbers is written it is in this basis. While there are other bases, they just represent concrete triples which you have to multiply by the corresponding coefficients and sum up, effectively transforming to $(e_1, e_2, e_3)$, if you insist on applying the scalar product formula above.



                                    The generalization to taking vectors not as triples of numbers, but as combinations of some abstract $e'_1$, $e'_2$, $e'_3$, then requires specifying what $e'_i cdot e'_j$ is for all $i$, $j$, as other answers have already said in a plenty of ways. If $(e_i)$ and $(e'_i)$ are two different bases, and you know the scalar product in one, the scalar product in the other can be computed from the relations between the basis vectors. And so can a formula for taking scalar products of two vectors, one in each of the two bases.



                                    The basic idea remains, though, and it is a good idea to get oneself familiarized with all the aspects of the above as deeply as possible: to understand the relation between scalar product and norm, orthogonality, expression of geometrical properties and relations (length, angle, distance), etc., before things get too abstract. That's why many texts just hold on to the simplest formula as long as they can.




                                    To actually answer your question: let



                                    $$vecx = x_1 vece_1 + x_2 vece_2 + x_3 vece_3$$
                                    $$vecy = y_1 vece_1' + y_2 vece_2' + y_3 vece_3'$$



                                    such that $(vece_1, vece_2, vece_3)$ is the standard basis. Let further



                                    $$vece_i' = sum_j=1^3 E_i,j vece_j,$$



                                    so using distributivity and linearity it holds that



                                    $$vece_i' cdot vece_k
                                    = left( sum_j=1^3 E_i,j vece_j right) cdot vece_k
                                    = sum_j=1^3 E_i,j left( vece_j cdot vece_k right)
                                    = sum_j=1^3 E_i,j delta_jk (**)
                                    = E_i,k,$$



                                    (also $vece_k cdot vece_i' = E_i,k$), so



                                    $$vecx cdot vecy
                                    = left( sum_i=1^3 x_i vece_i right) cdot left( sum_j=1^3 y_j vece_j' right)
                                    = sum_i=1^3 sum_j=1^3 x_i y_j left( vece_i cdot vece_j' right)
                                    = sum_i=1^3 sum_j=1^3 x_i y_j E_j,i.$$



                                    You can use this formula for taking dot products of two vertices in different bases.
                                    I'm not sure if this counts as not converting to the same basis or not: you will need the conversion matrix $(E_i,j)$ anyway. You won't need to explicitly write $vecy$ in the $(vece_i)$ basis beforehand, though.




                                    (*) Mathematically speaking, special relativity does not use an actual 'scalar product'. But for my example this suffices without further details.



                                    (**) $delta_jk$ is shorthand for "one when $j=k$ and zero otherwise".






                                    share|cite|improve this answer











                                    $endgroup$















                                      1












                                      1








                                      1





                                      $begingroup$

                                      The formula



                                      $$langle x_1,x_2,x_3rangle cdot langle y_1,y_2,y_3rangle = x_1 y_1 + x_2 y_2 + x_3 y _3$$



                                      is just a start and, as you go further down in physics, will need quite a few generalizations. The assumptions here are that your vectors are (a) real (b) three-dimensional (c) tuples (d) written in a "standard basis". There are points at which either of these are broken: for example, one of the first things you learn in special theory of relativity(*) is how to work with (b') four-dimensional vectors that (d') don't even allow a standard basis at all, so you get a different formula (of which this is a special case). Similarly, in quantum mechanics, depending on the text, you need to grasp in quantum mechanics are (a') complex vector spaces of (b'') infinite-dimensional things that (c') may not be tuples at all (although often can be written so, again allowing a formula of which this is a special case).



                                      You just yourself figured out that (d) will not always be the case, and that's a splendid job on your part.



                                      Before any of those generalizations take place, the assumptions (a - d) are taken for granted. That is, we are working in a basis
                                      $$e_1 equiv langle 1,0,0 rangle \
                                      e_2 equiv langle 0,1,0 rangle \
                                      e_3 equiv langle 0,0,1 rangle$$

                                      and
                                      $$e_1 cdot e_1 = 1, e_1 cdot e_2 = 0, e_1 cdot e_3 = 0 textetc.$$
                                      If a triple of numbers is written it is in this basis. While there are other bases, they just represent concrete triples which you have to multiply by the corresponding coefficients and sum up, effectively transforming to $(e_1, e_2, e_3)$, if you insist on applying the scalar product formula above.



                                      The generalization to taking vectors not as triples of numbers, but as combinations of some abstract $e'_1$, $e'_2$, $e'_3$, then requires specifying what $e'_i cdot e'_j$ is for all $i$, $j$, as other answers have already said in a plenty of ways. If $(e_i)$ and $(e'_i)$ are two different bases, and you know the scalar product in one, the scalar product in the other can be computed from the relations between the basis vectors. And so can a formula for taking scalar products of two vectors, one in each of the two bases.



                                      The basic idea remains, though, and it is a good idea to get oneself familiarized with all the aspects of the above as deeply as possible: to understand the relation between scalar product and norm, orthogonality, expression of geometrical properties and relations (length, angle, distance), etc., before things get too abstract. That's why many texts just hold on to the simplest formula as long as they can.




                                      To actually answer your question: let



                                      $$vecx = x_1 vece_1 + x_2 vece_2 + x_3 vece_3$$
                                      $$vecy = y_1 vece_1' + y_2 vece_2' + y_3 vece_3'$$



                                      such that $(vece_1, vece_2, vece_3)$ is the standard basis. Let further



                                      $$vece_i' = sum_j=1^3 E_i,j vece_j,$$



                                      so using distributivity and linearity it holds that



                                      $$vece_i' cdot vece_k
                                      = left( sum_j=1^3 E_i,j vece_j right) cdot vece_k
                                      = sum_j=1^3 E_i,j left( vece_j cdot vece_k right)
                                      = sum_j=1^3 E_i,j delta_jk (**)
                                      = E_i,k,$$



                                      (also $vece_k cdot vece_i' = E_i,k$), so



                                      $$vecx cdot vecy
                                      = left( sum_i=1^3 x_i vece_i right) cdot left( sum_j=1^3 y_j vece_j' right)
                                      = sum_i=1^3 sum_j=1^3 x_i y_j left( vece_i cdot vece_j' right)
                                      = sum_i=1^3 sum_j=1^3 x_i y_j E_j,i.$$



                                      You can use this formula for taking dot products of two vertices in different bases.
                                      I'm not sure if this counts as not converting to the same basis or not: you will need the conversion matrix $(E_i,j)$ anyway. You won't need to explicitly write $vecy$ in the $(vece_i)$ basis beforehand, though.




                                      (*) Mathematically speaking, special relativity does not use an actual 'scalar product'. But for my example this suffices without further details.



                                      (**) $delta_jk$ is shorthand for "one when $j=k$ and zero otherwise".






                                      share|cite|improve this answer











                                      $endgroup$



                                      The formula



                                      $$langle x_1,x_2,x_3rangle cdot langle y_1,y_2,y_3rangle = x_1 y_1 + x_2 y_2 + x_3 y _3$$



                                      is just a start and, as you go further down in physics, will need quite a few generalizations. The assumptions here are that your vectors are (a) real (b) three-dimensional (c) tuples (d) written in a "standard basis". There are points at which either of these are broken: for example, one of the first things you learn in special theory of relativity(*) is how to work with (b') four-dimensional vectors that (d') don't even allow a standard basis at all, so you get a different formula (of which this is a special case). Similarly, in quantum mechanics, depending on the text, you need to grasp in quantum mechanics are (a') complex vector spaces of (b'') infinite-dimensional things that (c') may not be tuples at all (although often can be written so, again allowing a formula of which this is a special case).



                                      You just yourself figured out that (d) will not always be the case, and that's a splendid job on your part.



                                      Before any of those generalizations take place, the assumptions (a - d) are taken for granted. That is, we are working in a basis
                                      $$e_1 equiv langle 1,0,0 rangle \
                                      e_2 equiv langle 0,1,0 rangle \
                                      e_3 equiv langle 0,0,1 rangle$$

                                      and
                                      $$e_1 cdot e_1 = 1, e_1 cdot e_2 = 0, e_1 cdot e_3 = 0 textetc.$$
                                      If a triple of numbers is written it is in this basis. While there are other bases, they just represent concrete triples which you have to multiply by the corresponding coefficients and sum up, effectively transforming to $(e_1, e_2, e_3)$, if you insist on applying the scalar product formula above.



                                      The generalization to taking vectors not as triples of numbers, but as combinations of some abstract $e'_1$, $e'_2$, $e'_3$, then requires specifying what $e'_i cdot e'_j$ is for all $i$, $j$, as other answers have already said in a plenty of ways. If $(e_i)$ and $(e'_i)$ are two different bases, and you know the scalar product in one, the scalar product in the other can be computed from the relations between the basis vectors. And so can a formula for taking scalar products of two vectors, one in each of the two bases.



                                      The basic idea remains, though, and it is a good idea to get oneself familiarized with all the aspects of the above as deeply as possible: to understand the relation between scalar product and norm, orthogonality, expression of geometrical properties and relations (length, angle, distance), etc., before things get too abstract. That's why many texts just hold on to the simplest formula as long as they can.




                                      To actually answer your question: let



                                      $$vecx = x_1 vece_1 + x_2 vece_2 + x_3 vece_3$$
                                      $$vecy = y_1 vece_1' + y_2 vece_2' + y_3 vece_3'$$



                                      such that $(vece_1, vece_2, vece_3)$ is the standard basis. Let further



                                      $$vece_i' = sum_j=1^3 E_i,j vece_j,$$



                                      so using distributivity and linearity it holds that



                                      $$vece_i' cdot vece_k
                                      = left( sum_j=1^3 E_i,j vece_j right) cdot vece_k
                                      = sum_j=1^3 E_i,j left( vece_j cdot vece_k right)
                                      = sum_j=1^3 E_i,j delta_jk (**)
                                      = E_i,k,$$



                                      (also $vece_k cdot vece_i' = E_i,k$), so



                                      $$vecx cdot vecy
                                      = left( sum_i=1^3 x_i vece_i right) cdot left( sum_j=1^3 y_j vece_j' right)
                                      = sum_i=1^3 sum_j=1^3 x_i y_j left( vece_i cdot vece_j' right)
                                      = sum_i=1^3 sum_j=1^3 x_i y_j E_j,i.$$



                                      You can use this formula for taking dot products of two vertices in different bases.
                                      I'm not sure if this counts as not converting to the same basis or not: you will need the conversion matrix $(E_i,j)$ anyway. You won't need to explicitly write $vecy$ in the $(vece_i)$ basis beforehand, though.




                                      (*) Mathematically speaking, special relativity does not use an actual 'scalar product'. But for my example this suffices without further details.



                                      (**) $delta_jk$ is shorthand for "one when $j=k$ and zero otherwise".







                                      share|cite|improve this answer














                                      share|cite|improve this answer



                                      share|cite|improve this answer








                                      edited 2 days ago

























                                      answered 2 days ago









                                      The VeeThe Vee

                                      883413




                                      883413



























                                          draft saved

                                          draft discarded
















































                                          Thanks for contributing an answer to Physics Stack Exchange!


                                          • Please be sure to answer the question. Provide details and share your research!

                                          But avoid


                                          • Asking for help, clarification, or responding to other answers.

                                          • Making statements based on opinion; back them up with references or personal experience.

                                          Use MathJax to format equations. MathJax reference.


                                          To learn more, see our tips on writing great answers.




                                          draft saved


                                          draft discarded














                                          StackExchange.ready(
                                          function ()
                                          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fphysics.stackexchange.com%2fquestions%2f479656%2fformal-definition-of-dot-product%23new-answer', 'question_page');

                                          );

                                          Post as a guest















                                          Required, but never shown





















































                                          Required, but never shown














                                          Required, but never shown












                                          Required, but never shown







                                          Required, but never shown

































                                          Required, but never shown














                                          Required, but never shown












                                          Required, but never shown







                                          Required, but never shown







                                          Popular posts from this blog

                                          Category:9 (number) SubcategoriesMedia in category "9 (number)"Navigation menuUpload mediaGND ID: 4485639-8Library of Congress authority ID: sh85091979ReasonatorScholiaStatistics

                                          Circuit construction for execution of conditional statements using least significant bitHow are two different registers being used as “control”?How exactly is the stated composite state of the two registers being produced using the $R_zz$ controlled rotations?Efficiently performing controlled rotations in HHLWould this quantum algorithm implementation work?How to prepare a superposed states of odd integers from $1$ to $sqrtN$?Why is this implementation of the order finding algorithm not working?Circuit construction for Hamiltonian simulationHow can I invert the least significant bit of a certain term of a superposed state?Implementing an oracleImplementing a controlled sum operation

                                          Magento 2 “No Payment Methods” in Admin New OrderHow to integrate Paypal Express Checkout with the Magento APIMagento 1.5 - Sales > Order > edit order and shipping methods disappearAuto Invoice Check/Money Order Payment methodAdd more simple payment methods?Shipping methods not showingWhat should I do to change payment methods if changing the configuration has no effects?1.9 - No Payment Methods showing upMy Payment Methods not Showing for downloadable/virtual product when checkout?Magento2 API to access internal payment methodHow to call an existing payment methods in the registration form?