- #1
stunner5000pt
- 1,461
- 2
Show that [tex] \partial'_{\alpha} A'^\alpha (x') = \partial _{mu} A^{\mu}(x') [/tex]
lets focus on the partial operator for now
[tex] \partial'_{\alpha} = \frac{\partial}{\partial x'^{\alpha}} = \frac{\partial}{L_{\nu}^{\alpha} \partial x^{\nu}} [/tex]
Now A represents the Scalar and vector fields of an EM field.
[tex] A'^{\alpha}(x') = L_{\sigma}^{\alpha} A^{\sigma}(x') [/tex]
is that fine?
when i put them together
[tex] \partial'_{\alpha} A'^\alpha (x') = \frac{\partial}{L_{\nu}^{\alpha} \partial x^{\nu}} L_{\sigma}^{\alpha} A^{\sigma}(x') [/tex]
the argumetn is that both the L s represent the same dimensions thus the they are the same thing?
But Since L is a matrix... i can't be int eh denominator... can it? Would it simply be represented as an inverse? The two Ls still turn into idnetity matrix which is simply 1.
your helpsi greatly appreciated!
lets focus on the partial operator for now
[tex] \partial'_{\alpha} = \frac{\partial}{\partial x'^{\alpha}} = \frac{\partial}{L_{\nu}^{\alpha} \partial x^{\nu}} [/tex]
Now A represents the Scalar and vector fields of an EM field.
[tex] A'^{\alpha}(x') = L_{\sigma}^{\alpha} A^{\sigma}(x') [/tex]
is that fine?
when i put them together
[tex] \partial'_{\alpha} A'^\alpha (x') = \frac{\partial}{L_{\nu}^{\alpha} \partial x^{\nu}} L_{\sigma}^{\alpha} A^{\sigma}(x') [/tex]
the argumetn is that both the L s represent the same dimensions thus the they are the same thing?
But Since L is a matrix... i can't be int eh denominator... can it? Would it simply be represented as an inverse? The two Ls still turn into idnetity matrix which is simply 1.
your helpsi greatly appreciated!