Basics
Describing and manipulating vectors using bra-ket notation.
Last updated
Describing and manipulating vectors using bra-ket notation.
Last updated
A matrix is a way of organizing information. Just like a system of equations, we can use the information in a matrix to denote information about the system overall. The matrix will have dimensions m and n, where m is the number of rows and n is the number of columns.
The individual values in the matrix are referred to as elements or components.
You may be used to thinking of vectors as having a magnitude and a direction. This is a classical way of doing things - usually we're talking about vectors in a physical context, and therefore we're usually using them to describe where something is going (Wikipedia - 2020).
Here we want to think of vectors in a more abstract way. Our vectors here are not really going anywhere in particular. In the same way that we use matrices to hold some kind of information, we do the same with vectors.
A vector contains multiple pieces of information, and a number holds exactly one. We call a plain old number a scalar, because it can be used to scale some other information (like multiply it by 2 for example - changing the scale from 1 to 2).
The vectors we're describing are the sort of column vectors you're likely already very familiar with.
This kind of vector is just a one column matrix, and as such we can preform all of the matrix operations on it.
A column vector has as many elements as it does dimensions - so for example a vector describing something in the real world would have 3-dimensions and therefore 3 elements.
For quantum systems, we'll describe vectors using Bra-Ket notation. Once you get the hang of it, it's simple enough, but for the uninitiated it might be a bit weird.
The basic idea is that vectors can be labeled with what looks like a line ( | ) on one side and an angle bracket ( < , > ) on the other.
So, for the vector x, the label would look like this: (Griffiths, D. - p.119) .
Actually, this is a "ket" vector. A "bra" vector looks like this: . You'll see later that we use this to our advantage when describing the inner product of two vectors: , but that's another tutorial.
You can describe different things about a vector this way, too. So, for the vector x over time t we might write:
For our purpose the only important kinds of matrix operations are addition and multiplication.
Adding two matrices together requires that they both have the same number of rows and columns. As you might expect, you just add the respective entries together:
Simple enough. You should note that subtraction is just a sort of special case of addition, where you started by multiplying the second matrix by -1.
The dot product of two vectors gives us a number representing the combination of the two. We denote this with the two vectors placed "back to back" - . For this operation, we multiply the corresponding entries and add all the results together, in order.
So, for example:
Note that this only works with single column matrices.
The dot product of two vectors that are opposite (perpendicular) of each other is always 0.
The inner product is the more general form of the dot product. To be very broad, you can take the inner product of two vectors and end up with a complex number instead of a real number. We'll talk a bit more about complex numbers later on.
Some interesting properties of the inner product include the following (assume that are scalars and are vectors.
It's linear:
Where ,
The inner product of a vector with itself:
Unless the vectors is 0 (where all of the elements in the vector are 0) -
There are two relevant types of multiplication:
Where we multiply each entry in the matrix by a number:
As you would expect, you could also multiply the matrix by a variable.
Where we multiply two matrices together.
For this operation, we take the dot product of the columns in the first matrix with the rows of the second matrix. This might be a difficult definition to understand, so let's lay out the rules one by one:
Multiplication goes row by column
Multiply the first entry of the first column of the first matrix by the first entry in the first row of the second matrix
Add this to the result of the next entries, respectively
Continue until there are no more entries
As you can see, this means that the number of entries in the rows of the first matrix have to be equal to the number of entries in the columns of the second matrix.
Here I'll include a lovely graphic that should clear things up significantly (Jean, H. - 2018):
There, much better. You should be able to see that you'll end up with a single column matrix at the end.