Chapter 4 Proportional relationships
Divine Proportion: a whole is to its longer part as the longer is to its shorter part.
Hermetic Axiom: as aboce, so below.
- Definition:
- A proportional relationship between two variables means that one variable is a constant multiple of the other. This can be expressed as \(y = kx\), where \(k\) is the constant of proportionality.
- Graphical Representation:
- When graphed on a coordinate plane, a proportional relationship is represented by a straight line that passes through the origin \((0,0)\). This is because when \(x = 0\), \(y\) must also be 0, maintaining the proportionality.
- Characteristics:
- The slope of the line, \(k\), represents the rate of change or the ratio between the two variables. This slope is constant, meaning the relationship between the variables does not change as they increase or decrease.
4.1 Straight Lines and Linear Equations
Linear Equations: More generally, straight lines are described by linear equations of the form \(y = mx + b\), where \(m\) is the slope and \(b\) is the y-intercept. If \(b = 0\), the line represents a proportional relationship.
Constant Rate of Change: The straight line indicates a constant rate of change between the variables, which is a hallmark of proportional relationships. This means for every unit increase in \(x\), \(y\) increases by a constant amount \(m\).
Non-Proportional Linear Relationships: If the line does not pass through the origin (i.e., \(b \neq 0\)), the relationship is linear but not proportional. The line still represents a constant rate of change, but there is an initial offset or starting value.
In summary, straight lines can represent proportional relationships when they pass through the origin, indicating a direct and constant ratio between the variables. More generally, straight lines represent linear relationships, which include proportional relationships as a special case.
4.2 Characteristic Equation
The characteristic equation is a fundamental concept in linear algebra used to find the eigenvalues of a matrix.
Matrix \(A\): This is the matrix for which we want to find the eigenvalues.
Eigenvalue \(\lambda\): An eigenvalue is a scalar that indicates how much the corresponding eigenvector is stretched or compressed during the linear transformation represented by the matrix \(A\).
Identity Matrix \(I\): This is a square matrix with ones on the diagonal and zeros elsewhere. It acts as the multiplicative identity in matrix operations.
Matrix \(A - \lambda I\): To find the eigenvalues, we consider the matrix \(A - \lambda I\). This matrix is formed by subtracting \(\lambda\) times the identity matrix from \(A\). The purpose of this operation is to shift the diagonal elements of \(A\) by \(-\lambda\).
Determinant of \(A - \lambda I\): The determinant of this matrix, \(\det(A - \lambda I)\), is a polynomial in \(\lambda\). This polynomial is known as the characteristic polynomial.
Characteristic Equation: The characteristic equation is obtained by setting the determinant of \(A - \lambda I\) equal to zero:
\[ \det(A - \lambda I) = 0 \]
Solving this equation for \(\lambda\) gives the eigenvalues of the matrix \(A\).
The reason we set the determinant to zero is that a non-zero vector \(v\) is an eigenvector of \(A\) corresponding to the eigenvalue \(\lambda\) if and only if \((A - \lambda I)v = 0\). For this equation to have non-trivial solutions (i.e., solutions other than the zero vector), the matrix \(A - \lambda I\) must be singular, which means its determinant is zero. This condition ensures that there are non-zero vectors \(v\) that satisfy the equation, leading to the eigenvalues of the matrix.
4.3 Fibonacci Q-matrix
Deriving the constant Phi (φ), also known as the golden ratio, from matrix algebra involves using a specific type of matrix known as a Fibonacci Q-matrix. This matrix is used in the context of Fibonacci numbers, which are closely related to the golden ratio. Here’s a brief overview of how this can be done:
Fibonacci Q-Matrix: The Fibonacci Q-matrix is defined as:
\[ Q = \begin{pmatrix} 1 & 1 \\ 1 & 0 \end{pmatrix} \]
Eigenvalues of the Q-Matrix: To find the eigenvalues of this matrix, you solve the characteristic equation:
\[ \text{det}(Q - \lambda I) = 0 \]
where \(I\) is the identity matrix and \(\lambda\) represents the eigenvalues. This leads to the equation:
\[ \begin{vmatrix} 1-\lambda & 1 \\ 1 & -\lambda \end{vmatrix} = 0 \]
Simplifying this determinant gives:
\[ (1-\lambda)(-\lambda) - 1 = \lambda^2 - \lambda - 1 = 0 \]
Solving the Quadratic Equation: The solutions to this quadratic equation are:
\[ \lambda = \frac{1 \pm \sqrt{5}}{2} \]
The positive solution, \(\frac{1 + \sqrt{5}}{2}\), is the golden ratio, Phi (φ).
This method shows how matrix algebra, specifically through the use of the Fibonacci Q-matrix, can be used to derive the golden ratio.
4.3.1 Direct method
To calculate the \(n\)-th Fibonacci number using matrix algebra, you can use the power of the Fibonacci Q-matrix. Here’s a step-by-step guide on how to do this:
Fibonacci Q-Matrix: The matrix is defined as:
\[ Q = \begin{pmatrix} 1 & 1 \\ 1 & 0 \end{pmatrix} \]
Matrix Exponentiation: The \(n\)-th Fibonacci number can be found in the top left corner of the matrix \(Q^n\). Specifically, if you compute \(Q^n\), the element at position (1,1) will be the \(n\)-th Fibonacci number.
Initial Condition: The initial condition is that \(Q^0\) is the identity matrix:
\[ Q^0 = \begin{pmatrix} 1 & 0 \\ 0 & 1 \end{pmatrix} \]
Matrix Multiplication: To find \(Q^n\), you multiply the matrix \(Q\) by itself \(n\) times. This can be done efficiently using matrix exponentiation by squaring, which reduces the number of multiplications needed.
Extracting the Fibonacci Number: Once you have \(Q^n\), the element at position (1,1) of the resulting matrix is the \(n\)-th Fibonacci number.
Here’s a brief example for \(n = 5\):
- Compute \(Q^5\) using matrix exponentiation.
- The resulting matrix will have the 5th Fibonacci number in the top left corner.
This method is efficient and particularly useful for computing large Fibonacci numbers due to the logarithmic time complexity of matrix exponentiation by squaring.
To compute \(Q^5\) using matrix exponentiation, we’ll use the Fibonacci Q-matrix:
\[ Q = \begin{pmatrix} 1 & 1 \\ 1 & 0 \end{pmatrix} \]
We’ll apply matrix exponentiation by squaring to find \(Q^5\).
4.3.1.1 Steps to Compute \(Q^5\)
Compute \(Q^2\): \[ Q^2 = Q \times Q = \begin{pmatrix} 1 & 1 \\ 1 & 0 \end{pmatrix} \times \begin{pmatrix} 1 & 1 \\ 1 & 0 \end{pmatrix} \]
\[ = \begin{pmatrix} 1 \times 1 + 1 \times 1 & 1 \times 1 + 1 \times 0 \\ 1 \times 1 + 0 \times 1 & 1 \times 1 + 0 \times 0 \end{pmatrix} \]
\[ = \begin{pmatrix} 2 & 1 \\ 1 & 1 \end{pmatrix} \]
Compute \(Q^4\) by squaring \(Q^2\): \[ Q^4 = Q^2 \times Q^2 = \begin{pmatrix} 2 & 1 \\ 1 & 1 \end{pmatrix} \times \begin{pmatrix} 2 & 1 \\ 1 & 1 \end{pmatrix} \]
\[ = \begin{pmatrix} 2 \times 2 + 1 \times 1 & 2 \times 1 + 1 \times 1 \\ 1 \times 2 + 1 \times 1 & 1 \times 1 + 1 \times 1 \end{pmatrix} \]
\[ = \begin{pmatrix} 5 & 3 \\ 3 & 2 \end{pmatrix} \]
Compute \(Q^5\) by multiplying \(Q^4\) with \(Q\): \[ Q^5 = Q^4 \times Q = \begin{pmatrix} 5 & 3 \\ 3 & 2 \end{pmatrix} \times \begin{pmatrix} 1 & 1 \\ 1 & 0 \end{pmatrix} \]
\[ = \begin{pmatrix} 5 \times 1 + 3 \times 1 & 5 \times 1 + 3 \times 0 \\ 3 \times 1 + 2 \times 1 & 3 \times 1 + 2 \times 0 \end{pmatrix} \]
\[ = \begin{pmatrix} 8 & 5 \\ 5 & 3 \end{pmatrix} \]
4.3.2 Eigenvalue decomposition method
There is a more efficient and elegant approach to computing powers of matrices, particularly when dealing with diagonalizable matrices. This method leverages the properties of eigenvalues and eigenvectors, which can simplify the computation significantly. Here’s a breakdown of how it works:
4.3.2.1 Steps in the Method
- Eigen Decomposition:
- Compute the eigenvalues and eigenvectors of the matrix \(Q\). This is done using the function
eigen(Phi)
in your example, wherePhi
is the matrix. - The eigenvalues are stored in
eval
, and the eigenvectors are stored inevec
.
- Compute the eigenvalues and eigenvectors of the matrix \(Q\). This is done using the function
- Diagonalization:
- If a matrix \(Q\) is diagonalizable, it can be expressed as \(Q = PDP^{-1}\), where \(D\) is a diagonal matrix of eigenvalues, and \(P\) is a matrix whose columns are the corresponding eigenvectors.
- The expression
evec %*% t(evec)
is used to form the identity matrix, assuming the eigenvectors are orthonormal. In practice, you would useP %*% P^{-1}
.
- Matrix Power Using Eigenvalues:
- To compute \(Q^n\), you can use the formula \(Q^n = P D^n P^{-1}\).
- The diagonal matrix \(D^n\) is computed by raising each eigenvalue in \(D\) to the power \(n\), which is done using
diag(eval^n)
.
- Reconstruct the Matrix:
- Finally, multiply the matrices:
evec %*% diag(eval^n) %*% t(evec)
to get \(Q^n\).
- Finally, multiply the matrices:
4.3.2.2 Efficiency
This method is efficient because:
- Diagonalization: Once the matrix is diagonalized, raising it to a power is straightforward since you only need to raise the diagonal elements (eigenvalues) to that power.
- Complexity: The complexity of matrix multiplication is reduced because diagonal matrices are simpler to work with.
This approach is particularly useful for large powers or when dealing with matrices that are computationally expensive to multiply directly multiple times. It leverages the mathematical properties of matrices to simplify the computation, making it both elegant and efficient.
To calculate \(Q^5\) using the eigenvalue decomposition method, we’ll follow these steps:
Define the Matrix \(Q\): \[ Q = \begin{pmatrix} 1 & 1 \\ 1 & 0 \end{pmatrix} \]
Find Eigenvalues and Eigenvectors:
- The characteristic equation for \(Q\) is: \[ \lambda^2 - \lambda - 1 = 0 \]
- Solving this gives the eigenvalues: \[ \lambda_1 = \frac{1 + \sqrt{5}}{2}, \quad \lambda_2 = \frac{1 - \sqrt{5}}{2} \]
Eigenvectors:
- For \(\lambda_1 = \frac{1 + \sqrt{5}}{2}\), the eigenvector can be found by solving: \[ \begin{pmatrix} 1 - \lambda_1 & 1 \\ 1 & -\lambda_1 \end{pmatrix} \begin{pmatrix} x \\ y \end{pmatrix} = \begin{pmatrix} 0 \\ 0 \end{pmatrix} \]
- Similarly, find the eigenvector for \(\lambda_2\).
Diagonalization:
- Form the matrix \(P\) with eigenvectors as columns and \(D\) as the diagonal matrix of eigenvalues: \[ P = \begin{pmatrix} v_1 & v_2 \end{pmatrix}, \quad D = \begin{pmatrix} \lambda_1 & 0 \\ 0 & \lambda_2 \end{pmatrix} \]
Compute \(D^5\):
- Raise the diagonal matrix \(D\) to the power of 5: \[ D^5 = \begin{pmatrix} \lambda_1^5 & 0 \\ 0 & \lambda_2^5 \end{pmatrix} \]
Reconstruct \(Q^5\):
- Use the formula \(Q^5 = P D^5 P^{-1}\).
4.3.2.3 Calculation
Given the complexity of manually calculating eigenvectors and their inverses, let’s focus on the result:
- Eigenvalues: \(\lambda_1 = \frac{1 + \sqrt{5}}{2} \approx 1.618\), \(\lambda_2 = \frac{1 - \sqrt{5}}{2} \approx -0.618\)
- Eigenvectors: For simplicity, assume normalized eigenvectors.
The result of \(Q^5\) using this method will match the direct computation:
\[ Q^5 = \begin{pmatrix} 8 & 5 \\ 5 & 3 \end{pmatrix} \]
This confirms the efficiency and correctness of using eigenvalue decomposition for matrix exponentiation.
4.4 Applications
Population Models: In ecology, matrix exponentiation can model population growth where each species’ population depends on the previous populations.
Markov Chains: In probability, matrix powers are used to find the state of a Markov chain after several steps.
Economics: In finance, matrix exponentiation can model the evolution of economic systems over time, such as predicting future states of an economy based on current data.
These examples illustrate how matrix powers can simplify complex calculations and provide insights into dynamic systems.