@ozair
1b500718-fb9a-40e5-a9f8-8cc3e51fe41e-image.png
29811036-84e6-4aa3-bf59-06ec0fbca461-image.png
@zaasmi said in MTH603 Grand Quiz Solution and Discussion:
For the system of equations; x =2, y=3. The inverse of the matrix associated with its coefficients is-----------.
⌈ 3 -2⌉ ⌈ x ⌉ = ⌈ 5 ⌉
⌊ 4 3⌋ ⌊ y ⌋ ⌊ -2 ⌋
⌈ x ⌉ = (1/17) ⌈ 3 2 ⌉ ⌈ 5 ⌉
⌊ y ⌋ ⌊ -4 3 ⌋ ⌊-2⌋⌋
= (1/17)⌈11⌉
⌊-26⌋
= ⌈11/17⌉
⌊-26/17⌋
x = 11/17 and y = -26/17
For the system of equations; x =2, y=3. The inverse of the matrix associated with its coefficients is-----------.
@zaasmi said in MTH603 Grand Quiz Solution and Discussion:
The linear equation: x+y=1 has --------- solution/solutions.
This is an stand alone linear equation in two variables. It would have infinite number or solutions. Assign any real value to x, there would be a corresponding real value of y. All these pairs of values are solution of the equation.
The linear equation: x+y=1 has --------- solution/solutions.
@zaasmi said in MTH603 Grand Quiz Solution and Discussion:
Gauss–Seidel method is also known as method of …………….
In numerical linear algebra, the Gauss–Seidel method, also known as the Liebmann method or the method of successive displacement, is an iterative method used to solve a system of linear equations.
Gauss–Seidel method is also known as method of …………….
@zaasmi said in MTH603 Grand Quiz Solution and Discussion:
Differences methods are iterative methods. yes or no
Differences methods are iterative methods. yes or no
@zaasmi said in MTH603 Grand Quiz Solution and Discussion:
Power method is applicable if the eigen vectors corresponding to eigen values are linearly _______.
The Jacobi’s method is a method of solving a matrix equation on a matrix that has no zeros along its main diagonal. Power method is applicable if the eigen vectors corresponding to eigen values are linearly independent.
Power method is applicable if the eigen vectors corresponding to eigen values are linearly _______.
@zaasmi said in MTH603 Grand Quiz Solution and Discussion:
Two matrices with the same characteristic polynomial need not be similar.
Two similar matrices have the same characteristic polynomial. The converse however is not true in general: two matrices with the same characteristic polynomial need not be similar. The matrix A and its transpose have the same characteristic polynomial. … In this case A is similar to a matrix in Jordan normal form.
Two matrices with the same characteristic polynomial need not be similar.
@zaasmi said in MTH603 Grand Quiz Solution and Discussion:
If A is a nxn triangular matrix (upper triangular, lower triangular) or diagonal matrix , the eigenvalues of A are the diagonal entries of A.
The eigenvalues of B are 1,4,6 since B is an upper triangular matrix and eigenvalues of an upper triangular matrix are diagonal entries. We claim that the eigenvalues of A and B are the same.
If A is a nxn triangular matrix (upper triangular, lower triangular) or diagonal matrix , the eigenvalues of A are the diagonal entries of A.
@zaasmi said in MTH603 Grand Quiz Solution and Discussion:
@zaasmi said in MTH603 Grand Quiz Solution and Discussion:
If n x n matrices A and B are similar, then they have the same eigenvalues (with the same multiplicities).
Since similar matrices A and B have the same characteristic polynomial, they also have the same eigenvalues. If B = PAP−1 and v = 0 is an eigenvector of A (say Av = λv) then B(Pv) = PAP−1(Pv) = PA(P−1P)v = PAv = λPv. Thus Pv (which is non-zero since P is invertible) is an eigenvector for B with eigenvalue λ.
The matrix B has the same A as an eigenvalue. M−1x is the eigenvector. If two matrices are similar, they have the same eigenvalues and the same number of independent eigenvectors (but probably not the same eigenvectors).
@zaasmi said in MTH603 Grand Quiz Solution and Discussion:
If n x n matrices A and B are similar, then they have the same eigenvalues (with the same multiplicities).
Since similar matrices A and B have the same characteristic polynomial, they also have the same eigenvalues. If B = PAP−1 and v = 0 is an eigenvector of A (say Av = λv) then B(Pv) = PAP−1(Pv) = PA(P−1P)v = PAv = λPv. Thus Pv (which is non-zero since P is invertible) is an eigenvector for B with eigenvalue λ.