QuasiEigen Systems 

In a previous note we discussed the equivalence between eigenvalues and eigenvectors, and briefly considered the mapping between the normalized components of an eigenvector and the ratios of the components (i.e., eigenvalues) of that same vector. In either of these interpretations, only the ratios of the components are of significance, not their individual absolute values. However, since observations are not arbitrarily rescaleable – at least not on the smallest scales – it’s interesting to consider the possibility that the individual components have absolute significance. There is a definite absolute scale associated with the existence of a minimal action greater than zero, i.e., a finite quantum of action, characterized by Planck’s constant h. To accurately represent physical phenomena, perhaps we ought to replace the 0 in equation (3) of the previous note with a vector of magnitude h, which we will denote as h. Having done this, the equation takes the inhomogeneous form 

_{} 

We call this a pseudoeigen system. Now, assuming the matrix inside the parentheses is not singular, we could simply multiply both sides by the inverse of that matrix to give the unique solution for x, but this corresponds to the trivial x = 0 solution in the traditional formulation, except this solution will be on the order of h rather than 0. Hence this still represents a kind of trivial null solution, whereas (just as in the traditional case) the equation also possesses nontrivial solutions, provided the combined matrix on the left side satisfies a certain condition. Specifically, there will be nontrivial solutions x provided the determinant of the combined matrix is on the order of h. Thus we require 

_{} 

To give a simple illustration, consider the two matrices 

_{} 

The condition for nontrivial solution is 

_{} 

If the right hand side were zero, we could divide through by b^{2}, and this would just be the characteristic equation with two roots, representing the two eigenvalues of the system. But since the right hand side is not zero, this equation gives a surface, which we will refer to as an eigensurface. In this simple example the surface is a conic locus as illustrated in the figure below. 


The solution x is of the form 

_{} 

For any point of the (a,b) plane (other than the origin) we have a solution, but since h is extremely small, the solution components will be correspondingly small unless the determinant is on the same order of magnitude as h. The eigensurface represents the locus of points on the (a,b) plane where the determinant equals h, so the leading factor in the expression for x is unity. However, the components of the solution must still be extremely small because a and b are extremely small on the eigensurface, except where the eigensurface approaches the original eigenvalues. On those asymptotes (and only there) we can make a and b arbitrarily large, but of course they must maintain essentially a constant ratio, which approaches the traditional eigenvalue the further we are from the origin. In that case the components of x approach a constant ratio to each other, i.e., they approach the traditional eigenvector. 

The eigensurface has several interesting features. It provides a continuous path from one eigenvalue (asymptote) to the other, but it also consists of two disjoint branches. Which pairs of asymptotes are connected depends on the particular conic that describes the locus. The hyperbolic nature of the locus is a consequence (in ordinary quantum mechanics) of the stipulation that the eigenvalues of observable operators are purely real. If h is zero (as it is traditionally taken to be), the discriminant of the characteristic equation must be positive, and hence the conic when h is nonzero is a hyperbola. However, since h is nonzero, it is no longer necessary to require a positive discriminant (so observable operators need not be Hermitian), since we can get real eigenvalues even for characteristic equations with negative discriminants. Thus we can allow for elliptical eigensurfaces as depicted below. 


Of course, the solution vectors x for such systems are necessarily extremely close to null, since there are no asymptotes extending far from the origin. If h is sufficiently small, we might never notice the existence of such systems. The potential for parabolic systems is also intriguing, as they would be singlevalued but allowing for solutions vectors of arbitrary size, and those vectors would not have components in constant proportion, since the parabola is not asymptotic to any fixed rays from the origin. 

In general the degree of the eigensurface equals the order of the system. For two square matrices A and B of order n, we can express the determinant aA+bB in terms of the determinants a_{j}A+b_{j}B with j = 1, 2, …, n+1, where the (a_{j},b_{j}) are n+1 independent “basis points” of the a,b plane. The relation can be written in the form 

_{} 

Equation (3) was written using the basis points (1,0), (0,1), and (1,1), but if we happen to know the n ordinary eigenvalues (m_{j},n_{j}) such that m_{j}A+n_{j}B = 0 we could use them as n of the required n+1 basis points. (Of course this can be done only if there are n distinct eigenvalues.) We can then freely select just one more basis point, in terms of which to express the eigensurface equation. In doing so we must take care to preserve the scale factor, since ultimately we are setting a determinant not to 0 but to h. The characteristic equation giving the ordinary eigenvalues doesn’t contain the scale information, since only the ratios are important, so we can’t simply identify the characteristic polynomial with the determinant. Writing the ordinary characteristic equation in the symmetrical form 

_{} 

we have 
_{} 

for arbitrary constants k_{1} and k_{2} (not both equal to zero), and hence we can write the equation of the eigensurface as 

_{} 

Naturally we can combine the “quantized eigen problem” with the generalization to multiple partitions discussed previously to represent interactions involving more than just two systems. It can certainly be argued that the conventional focus on purely binary interactions is not theoretically justified. If indeed there are “measurement” interactions involving (say) three systems, the relevant equation might be expressible in the form 

_{} 

for observables A, B, and C. If the matrices are of order two, the eigensurface would then be given by 

_{} 

This is the equation of a quadric surface in the threedimensional space of a,b,g. 

The key distinctions between systems described by an equation such as (4) and those described by an equation such as (5) in the previous note is that the former involves an absolute scale, so it’s important to correctly define the absolute magnitudes (not just the ratios) of the coefficients of the polynomials defining the characteristic solutions. Thus its useful to understand the form of the expressions for those coefficients. Consider the threepartition system of order n, for which we have the determinant 

_{} 

where m_{i,j} are the components of M, and s represents a permutation [s_{1},s_{2},…,s_{n}] of the indices 1 to n, and S is the set of all such permutations, and s(s) is either +1 or 1 accordingly as the permutation s is even or odd. In this example we make the substitutions 

_{} 

where a_{i,j}, b_{i,j}, and c_{i,j} are the components of A, B, C respectively, and we get a homogeneous polynomial of degree n in the parameters a,b,g. Thus for some column vector R we have 

_{} 

where 
_{} 

The components of R can be found by choosing 6 independent basis points (a_{j},b_{j},g_{j}) with j = 1, 2, …, 6, and solving the system of equations 

_{} 

where 

_{} 

Thus letting H_{(i=J)} denote the matrix H with the ith column replaced with J, we can write the components of R explicitly as 

_{} 

We also have R = H^{1}J, so we can express the desired determinant as a linear combination of the “basis determinants” (i.e., the components of J) by substituting into (5) to give 

_{} 

where K = UH^{1}. Hence we have KH = U, and letting H^{(i=U)} denote the matrix H with the ith row replaced with U, the components of K can be written explicitly as 

_{} 

The concept of quasieigen problems, according to which the determinant of the system is set not to 0 but to the (small) finite value h, might provide an alternative representation of quantum mechanics. It would be interesting to determine the correspondence between these quasieigen systems and classical Poisson brackets as well as their analogs in quantum mechanics. 
