The Laplace Equation and Harmonic Functions 

Given a scalar field j, the Laplace equation in Cartesian coordinates is 

_{} 

This equation arises in many important physical applications, such as potential fields in gravitation and electrostatics, velocity potential fields in fluid dynamics, etc. Typically we are given a set of boundary conditions and we need to solve for the (unique) scalar field j that is a solution of the Laplace equation and that satisfies those boundary conditions. A solution of Laplace's equation is called a "harmonic function" (for reasons explained below). Since the Laplace equation is linear, the sum of two or more individual solutions is also a solution. 

The usual approach to solving the Laplace equation is to seek a "separable" solution given by the product of independent function of x, y, and z, as follows 

_{} 

Every solution can be expressed as a sum of (possibly infinitely many) solutions of this form. Inserting this into the Laplace equation and evaluating the derivatives gives 

_{} 

Dividing through by the product A(x)B(y)C(z), this can be written in the form 

_{} 

Since x, y, and z can be varied independently, this equation can be identically satisfied only if each of the three terms is a constant, and these three constants sum to zero. It will be convenient to denote these constants by squared quantities, a^{2}, b^{2}, g^{2} respectively. Thus we have three separate ordinary differential equations 

_{} 

with the condition a^{2} + b^{2} + g^{2} = 0. Taking just the first of these, we can rewrite it in the form 

_{} 

which the familiar equation of harmonic motion. This is the reason solutions of Laplace's equation are called "harmonic functions". The characteristic polynomial is 

s^{2}  a^{2} = 0 

whose roots are ±a, so the general solution of the differential equation is 

_{} 

for constants a_{1} and a_{2}. We arrive at similar solutions for B(y) and C(z), so the general solution of the Laplace equation of the form A(x)B(y)C(z) is 

_{} 

where the a_{i} and b_{i} are arbitrary constants, and a, b, g are any complex constants such that a^{2} + b^{2} + g^{2} = 0. By summing terms of this general form using the techniques of Fourier analysis we can determine the values of these constants necessary to match any given boundary conditions. 

The above analysis is useful when the boundary conditions are easily expressible in terms of Cartesian coordinates, but we often encounter problems that are easier to express in terms of other, possibly curvilinear, coordinate systems. The general form of the Laplacian operator for an arbitrary (differentiable) coordinate system x^{k}, k = 1,2,3, with the metric tensor g_{ij} is

_{} 

where g is the determinant of the covariant metric tensor. If the basis vectors of the coordinate system are mutually orthogonal at each point, the fundamental line element has the diagonal form 

_{} 

where _{}. In this case we have _{} and g_{ij} = 0 for all i ¹ j. Also, the nonzero components of the contravariant metric are g^{ii} = 1/g_{ii} , so the Laplacian expressed in terms of orthogonal coordinates has the form 

_{} 

For example, ordinary spherical coordinates r, q, f where 

_{} 

are orthogonal, i.e., the basis vectors in the r, q and f directions are mutually orthogonal at each point, so we can express the Laplacian in terms of spherical coordinates by means of the preceding expression. The line element for spherical coordinates is found by simply evaluating the derivatives ds/dr, etc., along the coordinate axes. This gives 

_{} 

and we have h_{r} = 1, h_{q} = r, h_{f} = rsin(q). Inserting these into the general expression for the Laplacian in orthogonal coordinates gives 

_{} 

We are often interested in situations with axial symmetry, such as when determining the potential flow around a sphere. In such cases we can take the axis of symmetry over the angle f, so we have ¶j/¶f = 0, and Laplace's equation reduces to 

_{} 

As in the case of Cartesian rectangular coordinates, we can focus on separable solutions, i.e., solutions of the form j(r,q) = A(r)B(q). Evaluating the Laplacian of this, equating this to zero, and dividing through by A(r)B(q) gives 

_{} 

where dots signify derivatives of the function with respect to the argument. Since r and q are independently variable, this equation can be identically satisfied only if the sum of the first two terms is constant, and the sum of the last two terms is constant. Denoting these constant by m,n, we have the two separate ordinary linear differential equations 

_{} 

with the condition m + n = 0. We could solve these explicitly, but notice that the first is satisfied by setting A(r) = r^{n}, provided we set m = n(n+1). Also, if we set B(q) = cos(q) the second equation is satisfied with n = 0. Hence the function j(r,q) = r^{n} cos(q) is a solution of Laplace's equation provided n(n+1) = 0, which implies either n = 0 or n = 1. 

More generally, we can consider solutions that are linear combinations of functions of the form

_{} 

where P_{n}(x) is a polynomial. Inserting this into the axially symmetric form of Laplace's equation gives 

_{} 

where dotted functions signify derivatives of P_{n}(x) with respect to x. Dividing through by r^{n}, replacing cos(q) with "x", and noting that sin(q)^{2} = 1cos(q)^{2}, this becomes 

_{} 

There is a unique (up to a constant factor) polynomial P_{n} for each value of n(n+1) = k, which implies that each such polynomial is compatible with two distinct values of n, i.e., the two roots n and n' of n^{2} + n = k. Of course, n need not be an integer, but if it is, we can easily determine the polynomial P_{n}(x) that satisfies the above equation, simply by inserting a polynomial with undetermined coefficients into the equation and solving for the coefficients. The results for the first few values of n are as shown below: 



Up to a scale factor, these are the Legendre polynomials (discussed further in the note on Inverse Square Forces and Orthogonal Polynomials). We can express an axially symmetrical solution of the Laplace equation as a linear combination of these individual solutions as follows 

_{} 

where the a_{i} and b_{i} are arbitrary constants. The factors of the form P_{n}(cos(q)) in this expansion are called zonal harmonics. As an example, the velocity potential of the flow of an ideal, irrotational, frictionless fluid past a stationary sphere of radius R is given by the second zonal harmonic with a_{1} = U and b_{1} = UR^{3}/2, where U is the freestream velocity of the fluid moving in the positive x direction (assuming the x axis is our axis of symmetry). Thus we have 

_{} 

The first term is obviously just Ux in Cartesian coordinates, which represents a uniform flow field of constant velocity U in the positive x direction. The second term is a dipole potential, proportional to x/r^{3}. This is the form of the "far field" potential associated with two nearby oppositely charged particles in electrostatics (aligned along the axis of symmetry). It's interesting that the superposition of these two simple fields gives the potential flow field around a sphere, as shown by the fact that the radial velocity vanishes at r = R. (See the note on Potential Flow and d'Alembert's Paradox for more.) 

One of the interesting properties of harmonic functions, i.e., continuous functions j(x,y,z) that satisfy Laplace's equation, is that the mean value of j on any sphere centered on the point p is equal to the value of j at the point p. This is called the mean value theorem, and a formal proof of it is presented in Differential Operators and the Divergence Theorem. This theorem is often invoked to prove that a continuous differentiable harmonic function cannot contain a local maximum or minimum, because by definition such points have values of j properly greater then (resp. less than) the values of j for all the surrounding points, which would make it impossible for the average on the surface of a sphere surrounding the point to equal the value of j at that point. In fact, by the same reasoning, if the value of a harmonic function at any given point is increasing in one direction then it must be decreasing in some other direction. (Of course, we don’t actually need the mean value theorem to prove these facts; we can see at a glance that the Laplace equation automatically precludes local maxima and minima, because at any such point all three of the second partial derivatives with respect to x, y, and z would have the same sign, making it impossible for their sum to be zero.) 

From this it follows that if the value of a harmonic function j has the constant value c over an entire surface that completely encloses a region of space, then the value of j must be c throughout that region. This is because if the value was anything other than c at any point in the interior of the region, then the region must contain a properly highest (or lowest) value, so there must be a point from which the function is decreasing (or increasing) in some directions but not increasing (or decreasing) in any directions, which is impossible. 

Having established this, we can also see that the solution of Laplace's equation satisfying a complete set of boundary conditions is unique. To prove this, suppose two distinct harmonic functions j_{1} and j_{2} have the same values on a closed surface, but have different values in the interior of the enclosed region. We know that harmonic functions are additive, so j_{1}  j_{2} is also a harmonic function, and it's value is zero over the closed surface. But this implies the value of this difference is zero throughout the interior region enclosed by the surface (by the preceding proposition), so j_{1} and j_{2} are identical throughout the region. 

It’s also interesting to consider the mean value theorem in the context of the general power series of a harmonic function. If we expand j(x,y,z) into a power series about the point p (after shifting the coordinates to place p at the origin), we get an expression of the form 

_{} 

Recall that if q denotes the “latitude” (zero at the equator) and f the “longitude”, then the mean value of a function f(q,f) over the surface of a sphere of radius R is given by the double integral

_{} 

Making the substitutions x = Rcos(f)cos(q), y = Rsin(f)cos(q), z = Rsin(q) into the expression for j and evaluating the mean value over this sphere gives 

_{} 

In general we find that only even indices appear, and the mean value can be written as 

_{} 

The constant coefficient in this expression is c_{000}, which equals the value of j at the central point p, corresponding to R = 0. For all other values of R, the mean value theorem asserts that the mean value remains constant, so the coefficients of all nonzero powers of R must vanish. To show that they do, we can substitute the original power series for j into the Laplace equation and collect terms by powers of x,y,z. We find the coefficient of x^{q}y^{r}z^{s} is 

_{} 

Each of these coefficients must vanish. With q = r = s = 0 this gives 

_{} 

which proves that the coefficient of R^{2} in the series for j_{mean} is zero. Likewise for the (q,r,s) combinations (2,0,0), (0,2,0), and (0,0,2) we have the three conditions 

_{} 

Summing these three conditions and dividing by 12 gives 

_{} 

which proves that the coefficient of R^{4} in the double integral for j_{mean} is zero. Proceeding on to the next term, with the (q,r,s) combinations (4,0,0), (0,4,0), (0,0,4), (2,2,0), (2,0,2), and (0,2,2) we get the six conditions 

_{} 

Multiplying the first three by 3 and then adding all these expressions together and dividing by 90 gives 

_{} 

which proves that the coefficient of R^{6} in the double integral for j_{mean} is zero. This process can be continued indefinitely, corresponding to the fact that the coefficient of each positive power of R is zero, and hence j_{mean} = c_{000} for all R. 
