After facing much difficulty in the derivations from this post, and doing some cursory digging into the "advanced" techniques for solving the troublesome recurrence relations enclosed within, I uncovered some information about the Umbral Calculus, developed by one Rev. John Blissard1 in an 1860s series of articles on symbolic different "Generic Equations,"2,3 which is all together more esoteric and fascinating.
The idea is symbolic in nature, and aims to (amongst other things) find a relationship between discrete and continuous calculus. What it lacks in rigor, it more than makes up for with clever and seemingly magical symbolic derivations which may have foreshadowed Heiner's unified Algebra.4
This post is rather unstructured, and more or less serves as a meandering random walk through some applications of the Umbral Calculus which I found interesting and/or useful.
Operators
We'll begin by introducing some notation which simplifies some of the algebra to come. If we consider the fundamental theorem of calculus:
dxd∫f(x)dx=f(x)
we'll rewrite the derivative and integral operators as D and I:
DIf(x)=f(x)
Theory of Discrete Calculus
Similarly, we can rewrite the discrete analogs of derivations and integrations (differences and summations):
ΔΣf(x)=f(x)
Our goal is to find a relationship to bridge the gap between discrete and classical calculus via some additional operator we'll denote ϕ such that:
ϕDϕI=Δϕ=Σϕ
From pre-calc, we know that the derivative is a continuous rate of change of the line tangent to each point on a curve. For some function such as f(x)=x2, the derivative with respect to x is 2x:
In Discrete Calculus, we're concerned with the discrete rate of change which we measure with Δ:
x
f(x)
Δf(x)
...
...
...
-1
1
-1
0
0
+1
1
1
+3
2
4
+5
3
9
+7
...
...
...
If we know the value of Δf(x) for a specific value of x, we can also compute the value of f(x+1):
so we arrive at a value of 2x+1 for the discrete rate of change, where Δ is known at the forward difference operator, the general form of which is:
Δf(x)=f(x+1)−f(x)
Linear Operators
An important property of Δ is its linearity – or that it preserves addition and scalar multiplication such that:
Δf(x)+Δg(x)=f(x+1)+g(x+1)−f(x)−g(x)
Δaf(x)=af(x+1)−af(x)=a[f(x+1)−f(x)]=aΔf(x)
And, just like in classical, continuous calculus where differentiation is inversely related to integration, forward differencing is the inverse of summation:
DIf(x)=f(x)ΔΣf(x)=f(x)
This should be evident from our plot of f(x)=x2 above where:
9=32=1+3+5=n=0∑2Δn2=n=0∑x−1Δn2=n=0∑x−1(2n+1)
We can prove that this generalization holds by expanding successive terms of a summation involving Δ:
which looks suspiciously similar to the Fundamental Theorem of Calculus... This is useful because it means that finding a summation formula for e.g. n2 is equivalent to finding a function who's Δ is x2 like so:
an equation for x2 in terms of Δ. By linearity, we can factor out the Δ:
x2=Δx3(31x3−21x2+61x)
which is (drum roll) a function who's delta is x2.
If Frank Lloyd Wright tried his hand at exponentiation
As alluded to before, the inverse of the forward difference is given by indefinite summation:
ΔΣf(x)=f(x)
which is unique up to a constant term, like the indefinite integral. Again, recall that our overarching goal is to devise some method of ""interpolating"" between discrete and continuous calculus. The crux of this translation ϕ will be falling powers which is like a parameterized factorial function:
which is shaped the same as the power rule for differentiation, just with the exponent and subscript switched...
DxnΔxn◯=nxn−1=nxn-1◯
And therein lies the behavior of the umbral operator:
ϕ:Rxn→Rϕxn◯
We can observe how ϕ composes with our difference/differentiation operators by tracing the diagram. E.g. from top left to bottom right, first taking the derivative of some number then applying ϕ to it, we get:
ϕDxn=Δϕxn◯
And, by linearity, this holds for all polynomials, as well as a bunch of other transcendental functions that can be expressed by series expansion by polynomials, so for concision we might just write:
ϕD=Δϕ
which kind of begs the question, can't we just divide through by ϕ? No! similar to matrix multiplication it's not multiplicatively commutative. We can, however, multiply through by (apply, composably) its inverse to yield expressions like the following:
Δ=ϕDϕ−1Σ=ϕIϕ−1
(and of course, this also applies to summation and integrals by reading the diagram from bottom-right to top-left). This Umbral technique is roughly analogous to Galois Theory5,6 insofar as we're taking taking a hard problem like integration, translating it into a comparatively easier setting of discrete summation, solving the problem there, and then exfiltrating our findings back out into the "harder" context via this ϕ operator.
As one last trivial example, we could concretely derive the formula for ∑n2 as:
n=0∑x−1=ϕ∫0xϕ−1n2dn
Nontrivial ϕ
None of these techniques are useful unless we know ϕ and ϕ−1. For polynomials discussed so far, it's relatively straightforward to derive the values:
ϕx=x1◯=xϕx2=x2◯=x(x−1)ϕx3=x3◯=x(x−1)(x−2)⋮
Using the same substitution strategy used for computing the forward difference, we can attain the inverse:
And using this method, we can find ϕ and ϕ−1 for any polynomial:
ϕx=xϕx2=x2−xϕx3=x3−3x2+2xϕx4=x4−6x3+11x2−6x
ϕ−1x=xϕ−1x2=x2+xϕ−1x3=x3+3x2+xϕ−1x4=x4+6x3+7x2+x
These are known as the Signed Stirling Numbers of the first kind, and Unsigned Stirling Numbers of the second kind – respectively.
They have some nice properties. We can compute the coefficients of a row by multiplying the negated row index by the first term and adding it to the next term. For example, the second term of the fifth row would be:
For the unsigned Stirling numbers, we compute the next set of coefficients by multiplying the corresponding coefficient from the previous row by its exponent and add it to the next term:
In order to further underscore the apparent relationship to Pascal's triangle an binomial coefficients, we express these polynomials in brackets and braces:
These sequences lie at the heart of Umbral calculus, and have natural applications in combinatorics as well.
(kn)=k!nk◯
that is, n choose k can be though of as the number of k objects we can take from n objects and by the number of permutations of k objects, the quotient of which is the number of ways to choose k objects from n objects, ignoring their orderings. We can use this identity to compute ϕ for some other functions like:
which in turn give rise to other classical results:
Dex=exDsinx=cosx⟺Δ2x=2x⟺Δϕsinx=ϕcosx
We can take the Umbral a step further, taking a step upwards in the direction of Category Theory by generalizing what we've observed about calculus operations to number theory and arithmetic.
Operational Calculus
Briefly revisiting the behavior of Umbral operations in conjunction with linear operators, we'll motivate discussion of their applicability to differential equations, recurrence relations, exponential shifting, and more.
Consider the number line of reals, together with arithmetic multiplication and addition (i.e. the field ⟨F,+,×⟩). Instead of thinking of a number as a point on the line, let's shift our representational understanding of them to be operations themselves, and addition and multiplication to be shifts and stretches applied to the numeral operators. E.g., the point 2 together with other operators, can be thought of as:
(2,5)(−1,2)×10+1
The notion of addition can then be thought of as the translation which maps 0↦2 (shifting the whole number line accordingly), same with multiplication:
two points combined via the extraneous notion of ×:
(2,3)×6
two, as a point, on which we can apply the transformation ×3:
2×36
The converse: three, as a point, on which we can apply the transformation ×2:
3×26
and finally, the composition(s) of those transformations acting on 1:
1×2×36
Considering numbers themselves as transformations acting on functions, we can also write:
(2)(3)f(x)=6f(x)
Numbers as operators (distinct from the implicit shorthand multiplication where we place numbers next to one another sans ×,⋅) are themselves linear operators since they maintain:
additivity: a(b+c)=ab+ac
and commutative under scalar multiplication
So, the above expression is more explicitly written:
2∘3∘f(x)=6∘f(x)
And, similarly, this applies to addition as well:
2+32∘f(x)+3∘f(x)=5=5∘f(x)
Examining a non-trivial linear operation which purportedly ought to follow similar rules, can we make sense of something like:
D+1
In the context of composing operators, it's useful to divorce ourselves from the specifics of their application –only bearing in mind7 the category/domain of our contextual applicability, which for us so far is just \Real:
Visually, we don't have to know or care about the values of our unlabeled points, just the transformations connecting them. Let's zoom in on the derivative operator D (also colored red to emphasize the representational equivalence to numeral operators). As established before, it's also a linear operator, preserving scalar addition and distributing & commuting under scalar multiplication:
Da+D(af(x)+bf(x))=aD=Daf(x)+Dbf(x)
Note here that a,b don't ever have to be scalars. They could also be any other linear operator. So, given an expression like:
(D+2)(D+3)
we can apply intuitive rules of algebra to derive:
(D+2)(D+3)=(D+2)3+(D+2)D=D2+2D+3D+6=D2+5D+6
Which we can apply to a second order differential equation like so:
D2f(x)+5Df(x)+6f(x)
We can just factor it into two maybe-much-simpler equations:
(Note that the φ used here is the golden ratio and its conjugate, not to be confused with our transformation ϕ). Using the same superposition principle, we get a general solution:
f(x)=φxk1+(−φ)−xk2+
And the constant are given by the initial conditions:
Which means that we can shift a differential equation to make it easier to solve by multiplying through by an exponential. This is the exponential shift theorem.
eax(D+a)y=D(eaxy)
Series Expansion
Alternatively, we could've just expanded (1−D)−1 as a geometric series:
y=(1+D+D2+D3+⋯)x3+(1−D)−10=x3+3x2+6x+6+cex
We can use this approach to express any polynomial as a Taylor Series centered at an arbitrary point a like so:
f(x+a)=k=0∑∞k!xkDkf(a)
Here, D is still differentiation with respect to x, applied to f evaluated at a. Not that the left hand side of this equation is a unit shift of x units applied to a:
Txf(a)=k=0∑∞k!xkDkf(a)
and the sum on the right is just the sum we get if we expand exD:
Tx=exD
So, just like exponentiating a complex number of the form a+bi corresponds to the composition of a stretch action and an orthogonal rotation action, so too does exponentiation of the derivative operator produce a scalar stretch and orthogonal shift:
So, if we exponentiate a+bD, we get the composition of scalar stretch and orthogonal shift:
which finds application in describing iterated functions of the form:
Blissard, John. "Theory of generic equations." Quarterly Journal of Pure Applied Math. 4, 279–305, 1861. ↩
Blissard, John. "Theory of generic equations." Quarterly Journal of Pure Applied Math. 5, 58.75, 185-208, 1862.
And it's actually despicable that I can't find a PDF of these, nor the de facto biography on him and his methods (The history of Blissard’s symbolic method with a sketch of its inventor’s life by E.T. Bell) anywhere. Aaron Swartz wept. ↩