You are currently browsing Todd Trimble’s articles.
Huh — no solutions to POW-13 came in! I guess I was surprised by that.
Ah well, that’s okay. The problem wasn’t exactly trivial; there are some fairly deep and interesting things going on in that problem that I’d like to share now. First off, let me say that the problem comes from The Unapologetic Mathematician, the blog written by John Armstrong, who posted the problem back in March this year. He in turn had gotten the problem from Noam Elkies, who kindly responded to some email I wrote and had some pretty insightful things to say.
In lieu of a reader solution, I’ll give John’s solution first, and then mine, and then touch on some of the things Noam related in his emails. But before we do so, let me paraphrase what John wrote at the end of his post:
Here’s a non-example. Pick
points
and so on up to
. Pick
points
,
,
, and so on up to
. In this case we have
blocking points at
,
, and so on by half-integers up to
. Of course this solution doesn’t count because the first
points lie on a line as do the
points that follow, which violates the collinearity condition of the problem.
Here’s a picture of that scenario when :
Does this configuration remind you of anything? Did somebody say “Pappus’s theorem“? Good. Hold that thought please.
Okay, in the non-example the first points were on two lines, which is disallowed. Now the two lines here,
and
, form a degenerate conic
. Thinking in good algebraic geometry fashion, perhaps we can modify the non-solution by replacing the degenerate conic by an honest (smooth) nondegenerate conic, like an ellipse or something, so that at most two of the
points are on any given line. This should put one in the mood for our first solution.
John Armstrong writes: Basically, I took the non-example and imagined bending back the two lines to satisfy the collinearity (pair-of-lines = degenerate conic, so non-example is degenerate example). The obvious pair of curves to use is the two branches of a hyperbola. But hyperbolas can be hard to work with, so I decided to do a projective transformation to turn it into a parabola.
The line itself is . Which has the obvious
-intercept
. Now we need to pick a lot of
and
values to get repeated products. Place
points at
,
,
,
, and so on above the negative powers of two. Place
points at
,
,
,
, and so on above the positive powers of two. The blocking points are then at
,
,
,
, and so on up the
-axis by powers of two. Presto!
Very nice. My own solution was less explicit in the sense that I didn’t actually write down coordinates of points, but gave instead a general recipe which relies instead on the geometry of conics, in fact on a generalization of Pappus’s theorem known to me as “Pascal’s mystic hexagon“. I first learned about this from a wonderful book:
- C. Herbert Clemens, A Scrapbook of Complex Curve Theory (2nd Edition), Graduate Studies in Mathematics 55, AMS (2002).
Pascal’s Mystic Hexagon, version A: Consider any hexagon inscribed in a conic, with vertices marked . For
, mark the intersection of the lines
by
. Then
are collinear. (The reason for the strange indexing will be clear in a moment.)
Pascal’s Mystic Hexagon, version B: Consider any pentagon inscribed in a conic , with vertices marked
. Choose any line
through
, and define
and
. Then the intersection
is the sixth point of a hexagon inscribed in
.
The following solution uses version B.
Solution by Todd and Vishal: For the sake of explicitness, let the conic be a circle
and let the line (segment)
be the diameter along
. Choose two points
on
above
and a point
on
below
. The remaining points are determined recursively by a zig-zag procedure, where at each stage
is the intersection
, where
, where
, and
. We will show the blocking condition is satisfied: for all
, the point
is on the line
.
To get started, notice that the blocking condition is trivially satisfied up to the point where we construct . Mystic Hexagon B ensures that
, as defined above, is blocked from
by
and from
by
. Then define
as above. So far the blocking condition is satisfied.
Suppose the blocking condition is satisfied for the points . Define
, as above. Then Mystic Hexagon B, applied to the pentagon consisting of points
, shows that
is blocked from
by
and from
by
.
This shows that could have been defined to be
. Then Mystic Hexagon B, applied to the pentagon
, shows that
is blocked from
by
and from
by
. This shows
could have been defined to be
.
And so on up the inductive ladder: for , defining
, Hexagon B applied to the pentagon
shows that
could have been defined to be
. This shows the blocking condition is satisfied for
,
. We block
from
by defining
as above, to extend up to
.
Then, as prescribed above, we define , and repeat the preceding argument mutatis mutandis (interchanging
‘s and
‘s), to obtain the blocking conditions up to
,
. This completes the proof.
What this argument shows (with Hexagon B doing all the heavy lifting) is that no cleverness whatsoever is required to construct the desired points: starting with any nondegenerate conic , any secant line
, and any three initial points
on
to get started (with
‘s on one side of
and
on the other), the whole construction is completely forced and works no matter what!
Which may lead one to ask: what is behind this “miracle” called Pascal’s Mystic Hexagon? Actually, not that much! Let me give a seat-of-the-pants argument for why one might expect it to hold, and then give a somewhat more respectable argument.
Define a planar cubic to be the locus of any degree 3 polynomial
in two variables. (We should actually be doing this as projective geometry, so I ought to write
where
is homogeneous of degree 3, but I think I’ll skip that.) For example, the union of three distinct lines is a cubic where
is a product of three degree 1 polynomials. What is the dimension of the space of planar cubics? A cubic polynomial
,
has 10 coefficients. But the equation is equivalent to the equation
for any nonzero scalar
; modding out by scalars, there are 9 degrees of freedom in the space of cubics. What is the dimension of the space of cubics passing through a given point
? The condition
gives one linear equation on the coefficients
, so we cut down a degree of freedom, and the dimension would be 8. Similarly, if we ask for the dimension of cubics passing through 8 given points, we get a system of eight linear equations, and we cut down by eight degrees of freedom: in general, the space of cubics through 8 points is “expected” to be (is “almost always”) 1-dimensional, in fact, a (projective) line.
In the configuration for Pascal’s Mystic Hexagon, version A
we see three cubics passing through the 8 points , namely:
where the conic
is defined by a degree 2 polynomial
Since we expect that the space of cubics through these eight points is a line, we should have a linear relationship between the cubic polynomials used respectively to define
, and
above, hence we would get
for some scalars . Thus, if a ninth point
is in
and
, so that
, then
. Thus
lies in
as well, and if
isn’t on
, it must be on the line
. Thus, “generically” we expect
to be collinear, whence Hexagon A.
This rough argument isn’t too far removed from a slightly more rigorous one. There’s a general result in projective algebraic geometry called Bézout’s theorem, which says that a degree planar curve and a degree
planar curve either intersect in
points (if you count them right, “with multiplicity”) or they have a whole curve component in common. (Fine print: to make this generally true, you have to work in the projective plane, and you have to work over an algebraically closed field.) A much weaker result which removes all the fine print is that a degree
curve and a degree
curve either have a curve component in common, or they intersect in at most
points. In particular, in the notation above, the cubics
and
intersect in 9 points, 6 of which are on the conic
. Pick a seventh point
on
, away from those six, and let
and
. Then we see that the locus of the degree 3 polynomial
intersects the degree 2 conic in at least 7 points (namely,
and
), greater than the expected number
, which is impossible unless the loci of
and
have a component in common. But the conic
has just one component — itself — so one can conclude that its defining degree 2 polynomial (I’ll call it
) must divide
. Then we have
for some degree 1 polynomial , so the last three of the nine points of intersection
, which are zeroes of
and
, must be zeroes of the linear polynomial
, and hence are collinear. Thus we obtain Pascal’s Mystic Hexagon, version A.
It’s clear then that what makes the Mystic Hexagon tick has something to do with the geometry of cubic curves. With that in mind, I’m now going to kick the discussion up a notch, and relate a third rather more sophisticated construction on cubics which basically subsumes the first two constructions. It has to do with so-called “elliptic curves“.
Officially, an elliptic curve is a smooth projective (irreducible) cubic “curve” over the complex numbers. I put “curve” in quotes because while it is defined by an equation where
is a polynomial of degree 3, the coefficients of
are complex numbers as are the solutions to this equation. We say “curve” in the sense that locally it is like a “line”, but this is the complex line
we’re talking about, so from our real number perspective it is two-dimensional — it actually looks more like a surface. Indeed, an elliptic curve is an example of a Riemann surface. It would take me way too far afield to give explanations, but when you study these things, you find that elliptic curves are Riemann surfaces of genus 1. In more down-to-earth terms, this means they are tori (toruses), or doughnut-shaped as surfaces. Topologically, such tori or doughnuts are cartesian products of two circles.
Now a circle or 1-dimensional sphere carries a continuous (abelian) group structure, if we think of it as the set of complex numbers
of norm 1, where the group operation is complex multiplication. A torus
also carries a group structure, obtained by multiplying in each of the two components. Thus, given what we have said, an elliptic curve also carries a continuous group structure. But it’s actually much better than that: one can define a group structure on a smooth complex cubic
(in the complex plane
, or rather the projective complex plane
) not just by continuous operations, but by polynomially defined operations, and the definition of the group law is just incredibly elegant as a piece of geometry. Writing the group multiplication as addition, it says that if
are points on
, then
if are collinear. [To be precise, one must select a point 0 on
to serve as identity, and this point must be one of the nine inflection points of
. When
and
coincide (are “infinitesimally close”), the line through
and
is taken to be tangent to
; when
coincide, this is a line of inflection.]
This is rather an interesting thing to prove, that this prescription actually satisfies the axioms for an abelian group. The hardest part is proving associativity, but this turns out to be not unlike what we did for Pascal’s Mystic Hexagon: basically it’s an application of Bézout’s theorem again. (In algebraic geometry texts, such as Hartshorne’s famous book, the discussion of this point can be far more sophisticated, largely because one can and does define elliptic curves as certain abstract 1-dimensional varieties or schemes which have no presupposed extrinsic embeddings as cubic curves in the plane, and there the goal is to understand the operations intrinsically.)
In the special case where is defined by a cubic polynomial with real coefficients, we can look at the locus of real solutions (or “real points”), and it turns out that this prescription for the group law still works on the real locus, in particular is still well-defined. (Basically for the same reason that if you have two real numbers
which are solutions to a real cubic equation
, then there is also a third real solution
.) There is still an identity element, which will be an inflection point of the cubic.
Okay, here is a third solution to the problem, lifted from one of Noam Elkies’ emails. (The original formulation of the problem spoke in terms of “red” points (instead of my
),
“blue” points (instead of my
), and “blocking” points which play the role of my
.) The addition referred to is the addition law on an elliptic curve. I’ve taken the liberty of paraphrasing a bit.
“Choose points on the real points of an elliptic curve such that
is in-between
and
. Then set
- red points:
,
- blue points:
,
- blocking points:
,
where is a real point on the elliptic curve very close to the identity. The pair
,
is blocked by
, because these three points are collinear, and the smallness of P guarantees that the blocking point is actually between the red point and blue point, by continuity.”
Well, well. That’s awfully elegant. (According to Noam’s email, it came out of a three-way conversation between Roger Alperin, Joe Buhler, and Adam Chalcraft. Edit: Joe Buhler informs me in email that Joel Rosenberg’s name should be added. More at the end of this post.) Noam had given his own slick solution where again the red and blue points sit on a conic and the blocking points lie on a line not tangent to the conic, and he observed that his configuration was a degenerate cubic, leading him to surmise that his example could in a sense be seen as a special case of theirs.
How’s that? The last solution took place on a smooth (nondegenerate) cubic, so the degenerate cubic = conic+line examples could not, literally speaking, be special cases. Can the degenerate examples be seen in terms of algebraic group structures based on collinearity?
The answer is: yes! As you slide around in the space of planar cubics, nondegenerate cubics (the generic or typical case) can converge to cubics which are degenerate in varying degrees (including the case of three lines, or even a triple line), but the group laws on nondegenerate cubics based on collinearity converge to group laws, even in degenerate cases! (I hadn’t realized that.) You just have to be careful and throw away the singular points of the degenerate cubic, but otherwise you can basically still use the definition of the group law based on collineation, although it gets a little tricky saying exactly how you’re supposed to add points on a line component, such as the line of conic+line.
So let me give an example of how it works. It seems convenient for this purpose to use John Armstrong’s model which is based on the parabola+line, specifically the locus of . The singular points of its projective completion are at
and the point where the
-axis meets the line at infinity. After throwing those away, what remains is a disjoint union of four pieces: right half of parabola, left half of parabola, positive
-axis, negative
-axis.
We can maybe guess that since implies
collinear, that the two pieces of the
-axis form a subgroup for the group law we are after (also, these two pieces together should suggest the two halves of the multiplicative group of nonzero reals
, but don’t jump to conclusions how this works!). If so, then we notice that if
and
lie on the parabola, then the line between them intersects the
-axis at a point
, so then the parabolic part would not be closed under multiplication.
One is then led to consider that the group structure of this cubic overall is isomorphic to the group , with the linear part identified somehow with the subgroup
, and the parabolic part with
.
I claim that the abelian group structure on the punctured -axis should be defined by
so that the identity element on the cubic is , and the inverse of
is
. The remainder of the abelian group structure on the cubic is defined as follows:
Undoubtedly this group law looks a bit strange! So let’s do a spot check. Suppose ,
, and
are collinear. Then it is easily checked that
, and each of the two equations
is correct according to the group law, so the three collinear points do add to the identity and everything checks out.
All right, let’s retrieve John’s example as a special case. Take as a red point,
as a blue point, and
as blocking point. Take a point
“sufficiently close” to the identity, say
. Then
which was John’s solution.
Another long post from yours truly. I was sorry no solutions came from our readers, but if you’d like another problem to chew on, here’s a really neat one I saw just the other day. Feel free to discuss in comments!
Exercise: Given a point on a circle, show how to draw a tangent to the point using ruler only, no compass. Hint: use a mystic hexagon where two of the points are “infinitesimally close”.
Added July 25: As edited in above, Joel Rosenberg also had a hand in the elliptic curves solution, playing an instrumental role in establishing some of the conjectures, such as that is the minimal number of blocking points under certain assumptions, and (what is very nice) that the elliptic curves solution is the most general solution under certain assumptions. I thank Joe Buhler for transmitting this information.
It’s been an awfully long time since I’ve posted anything; time to finally break the silence.
This problem appeared elsewhere on the internet some months ago; some of you may have already seen it. I don’t want to say right away where I saw it, because there was some commentary which included some rough hints which I don’t want to give, but I’ll be sure to give credit when the solution is published. I’ll bet some of you will be able to find a solution, and will agree it’s quite cute. Here it is:
Given integers
, show that it is possible to construct a set of
points in the plane, let’s say
, so that no three points of the set are collinear, and for which there exist points
, all lying on a straight line, and arranged so that on the line between any
and any
, some
lies between them.
So no can “see” any
, because there’s always some
blocking the view. As the proposer observed, the problem would be easy if we had
‘s to play with, one for each pair
. But here there are only
‘s, so some of them will have to do more than their fair share, blocking the view between quite a few
-pairs simultaneously. Thus, you have to arrange the
‘s and
‘s so that a lot of the lines between them will be coincident at a point
, and subject to the constraint I italicized above.
Please submit solutions to topological[dot]musings[At]gmail[dot]com by Friday, July 17, 11:59 pm (UTC); do not submit solutions in Comments. Everyone with a correct solution will be inducted into our Hall of Fame! We look forward to your response.
This being a mathematics blog, I’m sure a lot of readers out like to play Minesweeper. I’ve just obtained a personal best today (94 seconds on expert level) which, as Minesweeper buffs know, is nowhere close to world-class levels, but which made me happy anyway, as I’d never broken into the double digits before today!
It seems to me that world-class competitors must know some tricks with the mouse which I’ve never bothered to master, particularly because my laptop doesn’t have a mouse but rather a touchpad. This being the case, I keep my right index finger on the touchpad to guide the cursor, and the left index finger to click. I always left-click: that is, in my style of play, I [practically] never flag squares for bombs; I click only on non-bomb squares. For it’s well-known, or at least it should be, that the program doesn’t care if you identify where the bombs are — you get full credit for only identifying all the numbered squares.
To play in this style well, one needs to be fluent in a number of tactical tricks, which I don’t have good names for, but which in my personal argot I call things like “1-2-1”, “1-2-2-1”, “rule of three”, to name just a few. But that’s not what I set out to discuss, really. What I’d really like to hear from good players is: what opening strategies do you use?
The personal best I just set came after deciding on a new opening strategy. What I had been doing is clicking along border squares. Under that strategy, one could of course just keep clicking until one opens up an area, but often I would add to that the observation that if one clicked on a 1, thus leading to, e.g.,
x 1 x (–> border row)
x x x
then one could then click on the non-border square just adjacent to the 1, with only a 1 in 5 chance of setting off a bomb. If one then hits another 1:
x 1 x
x 1 x
x x x
then one can immediately open up the line of squares on the third rank, leading to a configuration such as
x 1 x
x 1 x
1 1 1
or better. This is often a cheap and quick way of opening up squares or otherwise getting a tactical toehold.
The new strategy I began using today is not to click along the border, but to click along the ranks or files adjacent to the border. Under this strategy, if one lands on a 1, leading to
x x x (–> border row)
x 1 x
x x x
then one can click on the border square adjacent to the 1, with only a 1 in 8 chance of setting off a bomb. If one does not set off a bomb, that square has to be a 1:
x 1 x
x 1 x
x x x
and then one can proceed as before. So I’ve just lowered my odds of hitting a bomb, plus a very small fractional gain in processing time that comes with the certain knowledge that it’s a 1 if not a bomb. So far the strategy has paid off well!
I’d like to hear other people’s opening strategies, and also I’d like to know some statistics. For example, considered over the space of expert-level games, what is the probability of getting a 1, a 2, and so on? Does anyone know? (It seems this would be very difficult computing analytically — one is probably better off using a Monte Carlo simulation. But I don’t have the wherewithal to set that kind of thing up.)
The solutions are in! This problem of the week was interesting for me: the usual pattern has been that I pose problems that I’ve solved myself at some point in the past, giving me a kind of “inside edge” on understanding the solutions that come in. But, as I said in POW-12, the difference this time is that the solution I knew of came from someone else (Arin Chaudhuri). What was interesting for me — and given the circumstances, it was probably inevitable — is that some of the solutions we received forced me to make contact with some mathematics I’d only heard about but never studied. Let me get to that in a minute.
Another interesting thing for me — speaking now with my category theorist’s hat on — is how utterly simple and conceptual Arin’s proof was! I was pleased to see that regular problem solver Philipp Lampe also spotted it. Wait for it…
Solution I by Philipp Lampe, University of Bonn: The answer is 8.
Among the eight neighbors of an arbitrary vertex, all colors of an admissible coloring must occur. Thus, 8 is an upper bound for the maximum number of colors one can use. We have to show that there is an admissible coloring with eight colors.
The vertices of the 8-dimensional cube may be represented by vectors with
in
, in other words as vectors in
, the 8-dimensional vector space over the field
with two elements (i.e., the integers modulo 2). Two such vectors
are neighbors iff their coordinate vectors differ in exactly one place, in other words if
considered modulo 2, where
is one of the eight standard basis elements (with
-th coordinate 1, and the other coordinates 0).
Now let our “colors” be the 8 elements of
, the 3-dimensional vector space over
. Let the vertex “coloring”
be the unique -linear map such that
; that is, define the color of a vertex/vector by
Now, if is any vector, the colors of its neighbors are
. The colors of these neighbors are all distinct since the
are distinct. Hence all 8 colors are represented among the colors of the neighbors of any given vertex
, QED.
What I love about this solution it is how natural it is. I’ll say a little more about this in remarks below.
But I also learned a thing or two by studying the next solution. It relies on the theory of Hamming codes, with which I was not conversant. Up to some small edits, here is exactly what Sune Jakobsen submitted:
Solution II by Sune Jakobsen (first-year student), University of Copenhagen: Since each vertex only has 8 neighbors, the answer cannot be greater that 8.
Now we construct such a coloring with 8 colors. An 8-dimensional cube can be represented by the graph with vertices in , and with an edge between them iff the Hamming distance between them is 1. We color a vertex with color 8 if the seven first bits in the vertex is a “correct” Hamming message (cf. Hamming code (7,4)), and color it in color
if the first seven bits give a correct Hamming message upon changing bit
. This is a well-defined coloring, since each element in
is either a correct Hamming message, or is Hamming distance 1 away to exactly one correct Hamming message.
It remains to show that no vertex is neighbor to two vertices of the same color. The Hamming distance between these two vertices is 2, thus the Hamming distance between the first 7 bits of two neighbors must be 1 or 2. If two neighbors had the same color , then by definition one would get two correct Hamming messages by changing bit
in both of them, and the Hamming distance between these messages would be the same as before — either 1 or 2. But the distance between any two correct Hamming messages is at least 3. Contradiction.
Remarks:
1. Let me give a little background to Sune’s solution. Mathematically, the Hamming code called “(7, 4)” is the image of injective linear map
given by the matrix
1101
1011
1000
0111
0100
0010
0001
The Hamming code is what is known as an “error-correcting code”. Imagine that you want to send a 4-bit message (each bit being a 0 or 1) down a transmission line, but that due to noise in the line, there is the possibility of a bit being flipped and the wrong message being received. What can you do to help ensure that the intended message is received?
The answer is to add some “parity check bits” to the message. In the code (7, 4), one adds in three parity checks, so that the transmitted message is 7 bits long. What one does is apply the matrix to the 4-vector, to get a 7-vector (remember we’re working modulo 2), and this 7-vector is sent down the line. Assuming only a moderate amount of noise in the transmission line, perhaps the 7-bit message will remain intact or perhaps a single bit will be flipped, more rarely two bits will be flipped (and we’ll assume the event that more than two are flipped has negligible probability). Now, the Hamming encoding
is rigged up so that if the 7-bit vector is received as sent, then the parity checks will report 0 errors. If the parity checks report an error, they report precisely where the error occurred if the received vector was off by one bit. (If two flips occurred, then the receiver can still detect from the parity checks that an error occurred. The parity checks can never detect how many bits were flipped, but if the receiver assumes correctly that just one bit got flipped, he can restore the intended message with accuracy. If three or more got flipped, then the receiver got the wrong message, but he would never know it if the parity checks reported back 0 errors.)
How does this work? By having the receiver apply a parity-check matrix to the received message, given by the
matrix
1010101
0110011
0001111
In the first place, , and so if the message received belongs to the image of
(is a “correct Hamming message” in the terminology of Solution II), which will indeed be the case if there were no errors in the transmission, then
applied to the message will produce the zero vector. In the case where one bit got flipped in the transmission,
applied to the received vector will return a nonzero 3-vector, but the beauty of the Hamming code is that the 3-vector will spell out in binary the bit the error occurred (for example, if the output vector is
, then error occurred in the bit with binary number 011, that is bit 3). In that case, the receiver flips that bit to restore the original message. In these two cases, the original 4 bits are then read off as the 3rd, 5th, 6th, and 7th coordinates of the (possibly corrected) 7-vector.
By the way, Sune reminded us that Hamming codes also came up in a post by Greg Muller over at The Everything Seminar, who used the existence of a Hamming code in every dimension to solve general hat-guessing puzzles.
2. Within minutes of posting the original problem, we received a message from David Eppstein, who reported that the solution of 8 colors is essentially contained in a paper he coauthored with T. Givaris (page 4); his solution is close in spirit to Sune’s.
Arin Chaudhuri also surmised the connection to Hamming codes, and mentions that the problem (in a slightly different formulation) originally came from a friend of a friend, who blogs about a number of related problems on this page. Presumably the author had things like error-correcting codes in mind.
3. Sune and David noted that their solutions generalize to show that on the -dimensional cube (for any
), it is possible to color the vertices with
colors (the maximum number possible) so that all of the colors show up as colors of the neighbors of any given vertex.
This is very easy to see by adapting Philipp’s method. Indeed, for each just take the color set to be the
-vector space
. The set of vertices of the
-dimensional cube may identified with the
-vector space
generated by the set
(as basis), and the desired coloring is then just the
-linear map
that extends the identity function on
. As mathematical constructions go, you can’t get much more canonical than that! No fuss, no muss — it’s a purely categorical construction (categorists call it a counit of an adjunction). So then why didn’t I see that myself? 🙂
4. We’re still not sure what the story is in other dimensions. If is the maximum number of admissible colors in dimension
, then about all any one of us knows right now is that
for
,
is weakly monotone increasing, and that
if
is not a power of 2. There’s a conjecture afloat that
where
is the floor function, but for now that should be treated as a wild guess. If anyone has further information, please let us know!
Solved by Arin Chaudhuri, David Eppstein, Sune Jakobsen, and Philipp Lampe. Thanks to all those who wrote in!
Happy holidays, folks! And happy birthday to Topological Musings, which Vishal started just a little over a year ago — we’ve both had a really good time with it.
And in particular with the Problem of the Week series. Which, as we all know, doesn’t come out every week — but then again, to pinch a line from John Baez’s This Week’s Finds, we never said it would: only that each time a new problem comes out, it’s always the problem for that week! Anyway, we’ve been very gratified by the response and by the ingenious solutions we routinely receive — please keep ’em coming.
This problem comes courtesy of regular problem-solver Arin Chaudhuri and — I’ll freely admit — this one defeated me. I didn’t feel a bit bad though when Arin revealed his very elegant solution, and I’d like to see what our readers come up with. Here it is:
What is the maximum number of colors you could use to color the vertices of an 8-dimensional cube, so that starting from any vertex, each color occurs as the color of some neighbor of that vertex? (Call two vertices neighbors if they are the two endpoints of an edge of the cube.)
Arin, Vishal, and I would be interested and pleased if someone solved this problem for all dimensions .
Please submit solutions to topological[dot]musings[At]gmail[dot]com by Friday, January 2, 2009, 11:59 pm (UTC); do not submit solutions in Comments. Everyone with a correct solution will be inducted into our Hall of Fame! We look forward to your response.
After a long hiatus, I’d like to renew the discussion of axiomatic categorical set theory, more specifically the Elementary Theory of the Category of Sets (ETCS). Last time I blogged about this, I made some initial forays into “internalizing logic” in ETCS, and described in broad brushstrokes how to use that internal logic to derive a certain amount of the structure one associates with a category of sets. Today I’d like to begin applying some of the results obtained there to the problem of constructing colimits in a category satisfying the ETCS axioms (an ETCS category, for short).
(If you’re just joining us now, and you already know some of the jargon, an ETCS category is a well-pointed topos that satisfies the axiom of choice and with a natural numbers object. We are trying to build up some of the elementary theory of such categories from scratch, with a view toward foundations of mathematics.)
But let’s see — where were we? Since it’s been a while, I was tempted to review the philosophy behind this undertaking (why one would go to all the trouble of setting up a categories-based alternative to ZFC, when time-tested ZFC is able to express virtually all of present-day mathematics on the basis of a reasonably short list of axioms?). But in the interest of time and space, I’ll confine myself to a few remarks.
As we said, a chief difference between ZFC and ETCS resides in how ETCS treats the issue of membership. In ZFC, membership is a global binary relation: we can take any two “sets” and ask whether
. Whereas in ETCS, membership is a relation between entities of different sorts: we have “sets” on one side and “elements” on another, and the two are not mixed (e.g., elements are not themselves considered sets).
Further, and far more radical: in ETCS the membership relation is a function, that is, an element
“belongs” to only one set
at a time. We can think of this as “declaring” how we are thinking of an element, that is, declaring which set (or which type) an element is being considered as belonging to. (In the jargon, ETCS is a typed theory.) This reflects a general and useful philosophic principle: that elements in isolation are considered inessential, that what counts are the aggregates or contexts in which elements are organized and interrelated. For instance, the numeral ‘2’ in isolation has no meaning; what counts is the context in which we think of it (qua rational number or qua complex number, etc.). Similarly the set of real numbers has no real sense in isolation; what counts is which category we view it in.
I believe it is reasonable to grant this principle a foundational status, but: rigorous adherence to this principle completely changes the face of what set theory looks like. If elements “belong” to only one set at a time, how then do we even define such basic concepts as subsets and intersections? These are some of these issues we discussed last time.
There are other significant differences between ZFC and ETCS: stylistically, or in terms of presentation, ZFC is more “top-down” and ETCS is more “bottom-up”. For example, in ZFC, one can pretty much define a subset by writing down a first-order formula
in the language; the comprehension (or separation) axiom scheme is a mighty sledgehammer that takes care of the rest. In the axioms of ETCS, there is no such sledgehammer: the closest thing one has to a comprehension scheme in the ETCS axioms is the power set axiom (a single axiom, not an axiom scheme). However, in the formal development of ETCS, one derives a comprehension scheme as one manually constructs the internal logic, in stages, using the simple tools of adjunctions and universal properties. We started doing some of that in our last post. So: with ZFC it’s more as if you can just hop in the car and go; with ETCS you build the car engine from smaller parts with your bare hands, but in the process you become an expert mechanic, and are not so rigidly attached to a particular make and model (e.g., much of the theory is built just on the axioms of a topos, which allows a lot more semantic leeway than one has with ZF).
But, in all fairness, that is perhaps the biggest obstacle to learning ETCS: at the outset, the tools available [mainly, the idea of a universal property] are quite simple but parsimonious, and one has to learn how to build some set-theoretic and logical concepts normally taken as “obvious” from the ground up. (Talk about “foundations”!) On the plus side, by building big logical machines from scratch, one gains a great deal of insight into the inner workings of logic, with a corresponding gain in precision and control and modularity when one would like to use these developments to design, say, automated deduction systems (where there tend to be strong advantages to using type-theoretic frameworks).
Enough philosophy for now; readers may refer to my earlier posts for more. Let’s get to work, shall we? Our last post was about the structure of (and relationships between) posets of subobjects relative to objects
, and now we want to exploit the results there to build some absolute constructions, in particular finite coproducts and coequalizers. In this post we will focus on coproducts.
Note to the experts: Most textbook treatments of the formal development of topos theory (as for example Mac Lane-Moerdijk) are efficient but highly technical, involving for instance the slice theorem for toposes and, in the construction of colimits, recourse to Beck’s theorem in monad theory applied to the double power-set monad [following the elegant construction of Paré]. The very abstract nature of this style of argumentation (which in the application of Beck’s theorem expresses ideas of fourth-order set theory and higher) is no doubt partly responsible for the somewhat fearsome reputation of topos theory.
In these notes I take a much less efficient but much more elementary approach, based on an arrangement of ideas which I hope can be seen as “natural” from the point of view of naive set theory. I learned of this approach from Myles Tierney, who was my PhD supervisor, and who with Bill Lawvere co-founded elementary topos theory, but I am not aware of any place where the details of this approach have been written up before now. I should also mention that the approach taken here is not as “purist” as many topos theorists might want; for example, here and there I take advantage of the strong extensionality axiom of ETCS to simplify some arguments.
The Empty Set and Two-Valued Logic
We begin with the easy observation that a terminal category, i.e., a category with just one object and one morphism (the identity), satisfies all the ETCS axioms. Ditto for any category
equivalent to
(where every object is terminal). Such boring ETCS categories are called degenerate; obviously our interest is in the structure of nondegenerate ETCS categories.
Let be an ETCS category (see here for the ETCS axioms). Objects of
are generally called “sets”, and morphisms are generally called “functions” or “maps”.
Proposition 0: If an ETCS category is a preorder, then
is degenerate.
Proof: Recall that a preorder is a category in which there is at most one morphism for any two objects
. Every morphism in a preorder is vacuously monic. If there is a nonterminal set
, then the monic
to any terminal set defines a subset
distinct from the subset defined by
, thus giving (in an ETCS category) distinct classifying maps
, contradicting the preorder assumption. Therefore all objects
are terminal.
Assume from now on that is a nondegenerate ETCS category.
Proposition 1: There are at least two truth values, i.e., two elements , in
.
Proof: By proposition 0, there exist sets and two distinct functions
. By the axiom of strong extensionality, there exists
such that
. The equalizer
of the pair
is then a proper subset of
, and therefore there are at least two distinct elements
.
Proposition 2: There are at most two truth values ; equivalently, there are at most two subsets of
.
Proof: If are distinct subsets of
, then either
or
, say the former. Then
and
are distinct subsets, with distinct classifying maps
. By strong extensionality, there exists
distinguishing these classifying maps. Because
is terminal, we then infer
and
, so
as subsets of
, and in that case only
can be a proper subset of
.
By propositions 1 and 2, there is a unique proper subset of the terminal object . Let
denote this subset. Its domain may be called an “empty set”; by the preceding proposition, it has no proper subsets. The classifying map
of
is the truth value we call “false”.
Proposition 3: 0 is an initial object, i.e., for any there exists a unique function
.
Proof: Uniqueness: if are maps, then their equalizer
, which is monic, must be an isomorphism since 0 has no proper subsets. Therefore
. Existence: there are monos
where is “global truth” (classifying the subset
) on
and
is the “singleton mapping
” on
, defined as the classifying map of the diagonal map
(last time we saw
is monic). Take their pullback. The component of the pullback parallel to
is a mono
which again is an isomorphism, whence we get a map
using the other component of the pullback.
Remark: For the “purists”, an alternative construction of the initial set 0 that avoids use of the strong extensionality axiom is to define the subset to be “the intersection all subsets of
“. Formally, one takes the extension
of the map
where the first arrow represents the class of all subsets of , and the second is the internal intersection operator defined at the end of our last post. Using formal properties of intersection developed later, this intersection
has no proper subsets, and then the proof of proposition 3 carries over verbatim.
Corollary 1: For any , the set
is initial.
Proof: By cartesian closure, maps are in bijection with maps of the form
, and there is exactly one of these since 0 is initial.
Corollary 2: If there exists , then
is initial.
Proof: The composite of followed by
is
, and
followed by
is also an identity since
is initial by corollary 1. Hence
is isomorphic to an initial object
.
By corollary 2, for any object the arrow
is vacuously monic, hence defines a subset.
Proposition 4: If , then there exists an element
.
Proof: Under the assumption, has at least two distinct subsets:
and
. By strong extensionality, their classifying maps
are distinguished by some element
.
External Unions and Internal Joins
One of the major goals in this post is to construct finite coproducts in an ETCS category. As in ordinary set theory, we will construct these as disjoint unions. This means we need to discuss unions first; as should be expected by now, in ETCS unions are considered locally, i.e., we take unions of subsets of a given set. So, let be subsets.
To define the union , the idea is to take the intersection of all subsets containing
and
. That is, we apply the internal intersection operator (constructed last time),
to the element that represents the set of all subsets of
containing
and
; the resulting element
represents
. The element
corresponds to the intersection of two subsets
Remark: Remember that in ETCS we are using generalized elements:
really means a function
over some domain
, which in turn classifies a subset
. On the other hand, the
here is a subset
. How then do we interpret the condition “
“? We first pull back
over to the domain
; that is, we form the composite
, and consider the condition that this is bounded above by
. (We will write
, thinking of the left side as constant over
.) Externally, in terms of subsets, this corresponds to the condition
.
We need to construct the subsets . In ZFC, we could construct those subsets by applying the comprehension axiom scheme, but the axioms of ETCS have no such blanket axiom scheme. (In fact, as we said earlier, much of the work on “internalizing logic” goes to show that in ETCS, we instead derive a comprehension scheme!) However, one way of defining subsets in ETCS is by taking loci of equations; here, we express the condition
, more pedantically
or
, as the equation
where the right side is the predicate “true over “.
Thus we construct the subset of
via the pullback:
{C: A ≤ C} -------> 1 | | | | t_X V chi_A => - V PX -----------> PX
Let me take a moment to examine what this diagram means exactly. Last time we constructed an internal implication operator
and now, in the pullback diagram above, what we are implicitly doing is lifting this to an operator
The easy and cheap way of doing this is to remember the isomorphism we used last time to uncover the cartesian closed structure, and apply this to
to define . This map classifies a certain subset of
, which I’ll just write down (leaving it as an exercise which involves just chasing the relevant definitions):
Remark: Similarly we can define a meet operator
by exponentiating the internal meet
. It is important to know that the general Heyting algebra identities which we established last time for
lift to the corresponding identities for the operators
on
. Ultimately this rests on the fact that the functor
, being a right adjoint, preserves products, and therefore preserves any algebraic identity which can be expressed as a commutative diagram of operations between such products.
Hence, for the fixed subset (classified by
), the operator
classifies the subset
Finally, in the pullback diagram above, we are pulling back the operator against
. But, from last time, that was exactly the method we used to construct universal quantification. That is, given a subset
we defined to be the pullback of
along
. Putting all this together, the pullback diagram above expresses the definition
that one would expect “naively”.
Now that all the relevant constructions are in place, we show that is the join of
and
in the poset
. There is nothing intrinsically difficult about this, but as we are still in the midst of constructing the internal logic, we will have to hunker down and prove some logic things normally taken for granted or zipped through without much thought. For example, the internal intersection operator was defined with the help of internal universal quantification, and we will need to establish some formal properties of that.
Here is a useful general principle for doing internal logic calculations. Let be the classifying map of a subset
, and let
be a function. Then the composite
classifies the subset
so that one has the general identity . In passing back and forth between the external and internal viewpoints, the general principle is to try to render “complicated” functions
into a form
which one can more easily recognize. For lack of a better term, I’ll call this the “pullback principle”.
Lemma 1: Given a relation and a constant
, there is an inclusion
as subsets of . (In traditional logical syntax, this says that for any element
,
implies
as predicates over elements . This is the type of thing that ordinarily “goes without saying”, but which we actually have to prove here!)
Proof: As we recalled above, was defined to be
, the pullback of global truth
along the classifying map
. Hold that thought.
Let
be the map which classifies the subset . Equivalently, this is the map
under the canonical isomorphisms ,
. Intuitively, this maps
, i.e., plugs an element
into an element
.
Using the adjunction of cartesian closure, the composite
transforms to the composite
so by the pullback principle, classifies
.
Equivalently,
Also, as subsets of , we have the inclusion
[this just says that belongs to the subset classified by
, or equivalently that
is in the subset
]. Applying the pullback operation
to (2), and comparing to (1), lemma 1 follows.
Lemma 2: If as subsets of
, then
.
Proof: From the last post, we have an adjunction:
if and only if
for any subset of . So it suffices to show
. But
where the first inclusion follows from .
Next, recall from the last post that the internal intersection of was defined by interpreting the following formula on the right:
Lemma 3: If , then
.
Proof: classifies the subset
, i.e.,
is identified with the predicate
in the argument
, so by hypothesis
as predicates on
. Internal implication
is contravariant in the argument
[see the following remark], so
Now apply lemma 2 to complete the proof.
Remark: The contravariance of
, that is, the fact that
implies
is a routine exercise using the adjunction [discussed last time]
if and only if
Indeed, we have
where the first inequality follows from the hypothesis
, and the second follows from
. By the adjunction, the inequality (*) implies
.
Theorem 1: For subsets of
, the subset
is an upper bound of
and
, i.e.,
.
Proof: It suffices to prove that , since then we need only apply lemma 3 to the trivially true inclusion
to infer , and similarly
. (Actually, we need only show
. We’ll do that first, and then show full equality.)
The condition we want,
is, by the adjunction , equivalent to
which, by a –
adjunction, is equivalent to
as subsets of . So we just have to prove (1). At this point we recall, from our earlier analysis, that
Using the adjunction , as in the proof of lemma 2, we have
which shows that the left side of (1) is contained in
where the last inclusion uses another –
adjunction. Thus we have established (1) and therefore also the inclusion
Now we prove the opposite inclusion
that is to say
Here we just use lemma 1, applied to the particular element : we see that the left side of (**) is contained in
which collapses to , since
. This completes the proof.
Theorem 2: is the least upper bound of
, i.e., if
is a subset containing both
and
, then
.
Proof: We are required to show that
Again, we just apply lemma 1 to the particular element : the left-hand side of the claimed inclusion is contained in
but since is true by hypothesis (is globally true as a predicate on the implicit variable
), this last subset collapses to
which completes the proof.
Theorems 1 and 2 show that for any set , the external poset
admits joins. One may go on to show (just on the basis of the topos axioms) that as in the case of meets, the global external operation of taking joins is natural in
, so that by the Yoneda principle, it is classified by an internal join operation
namely, the map which classifies the union of the subsets
and this operation satisfies all the expected identities. In short, carries an internal Heyting algebra structure, as does
for any set
.
We will come back to this point later, when we show (as a consequence of strong extensionality) that is actually an internal Boolean algebra.
Construction of Coproducts
Next, we construct coproducts just as we do in ordinary set theory: as disjoint unions. Letting be sets (objects in an ETCS category), a disjoint union of
and
is a pair of monos
whose intersection is empty, and whose union or join in is all of
. We will show that disjoint unions exist and are essentially unique, and that they satisfy the universal property for coproducts. We will use the notation
for a disjoint union.
Theorem 3: A disjoint union of and
exists.
Proof: It’s enough to embed disjointly into some set
, since the union of the two monos in
would then be the requisite
. The idea now is that if a disjoint union or coproduct
exists, then there’s a canonical isomorphism
. Since the singleton map
is monic, one thus expects to be able to embed and
disjointly into
. Since we can easily work out how all this goes in ordinary naive set theory, we just write out the formulas and hope it works out in ETCS.
In detail, define to be
where is the singleton mapping and
classifies
; similarly, define
to be
Clearly and
are monic, so to show disjointness we just have to show that their pullback is empty. But their pullback is isomorphic to the cartesian product of the pullbacks of the diagrams
so it would be enough to show that each (or just one) of these two pullbacks is empty, let’s say the first.
Suppose given a map which makes the square
A -------> 1 | | h | | chi_0 V sigma_X V X -------> PX
commute. Using the pullback principle, the map classifies
which is just the empty subset. This must be the same subset as classified by (where
is the diagonal), which by the pullback principle is
An elementary calculation shows this to be the equalizer of the pair of maps
So this equalizer is empty. But notice that
equalizes this pair of maps. Therefore we have a map
. By corollary 2 above, we infer
. This applies to the case where
is the pullback, so the pullback is empty, as was to be shown.
Theorem 4: Any two disjoint unions of are canonically isomorphic.
Proof: Suppose is a disjoint union. Define a map
where classifies the subset
, and
classifies the subset
. Applying the pullback principle, the composite
classifies
which is easily seen to be the diagonal on . Hence
. On the other hand,
classifies the subset
which is empty because and
are disjoint embeddings, so
. Similar calculations yield
Putting all this together, we conclude that and
, where
and
were defined in the proof of theorem 3.
Next, we show that is monic. If not, then by strong extensionality, there exist distinct elements
for which
; therefore,
and
. By the pullback principle, these equations say (respectively)
If , then both
factor through the mono
. However, since
is monic, this would imply that
, contradiction. Therefore
. By similar reasoning,
. Therefore
where is the negation operator. But then
. And since
is the union
by assumption,
must be the top element
, whence
is the bottom element 0. This contradicts the assumption that the topos is nondegenerate. Thus we have shown that
must be monic.
The argument above shows that is an upper bound of
and
in
. It follows that the join
constructed in theorem 3 is contained in
, and hence can be regarded as the join of
and
in
. But
is their join in
by assumption of being a disjoint union, so the containment
must be an equality. The proof is now complete.
Theorem 5: The inclusions ,
exhibit
as the coproduct of
and
.
Proof: Let ,
be given functions. Then we have monos
Now the operation certainly preserves finite meets, and also preserves finite joins because it is left adjoint to
. Therefore this operation preserves disjoint unions; we infer that the monos
exhibit as a disjoint union of
. Composing the monos of (1) and (2), we have disjoint embeddings of
and
in
. Using theorem 4,
is isomorphic to the join of these embeddings; this means we have an inclusion
whose restriction to yields
and whose restriction to
yields
. Hence
extends
and
. It is the unique extension, for if there were two extensions
, then the equalizer of
and
would be an upper bound of
in
, contradicting the fact that
is the least upper bound. This completes the proof.
I think that’s enough for one day. I will continue to explore the categorical structure and logic of ETCS next time.
The solutions to POW-11 are in! Or make that solution, unless you also count your hosts’: only one reader responded. Perhaps everyone has been distracted by the upcoming US presidential election?!
Composite solution by Philipp Lampe, University of Bonn, and Todd and Vishal: We claim there are exactly two functions which satisfy the identity
: the identity function and the function which is identically zero.
Clearly forces
, whence
. Therefore
or
, and the claim above follows if we show that
for all
. The first few cases are built up as follows:
where in a few places we have invoked the simple lemma that [since
]. The remainder of the argument proceeds by strong induction. For odd
, we have
Since preserves sums of two squares, we derive
Therefore, using the inductive hypothesis that for
, we have
From equation (1) we also have
Comparing equations (2) and (3), we infer . In particular,
for
, and we have the case
from before. This will be used to start the induction for even
, starting from
.
For even , where
, we have
and by arguing as we did above, we derive the two equations
where the second follows from the first by inductive hypothesis. Multiplying (4) through by and comparing with the last equation, we conclude
, and this completes the proof.
Remarks:
The underlying idea is to exploit the fact that many numbers can be expressed as a sum of two squares in more than one way, so that one can generate enough relations of the form to nail down all the
[in terms of
]. A powerful way of generating equations of the form
is to consider prime factorizations in the ring of Gaussian integers
. For example, in the Gaussian integers we have the prime factorizations
so that there are two ways of writing their product 65 in the form :
Here we have, respectively, ,
, whence
. Extrapolating from this, if we set
then we have , whence we derive
This gives a large supply of available relations to work with, and from there it is not too bad to get an inductive argument up and flying, once some base cases have been cleared out of the way with some numerical experimentation.
[Although it’s a little unusual to need such an outlay of base cases (up to at least ) before one can start the induction; it creates a psychological hurdle, I feel. It reminds me of POW-6, where no complete solutions were received: one has to fiddle with a number of base cases before the general argument can be pushed through.]
Incidentally, while researching this problem, I found the wonderful website MathPages, created by an all but anonymous author named Kevin Brown. I really recommend checking out this website, and I hope that it is somehow kept alive: it would be a shame for such a treasure trove to be buried and forgotten in some deep hollow of cyberspace.
Time for another problem of the week! This one doesn’t necessarily require any number-theoretic knowledge, although a little bit might help point you in the right direction. As usual, let be the set of natural numbers, i.e., the set of nonnegative integers.
Describe all functions
such that
for all
.
Please submit solutions to topological[dot]musings[At]gmail[dot]com by Sunday, November 2, 11:59 pm (UTC); do not submit solutions in Comments. Everyone with a correct solution will be inducted into our Hall of Fame! We look forward to your response.
A reader brought up essentially this question: does anyone happen to know a proof that does not possess an elementary antiderivative? By “elementary”, I mean a member of the class of functions which contains all constants valued in the complex numbers, the identity function, the exponential and log functions, and closed under the four basic arithmetic operations and composition.
The solutions are in! As is so often the case, solvers came up with a number of different approaches; the one which follows broadly represents the type of solution that came up most often. I’ll mention some others in the remarks below, some of it connected with the lore of Richard Feynman.
Solution by Nilay Vaish: The answer to POW-10 is . Put
and integrate by parts: we have
The first term vanishes by a simple application of L’hôpital’s rule. We now have
where the second equation follows from , and the general elementary fact that
Adding the last two displayed equations, we obtain
This gives
after a simple substitution. The last integral splits up as two integrals:
but these two integrals are equal, using the identity together with the general integration fact cited above. Hence the two sides of (3) equal
recalling equation (1) above. Substituting this for the last integral in equation (2), we arrive at
whence we derive the value of the desired integral .
Remarks:
1. A number of solvers exploited variations on the theme of the solution above, which could be summarized as involving symmetry considerations together with a double angle formula. For example, Philipp Lampe and Paul Shearer in their solutions made essential use of the identity
in conjunction with the complementarity , and the general integration fact cited above.
2. Arin Chaudhuri (and Vishal in private email) pointed out to me that the evaluation of the integral
is actually fairly well-known: it appears for example in Complex Analysis by Ahlfors (3rd edition, p. 160) to illustrate contour integration of a complex analytic function via the calculus of residues, and no doubt occurs in any number of other places.
3. Indeed, Simon Tyler in his solution referred to this as an integral of Clausen type, and gave a clever method for evaluating it: we have
which works out to
The last integral can be expanded as a series
where the summands for odd vanish; the other terms can be collected and then resummed to yield
and by substituting this for the integral in (*), the original integral is easily evaluated.
4. Arin C. later described still another method which he says he got from an exercise in a book by Apostol — it’s close in spirit to the one I myself had in mind, called “differentiation under the integral sign”, famously referred to in Surely You’re Joking, Mr. Feynman!. As Feynman recounts, he never really learned fancy methods like complex contour integration, but he did pick up this method of differentiating under the integral sign from an old calculus book. In typical Feynman style, he writes,
“It turns out that’s not taught very much in the universities; they don’t emphasize it. But I caught on how to use that method, and I used that one damn tool again and again. So because I was self-taught using that book [Wilson’s Advanced Calculus], I had peculiar methods of doing integrals. The result was, when guys at MIT or Princeton had trouble doing a certain integral, it was because they couldn’t do it with the standard methods they had learned in school. If it was contour integration, they would have found it; if it was a simple series expansion, they would have found it. Then I come along and try differentiating under the integral sign, and often it worked. So I got a great reputation for doing integrals, only because my box of tools was different from everybody else’s, and they had tried all their tools on it before giving the problem to me.”
So, what is this method? Wikipedia has a pretty good article on it; the idea is to view a given definite integral as an instance of a smoothly parametrized family
of integrals, where the extra parameter
is inserted at some judicious spot, in such a way that
has an easy-to-integrate derivative. Then one figures out
, and evaluates it at the
which yields the original definite integral.
The best way to understand it is through an example. Pretend you’re Feynman, and someone hands you
[I suppose there may be a contour integration method to handle that one, but never mind — you’re Feynman now, and not supposed to know how to do that!] After fiddling around a little, you find that by inserting a parameter in the exponent,
this has a very simple derivative:
i.e., by differentiating under the integral sign, you manage to kill off that annoying factor in the denominator:
We therefore have
and notice that from the original definition of , we have
. Thus
, and the problem integral evaluates to
. Bam!
It takes a little experience to judge where or how to stick in the extra parameter to make this method work, but the Wikipedia article has some good practice problems, including the integral of POW-10. For this problem they recommend considering
and I’ll let you the reader take it from there.
The point is not that this method is much more elegant than others, but that a little practice with it should be enough to convince one that it’s incredibly powerful, and often succeeds where other methods fail. (Not always, of course, and Feynman had to concede defeat when in response to a challenge, his pal Paul Olum concocted some hellacious integral which he said could only be done via contour integration.)
Doron Zeilberger is also fond of this method (as witnessed by this paper). Something about it makes me wonder whether it’s secretly connected with the idea of “creative telescoping” and the powerful methods (e.g., see here) developed by Gosper, Wilf, Zeilberger, and others to establish identities of hypergeometric type. But I haven’t had time to consider this carefully.
Also solved by Arin Chaudhuri, Philipp Lampe (University of Bonn), Paul Shearer (University of Michigan), and Simon Tyler. Thanks to all who wrote in!
Recent Comments