This site will be migrated to mhabic.github.io soon. Please plan accordingly.
Last week at the CUNY set theory seminar I presented a proof of an old theorem of Bukovský, which characterises those pairs of models which are set-generic extensions. It turns out that these are precisely those pairs which satisfy a familiar covering property. This result, which seems to have been forgotten for a while, has gathered attention recently, particularly (as far as many of the people at CUNY are concerned) due to its use by Usuba in his work on set-theoretic geology and proving the downward directed grounds hypothesis. In my presentation I didn’t quite get to lay out everything as neatly as I would have liked, so I am writing this post in the hope of giving a fuller account.
As mentioned, the key property that Bukovský isolated is a kind of covering property between a pair of models. We may focus on the case of the relationship between the universe and an inner model ; for us an inner model will be a transitive proper class model of ZFC. We work throughout in GBC (we can even dispense with class choice), so, in particular, there is no need for inner models to be definable.
Definition 1 Let be an inner model and let be regular in . We say that uniformly -covers if for any function with an ordinal and there is another function in such that and for all .
That is to say, covers the function , giving fewer than many guesses at each coordinate. This property is well-known to all students of forcing and all the introductory texts that I know present the following key fact.
Proof: Fix a function in and a name for it. There is some so that it is forced that maps into . We can thus find, for each , a maximal antichain deciding the values of . Now we just let be the set of the possible values given by the conditions in .
Bukovský’s remarkable result is that the uniform covering property precisely characterises the set-forcing extensions.
Theorem 3 (Bukovský, 1973) Let be an inner model and let be regular in . The following are equivalent:
- is a -cc generic extension of ; that is, there are a poset and a filter such that is -cc in and is generic over and .
- uniformly -covers .
We already saw the forward implication as theorem 2. The majority of the work, not surprisingly, goes into proving the converse. Starting from the hypothesis that uniformly -covers , the argument can be divided into three fairly self-contained steps:
- If is a set of ordinals (or even ) then is a -cc generic extension of . Here is the least transitive model of ZFC extending and containing .
- There is a set of ordinals such that is a terminal -cc generic extension of ; that is, there is no inner model such that and is a -cc generic extension.
- If is the terminal extension from the previous step, then .
As it turns out, step 1 is the hardest to prove. Therefore we shall make the steps in reverse, starting from 3 and working our way back to 1.
Step 3 follows quite easily from the other two. Specifically, assume that . Since all of our models satisfy choice, there must be a set of ordinals . If we knew that was -cc generic over , this would contradict our assumption on the maximality of .
Proof: Fix a function in . There is an ordinal such that the range of is contained in . Let us pick an injection in . Since uniformly -covers we can find a covering function for in . But then the function is a covering function for in .
Returning to the argument from before, we can now apply step 1 to the pair to conclude that really is -cc generic over , giving us the contradiction and completing step 3.
The key realisation for step 2 is the following lemma:
Proof: It shall suffice to prove that is not -distributive. This is because, if adds a function , then the range of can be covered by a set of size in , by theorem 2. Modulo some coding, the function is then determined by a subset of this covering set.
So now assume toward a contradiction that is -distributive. This means that any family of at most many maximal antichains in has a common refinement. This allows us to build a tree in as follows:
- the root of is the top condition of ;
- every node in has at least two immediate successors;
- every level of is a maximal antichain in .
We can build inductively; the successor steps are easy and we use distributivity to pass through limit steps. Specifically, -distributivity allows us to build the tree up to height (at least) . But now take any condition from the -th level of and consider the branch it determines through . Since every node on the branch is splitting, any one-off-the-branch antichain determined by this branch will have size . But this of course contradicts the -cc.
Seen differently, the proof basically shows that a -cc poset cannot be -distributive.
Let be a set of ordinals coding , so that . We claim that is the required terminal -cc extension of . The set is definitely -cc generic over by step 1. On the other hand, if were not terminal, there would be a further -cc generic extension and, by lemma 5, there must be a new subset of in . But contains all the subsets of in by construction. This completes step 2.
In order to deal with step 1 with some elegance we need to introduce some terminology.
Definition 6 Let be cardinals. We shall denote by the free -complete Boolean algebra on the generators for .
By -complete we mean that suprema of sets of size less than exist. Freeness can be interpreted in two ways. We can think of it in terms of a universal property: any map of the generators into a -complete Boolean algebra extends to a -complete homomorphism on the whole . Alternatively, we can think of as being built in stages, starting from the generators and at each stage taking complements and size suprema of what we had constructed before, and modding out by the minimal obvious relations. This latter viewpoint suggest yet another one. We can also view the generators for as representing the propositional formulas , where is a predicate, and being the Lindenbaum algebra of infinitary formulas built from these atomic formulas via negations and size disjunctions. In my talk I suggested that could, for intuitive purposes, be replaced by , but this is misleading, as I will discuss after theorem 7.
Note that, if , then embeds as a -complete subalgebra into . We will henceforth see these algebras as nested. There is also another kind of nesting.
Theorem 7 Let be an inner model. Then embeds into as an –-complete subalgebra; that is, there is a homomorphism such that for any set in of size we have . In fact, can be taken to extend the identity map on the generators.
In particular, the theorem implies that maps small maximal antichains of into maximal antichains in . I will not give a proof of this result, but I want to point out a subtlety in the statement, which caused a lot of issues in my talk. It may seem that one can give a simple counterexample to the theorem. The one brought up in my talk was as follows: supposing that and were sufficiently large, let , for some real , be the conjunction of the generators or their complements according to . Then it should be the case that in . But if has more reals than , then in , contradicting the theorem. The subtlety arises in the claim that in . This would be the case if were -distributive, since then , but in general this will fail. In fact, by freeness, there is a which is compatible with every and their complements but incompatible with every .
We now know that is essentially a –-complete subalgebra of . This will allow us to pull down information from the algebra in to the algebra in . The foremost is the following lemma.
Proof: Given an ultrafilter , we can simply assign to it the set . Conversely, given a set in there is, by freeness, a unique -complete ultrafilter on which contains or omits the generators according to . But then is an –-complete ultrafilter on , since the latter is an –-complete subalgebra of . It is easy to check that these two assignments are inverse to each other.
Given we will denote by the corresponding ultrafilter on . Clearly , but there is no reason to believe that is in any way generic over . In fact, will not be -cc, so we have not quite finished with step 1 of theorem 2. We will use the uniform covering property to thin out to a -cc poset which the residue of will be generic for.
Let be the function that, given a maximal antichain in , gives the element of the intersection , if such an element exists, and gives some arbitrary element of if not. This function will, in general, only exist in . But since uniformly -covers , we can find a covering function for in ; we may also assume that for all . Essentially, the function represents trying to guess where the potential generic filter coding will meet the antichains of . We will use this guess to remove the inessential parts of the antichain .
We work for a while in . Let be a sufficiently large cardinal (much larger than ). Define a subset of by
It should be noted that, while in , it need not be the case that in .
We wish to see that in . This will follow if we can show that for all , by the –-completeness of . So assume that . By the completeness of , we can find a . But note that also , so . Therefore .
We are ready to define the poset which will add the set over the model . Let
In the Lindenbaum algebra view of , the poset is essentially the Lindenbaum algebra where we add the infinitary inference rules coming from ; that is, from we are allowed to infer .
Proof: Work in and let be a maximal antichain. Then is a subset of of size . We shall show that any element of is compatible with some element of . It will follow that has size .
So pick and fix some -complete ultrafilter on with . Then also , so, in particular, . Since we also have and, by -completeness, there is some . But then , meaning that is compatible with in .
Lemma 10 The filter is -generic over .
Proof: Let be a maximal antichain in . Then , so the join exists in and agrees with the join in . It follows that and, since was maximal, must be the top condition of . Then , so we can use the -completeness of to find a .
Since and are interdefinable over (we have just constructed from and and it is easy to read off from ), this ultimately shows that is a -cc generic extension of , finishing the proof of step 1.
I firmly believe that when one is stuck on a research problem one should tell as many people as possible about it, because one of two things will happen: either someone will solve your problem (and you will have contributed to the store of mathematical knowledge) or you will frustrate all of your mathematician friends. Either of those is a good thing. For this reason I’ve decided to write a couple of posts sketching my current project and some of the points of frustration.
The fundamental idea arose from the following fact about Laver functions on a supercompact cardinal, which was given to me as an exercise at some point by Joel Hamkins:
Theorem. If is supercompact then there are (the maximal possible number) many Laver functions such that this sequence is jointly Laver, i.e. such that for any and any sequence of sets in there is a -supercompactness embedding with critical point which satisfies for all .
It is not difficult to see that this is true; furthermore, it will be interesting to give an alternative proof of the weaker claim that there is a jointly Laver sequence of length . This can be accomplished by simply coding everything appropriately. Specifically, start with a single Laver function and let . Given a sequence in , we fix a -supercompactness embedding such that . It is then easy to check that for all .
For the length case, we reindex our sequences to use elements of instead of . Still working with a given Laver function we define for . Given a sequence in we fix a -supercompactness embedding such that and check that this makes everything work as required. If we had we should, at the end, also factor our a supercompactness embedding of the appropriate degree.
There are two things we should take away from this proof:
- The short case was fairly easy, requiring merely some coding. This suggests that whenever we have any Laver function-like object on a cardinal we should have many joint such objects, whatever that might mean;
- In the long case we seemingly only used the -supercompactness of . If we then consider only partially supercompact cardinals, this raises the question whether there is any strength in having a length jointly Laver sequence or whether such things just always exist (provided there is an appropriate Laver function in the first place).
As alluded to in point 1, questions about jointly Laver sequences make sense whenever a Laver function-like object makes sense. This ties together nicely with the various Laver diamond principles, introduced for many large cardinals in an (as of yet unpublished) paper by Hamkins. Building on his work and also on the work of Apter-Cummings-Hamkins on the number of measures problem, some answers have been forthcoming.
Let me illustrate the main results about these joint Laver sequences in the case of measurable cardinals, where many of the interesting phenomena already occur. The supercompact case is fairly similar, with some complications.
To be concrete, if is measurable we call a function a Laver diamond (for measurability) if for every there is an elementary embedding with critical point such that . We call a sequence a joint Laver diamond sequence (for measurability) if for every sequence of sets in there is an elementary embedding with critical point such that for every .
Theorem. If is measurable and has a Laver diamond then it has a joint Laver diamond sequence of length . In general, if is measurable then there is a forcing extensions in which remains measurable and has a joint Laver diamond sequence of length .
This is quite simple. If there is a Laver diamond for then we can simply do the coding we did before and get a joint Laver diamond sequence. The point is that if there is no Laver diamond for we can always force to add one. This can be done in one of several (nonequivalent) ways, e.g. by Woodin’s fast function forcing or by first doing a preparatory forcing and then adding a Cohen subset to .
Theorem. If is measurable then there is a forcing extension in which remains measurable and has a joint Laver diamond sequence of length .
This builds on the construction of adding a single Laver diamond. We first force the GCH to hold at if necessary. We next prepare by doing a Silver-style iteration up to where we add many Cohen subsets to inaccessible . Finally we add many Cohen subsets to . An argument as in the previous case shows that the Cohen subsets of can be decoded into a joint Laver diamond sequence, and since GCH still holds at at the end, there are many. The crucial issue is showing that remains measurable after this forcing. The usual lifting argument via master conditions doesn’t work since the generic is too big to be distilled down to a master condition. To solve this we use what has been called in the literature the “master filter argument”, where instead of building a single master condition we build a descending sequence of partial master conditions, which encode larger and larger pieces of the generic. The construction is quite sensitive and exploits, among other things, the continuity of the embedding at (this becomes relevant in the supercompactness argument).
The fact that in the resulting model GCH holds in is unavoidable without stronger hypotheses. The following question is still open.
Question. Given a model where is measurable, GCH fails at and has a Laver diamond, is there a forcing extension preserving these facts where has a joint Laver diamond sequence of length ?
The final result on measurables is a separation of the conclusions of the previous two theorems. Therefore, while having a joint Laver diamond sequence of length is no weaker in consistency strength than having a joint Laver diamond sequence of length , the outright implication still fails.
Theorem. If is measurable then there is a forcing extension in which remains measurable and has a joint Laver diamond sequence of length but no joint Laver diamond sequence of length .
The key observation here is that, in order to have a joint Laver diamond sequence of length , there must be at least many normal measures on , since every binary -sequence must be guessed by some embedding and, of course, each embedding corresponds to a single sequence. The argument now proceeds by first forcing to add a Laver diamond to as before and then using a result of Apter-Cummings-Hamkins by which we can force over a model with a measurable preserving measurability but making only carry many normal measures. By our argument before cannot possibly have a joint Laver diamond sequence of length greater than . It then remains to check that the single Laver diamond survived this final forcing and this gives us a joint Laver diamond sequence of length as in our first theorem above.
The main gap in these results concerns the lack of control over . One would like to be able to push high and still talk about joint Laver diamond sequences of intermediate length. Of course, this requires higher consistency strength than merely measurability, but I would guess that we get equiconsistency at that level again.
Next time (whenever that might be) I will discuss similar results on (partially) supercompact cardinals and perhaps some others (like weakly compact or strong or strongly unfoldable).
This is a short summary of some recent work on a principle I call the grounded Martin’s axiom. I gave a talk on this material in the CUNY Set Theory seminar a few days ago and a preprint will be available in the near future.
The grounded Martin’s axiom (or grMA) states that the universe is a ccc forcing extension of some ground model and that for any poset which is ccc in and any collection of less than continuum many dense subsets of there is a -generic filter on .
This concept appears naturally when one analyses the Solovay-Tennenbaum proof of the consistency of MA (with the continuum being a regular cardinal ). There we iterate, in many steps, through all the available ccc posets of size and use a suitable bookkeeping device to make sure that we have taken care of not only the posets in the ground model but also the posets that arise in all of the many intermediate models as well. This bookkeeping device (basically a bijection between and ) will necessarily be wildly discontinuous and, in my opinion, distracts from the essence of the argument. Thus I have in the past suggested a reorganization of the proof which eliminates the need for (at least this part of the) bookkeeping by making the iteration slightly longer. Specifically, we construct a finite support iteration of length as follows: starting in a suitable model (satisfying GCH or at least ) we iterate the many small ccc posets from this model, taking care to only take posets which remain ccc in the extension obtained so far; after the first many steps we repeat the process, considering now the small ccc posets in this extension. And we do it again and again, many times. The usual arguments show that what we get in the end is a model of MA and the continuum has size .
However, a new question now arises. Did we need to repeat this process many times? Did we need to repeat it at all? Might we already have MA after the first steps of the new iteration? The answer is no (assuming ). To see why, notice that the forcing up to that point is an iteration of ground model posets, so it is basically a product. Since the forcing to add a single Cohen real will have inevitably appeared as a factor somewhere in this product, the model obtained is a Cohen extension of some intermediate model, but it is well known that MA fails in any Cohen extension where CH fails.
So MA fails in this model, but on the other hand, it looks perfectly crafted to satisfy grMA. Well, almost. What we have ensured by construction is that the restriction of grMA to posets of size less than holds. The same issue arises in the usual MA argument and an easy Löwenheim-Skolem argument shows that there the two versions are equivalent. We cannot simply transpose the argument to the present context since the appropriate elementary substructure of the poset is now in the wrong model, but fortunately a modification of the argument gives the analogous result for grMA.
Having now what might be called a canonical model of grMA, we can also determine some cardinal characteristics in this model. Since grMA clearly implies MA(Cohen) we must have , but since, as before, the model is obtained by adding many Cohen reals to an intermediate extension, we can also conclude that in this model. These two equalities now resolve the whole of Cichoń’s diagram and also show that grMA is less rigid than MA with respect to some of the smaller cardinal characteristics.
Another noteworthy observation is that, while MA implies that the continuum is regular, grMA is consistent with a singular continuum. In particular, it is possible in a model of grMA to have , violating the generalized Luzin hypothesis. An interesting open question here is whether grMA implies that . While this equality holds in the canonical model, I do not know whether it holds in general.
The remainder of the current results on grMA concern its robustness under forcing. It is known that MA is destroyed by very mild forcing, adding either a Cohen or a random real (assuming CH fails). At the same time some fragments of MA are known to be preserved by such forcing. To determine the behaviour of grMA under such forcing, a variation of termspace forcing was utilized.
Termspace forcing (due to Laver and possibly independently Woodin and other people) is a construction for taking a two step forcing iteration and trying to approximate the poset named by by a poset in the ground model. This gives the poset , consisting of -names which are forced by every condition to be in and where extends if this is forced by every condition. It can then be proved that forcing with adds a sort of doubly generic object for . More precisely, forcing with gives a name which, when interpreted by any -generic for , names a -generic for . In particular, the iteration embeds into the product .
The crucial issue, however, is that might not have any nontrivial chain conditions. This is clearly problematic for us, since we are dealing with an axiom that concerns only ccc posets. To fix this flaw we need to restrict the names we consider in the termspace forcing and for this purpose the notion of finite mixtures is introduced. A finite mixture is a -name for an element of the ground model which is decided by some finite maximal antichain (the term finite mixture suggests that these names are obtained by applying the mixing lemma to finitely many check names). The subposet of , consisting only of finite mixtures, has a much better chance of having a good chain condition. In particular, it can be seen that if is just the forcing to add a single Cohen real, then is Knaster if is (here is assumed to be in the ground model). This is the key step in showing that grMA is preserved by adding a single Cohen real (in fact it is preserved with respect to the same ground model). By slightly modifying the notion of a finite mixture to exploit the measure theory involved, a similar approach also shows that grMA is preserved by adding a random real (again, even with respect to the same ground model).
The question still remains whether grMA is preserved when adding more generic reals. For example, what happens if we add many Cohen reals? The methods used for a single real hinge on certain antichain refinement properties of the Cohen poset which are no longer there when adding more reals. Similar question can also be asked for random reals. In that case, at least, we do have an upper bound for preservation, as it is known that adding more than many random reals will destroy MA(Cohen) and thus also grMA, but nothing is known about adding a smaller number.
I want to use this post to once and for all clear up any confusion I might have about what could be called the guessing principles of set theory. A secondary goal is to finally be able to claim that I have used the unique and amusing notation related to these principles.
I will restrict my attention to ; all of the principles generalize to larger cardinals (and restrict to stationary subsets etc.), but this basic case should be enough for an illustration.
So let’s start. I should first explain what I meant by “guessing principles”. Another fitting name would be “anticipatory principles”. Imagine a situation where one is performing an inductive construction of length , with the goal being that the final object will satisfy some universal property. A reasonable strategy is then to diagonalize against all possible counterexamples while performing the construction. However, it might happen that there are simply too many possible counterexamples to deal with in many steps. Nevertheless, it is often the case that we do not in fact need the entire putative counterexample to prevent it becoming a true counterexample, but only some small fragment of it. If we can additionally ensure that these are never resurrected as possible counterexamples, our plan will go through. The only missing part is a coherent way of producing the fragments of possible counterexamples and here is where guessing principles come in.
The simplest guessing principle is (called the stick principle). A -sequence is a sequence of infinite subsets of such that every uncountable contains some . We say that holds if there is a -sequence.
It is easily seen that CH implies . Indeed, CH implies that there are only many countable subsets of , so all of these can be taken for our -sequence. On the other hand, implies (the hypothesis here can be considerably weakened). To see this, let be a sequence of infinite subsets of and consider the poset , seen as adding a subset of . For any given it is dense in this poset that is not contained in the generic object and thus, by , the sequence of the fails to be a -sequence.
A stronger guessing principle is (called the club principle). A -sequence is a sequence of sets , indexed by limit ordinals , such that is a cofinal subset of and each uncountable subset of contains some . We say that holds if there is a -sequence.
Of course, implies . On the other hand, since both and are consistent (as shown by Jensen and Shelah, respectively), is in fact strictly weaker than .
We can, without too much effort, extract an apparently stronger formulation of . The claim is that a -sequence actually gets into any uncountable subset of stationarily often. To see this, let be uncountable and be club. By thinning out if necessary, we can assume that any limit point of is also a limit point of , and is thus in . Since the form a -sequence, we have for some . But then is a limit point of and is in .
The third and most renowned guessing principle is (the diamond principle). A -sequence is a sequence of sets such that and for each we have for stationarily many . We say that holds if there is a -sequence.
It is not difficult to see that implies . In fact, these two principles are equivalent in the presence of CH. To see this, we let be an enumeration of all countable subsets of with cofinal repetition and define , where is an element of the -sequence. One can then show fairly easily that the form a -sequence.
There is a multitude of variations of , where one is either allowed countably many guesses at each stage, or one attempts to guess club often or even guess club often while simultaneously guessing the club itself, but I think this short description will suffice for now.