Article
Author: Christopher J. G. Meacham (University of Massachusetts Amherst)
In recent work, Adlam (2022b), Chen and Goldstein (2022), and Meacham (2023) have defended accounts of laws that take laws to be primitive global constraints. A major advantage of these accounts is that they're able to accommodate the many different kinds of laws that appear in physical theories. In this paper I'll present these three accounts, highlight their distinguishing features, and note some key differences that might lead one to favor one of these accounts over the others. I'll conclude by briefly discussing a version of a “constraint” account that I think is especially attractive.
Keywords:
How to Cite: Meacham, C. J. (2025) “Constraint Accounts of Laws”, Ergo an Open Access Journal of Philosophy. 12(0). doi: https://doi.org/10.3998/ergo.7427
A number of different accounts of laws have been offered in the literature. The most popular accounts include Humean accounts (e.g., Lewis (1994)), Necessitation Relation accounts (e.g., Armstrong (1983)), Dispositional Essentialist accounts (e.g., Bird (2007)), and Primitive Local Dynamics accounts (e.g., Maudlin (2007)). Although these accounts have different pros and cons, they have a shared weakness. All of these accounts have difficulty accommodating the full variety of nomic possibilities and laws that physicists have taken seriously.
Adlam (2022b), Chen and Goldstein (2022), and Meacham (2023) have recently proposed novel accounts of laws which can accommodate these possibilities. Although these accounts differ in detail, all three suggest a similar picture of laws. First, all three accounts posit primitive nomic features of the world. Second, all three accounts take these nomic features to be global—i.e., features of the world as a whole. Third, all three accounts suggest a picture of laws in which they’re best thought of as things that constrain the world, instead of things that (say) produce or generate the world.
This paper aims to examine these primitive global constraint accounts. The goal of the paper is to consolidate a recent trend of research on constraint accounts, and lay out a framework for debate. In §2 I’ll describe the features that these three accounts have in common, and contrast them with some popular accounts of laws. In §3 I’ll provide a more detailed description of these three accounts. In §4 I’ll explore some of the key differences between these accounts, and assess the pros and cons of these differences. In light of this discussion, I’ll briefly present a Primitive Global Constraint account that I think looks particularly appealing in §5. I’ll conclude in §6.
Adlam (2022b), Chen and Goldstein (2022), and Meacham (2023) have recently advanced accounts of laws that resemble each other in a number of respects. I’ll call them:
Primitive Global Constraint Accounts: Primitive Global Constraint accounts of laws maintain that (i) the world has primitive nomic features, (ii) these nomic features are “global”; they work at the level of the world as a whole, and (iii) while these nomic features constrain what the world must be like, they don’t do anything more than that—e.g., they don’t engage in substantive time asymmetric relations, like production or generation.
Thus all Primitive Global Constraint accounts (or “constraint accounts”, for short) take there to be fundamental irreducible nomic features of the world. And they take these nomic features to do something like constrain what the world is like; where “constraining” is like “governing”, but without the temporal connotations.1 Thus these nomic features do something more metaphysically heavyweight than merely describing what the world is like, but not as metaphysically heavyweight as “generating” or “producing” future states of the world from past ones. Though none of the proponents of constraint accounts provide a reductive definition of the notion of constraining, one might try to spell it out in terms of essences or grounding relations. The question of how to best spell it out in such terms would lead us too far afield, however, so I’ll limit my discussion of these options to a footnote.2
Constraint accounts take these nomic features to be “global,” in the following sense. Let one region be a full copy of another if they contain the same number of things, those things have the same fundamental properties and relations, those properties and relations have the same higher-order fundamental properties and relations, and so on.3 Let’s say that an account of laws is local if any regions that are full copies of one another will instantiate the same laws and chances, and global otherwise. Primitive global constraint accounts are global in this sense—intuitively, you can know everything fundamental that there is to know about a region, its qualities, its quality’s qualities, and so on, without knowing what laws and chances are instantiated in that region.
To help us get a clearer picture of this view, let’s look at how constraint accounts differ from some standard accounts of laws.
On Humean accounts, such as Lewis’s (1994), the laws are just helpful descriptions of regularities found in the distribution of local occurrent facts. Humean accounts join constraint accounts in taking the laws to be global features of the world. But they don’t take the laws to be doing something as metaphysically substantive as constraining the world.4 And they deny that the world has primitive nomic features – any nomic features the world has are reducible to facts about the distribution of local occurrent properties.
On Necessitation Relation accounts, such as Armstrong’s (1983), the laws are determined by a fundamental “necessitation relation” that holds between fundamental properties, where if holds at a world, then anything at that world that has property will also have property .5 Like constraint accounts, Necessitation Relation accounts take worlds to have irreducible nomic features (certain instantiations of the necessitation relation). And they take these nomic features to do something like constrain what the world is like. However, unlike constraint accounts, Necessitation Relation accounts are local. For on Necessitation Relation accounts, regions that are full copies of each other will instantiate the same laws and chances, since they’ll instantiate the same fundamental properties and the same necessitation relations between them.
On Dispositional Essentialist accounts, such as Bird’s (2007), the laws are determined by the essential dispositions of fundamental properties. Dispositional Essentialist accounts join constraint accounts in taking there to be irreducible nomic features (the dispositions of fundamental properties). Likewise, they take these nomic features to do something like constrain what the world is like. But unlike constraint accounts, Dispositional Essentialist accounts are local. Since the dispositions of fundamental properties are essential—the same at all worlds – regions that are full copies of each other will instantiate the same laws and chances.
On Primitive Local Dynamics accounts, like Maudlin’s (2007), laws are primitive features of regions that determine how future regions are generated or produced from prior ones. (Thus on these accounts, all fundamental laws are laws of temporal evolution.) Like constraint accounts, Primitive Local Dynamics accounts take worlds to have primitive nomic features. But they differ from constraint accounts in being local; since the laws of temporal evolution that govern a region are features of that region, any full copy of a region will instantiate the same laws and chances.6 Likewise, they differ from constraint accounts in taking the laws to do more than merely constrain what the world is like. For the laws not only constrain the future states of the world, they also produce or generate these states.
In §2 we saw how constraint accounts differ from some standard accounts of laws. Now let’s look at why one might prefer a constraint account over these more established alternatives. The most obvious advantage of constraint accounts, relative to these alternatives, is that they have an easier time accommodating the variety of nomic possibilities and fundamental laws considered by the scientific community. To see this, let’s briefly rehearse some difficulties facing the standard accounts here, and how the constraint accounts do better.7
Let’s start with Humean accounts. Humean accounts can’t recognize laws and nomic possibilities that don’t supervene on the distribution of local occurrent facts. Thus, for example, Humean accounts are unable to distinguish between otherwise identical worlds where there’s a single chancy coin flip, but the chance of heads is 0.5 at one world and 0.6 at the other. Likewise, Humean accounts have difficulty recognizing worlds with uninstantiated laws, such as laws governing charge at a world where no charges are instantiated.8
Necessitation Relation accounts run into difficulties because most of the fundamental laws scientists have entertained are hard to characterize in terms of necessitation relations between fundamental properties. For example, it’s unclear how to characterize laws concerning particular times or places, like laws regarding initial conditions (e.g., the Past Hypothesis), since such laws generally aren’t reducible to claims about entailments between fundamental properties.9 Likewise, it’s unclear how to characterize laws that concern holistic rather than particular facts about the instantiations of properties, like conservation laws (e.g., charge conservation), since global claims about the total amount of charge don’t seem reducible to claims about the co-occurrence of different fundamental properties.10 And it’s unclear how to provide a satisfactory characterization of functional laws (e.g., Coulomb’s law), since it’s difficult to see how to characterize laws relating different magnitudes of properties in terms of claims about the co-occurrence of particular properties.11
Dispositional Essentialist accounts run into problems because many of the laws physicists have considered aren’t easily derivable from or consequences of the dispositions of fundamental properties. In fact, Dispositional Essentialist accounts and Necessitation Relation accounts have difficulties with many of the same kinds of laws. For example, it’s unclear how to use the dispositions of fundamental properties to derive laws concerning initial conditions, since it’s hard to find properties that plausibly (i) are fundamental, (ii) yield dispositions that impose the right constraints on initial conditions, and (iii) apply to the kinds of things which these would need to have those dispositions.12 Likewise, it’s hard to see how to use the dispositions of fundamental properties to derive holistic laws like charge conservation, since constraints on global facts about the total amount of charge don’t seem reducible to facts about the dispositions of particular charged objects.13 And it’s unclear how to use the dispositions of fundamental properties to yield the functional laws posited by physics, since it’s hard to provide a derivation of such laws that (i) is appropriately unified, (ii) assigns plausible dispositions to properties, and (iii) appeals only to the usual kinds of fundamental properties.14
Primitive Local Dynamics accounts run into difficulty because a number of the laws physicists have taken seriously are either hard to understand dynamically, hard to understand locally, or conflict with the idea of temporally asymmetric production. For example, it’s hard for Primitive Local Dynamics accounts to accommodate laws regarding initial conditions, since such laws are not dynamical. Likewise, it’s hard for Primitive Local Dynamics accounts to accommodate laws whose solutions are complete histories (such as Einstein’s Equation), since such laws are neither dynamical nor local. And it’s hard to square Primitive Local Dynamics accounts with retrocausal laws (such as the Wheeler-Feynman absorber theory, or the two-state vector accounts of quantum mechanics), or possibilities involving closed time-like curves (such as Gödel solutions in General Relativity), since such laws seem incompatible with a picture of temporally asymmetric production.15
Primitive Global Constraint accounts, by contrast, don’t face any of these difficulties. Constraint accounts avoid the worries facing Humean accounts because they don’t take laws to supervene on the distribution of local occurrent facts. Thus constraint accounts have no problems recognizing possibilities that are identical with respect to their occurrent facts, but differ with respect to their laws. Constraint accounts also avoid the worries facing Necessitation Relation and Dispositional Essentialist accounts, because they don’t need the laws to be something you can derive from entailments between or dispositions of fundamental properties. Thus they have no problem accommodating, e.g., initial condition laws, conservation laws, or functional laws. And constraint accounts avoid the worries facing Primitive Local Dynamics accounts because they don’t require laws to be dynamical and local, and they don’t presuppose some temporally asymmetric notion of production. Thus they have no problems recognizing laws that aren’t easily formulated in terms of local dynamical laws, or recognizing retrocausal laws or laws which admit the possibility of closed time-like curves.
Adlam (2022b), Chen and Goldstein (2022), and Meacham (2023) have all advocated adopting Primitive Global Constraint accounts of laws. But while these accounts are similar in a number of respects, they’re distinct proposals. Let’s take a more careful look at these views.
On Adlam’s (2022b) proposal, the laws of nature are parts of the objective modal structure of the world. Adlam characterizes the modal structure associated with the laws in terms of constraints induced by these laws. And these constraints metaphysically necessitate that the world be certain ways.
More formally, let a “Humean mosaic” be a distribution of local categorical properties across spacetime.16 Adlam proposes to associate laws with functions that take sets of Humean mosaics as inputs and yield real numbers as outputs, in a manner that satisfies the probability axioms. For non-probabilistic laws, these values are all 0s or 1s, and the function serves to identify the set of Humean mosaics that this law requires the actual world to “belong” to.17 For probabilistic laws, these values can take any value in the interval, and the function serves to determine the likelihood of different sets of Humean mosaics being the ones the actual world has to belong to.18 Note that these laws do not need to be the complete laws of a world—many different laws can obtain at the same world. When there are multiple laws that obtain at a world, the combined effect of the laws is to require that the world belong to the intersection of the set of Humean mosaics picked out (deterministically or probabilistically) by each law.
Two notes. First, Adlam spells out the account in terms of functions over Humean mosaics instead of worlds because worlds contain more information then we want, such as information about what the laws and chances are. And characterizing laws and chances in terms of functions over worlds which are partially individuated by what laws and chances obtain raises worries about circularity/triviality. Second, Adlam characterizes these functions in terms of sets of Humean mosaics instead of individual Humean mosaics because laws generally don’t tell you precisely what Humean mosaic the actual world must belong to. For example, the gravitational force law will pick out a set of Humean mosaics that satisfy that law, but doesn’t say anything further about which of those worlds is the actual one.
Adlam takes one of the advantages of her account to be its neutrality with respect to many of the other questions one might ask about laws and chances. For example, the account leaves open what, if anything, the laws of nature add to our ontology. The account leaves open whether laws of nature are qualitative intrinsic features of the world, and whether they’d be preserved by duplication. The account leaves open how we should think about the modal status of the laws of nature.19 And so on.
So, to sum up, Adlam’s account posits constraints that metaphysically necessitate that the Humean mosaic have certain features; constraints whose content can be characterized using probability functions over sets of Humean mosaics. And Adlam’s account associates/identifies laws with these constraints.
Chen and Goldstein (2022) propose to take the laws of nature to be primitive facts about the world. Their discussion suggests that these laws have something like sentential form— for example, they hold that the laws refer to particular properties.20 As a consequence, they take the laws to be relatively fine-grained. E.g., a law that refers to the properties green and blue can be distinct from an intensionally equivalent law that refers to the properties grue and bleen.21 That said, Chen and Goldstein are explicitly agnostic about how fine-grained the laws are, and don’t provide a precise characterization of the form of the laws or how they’re individuated.22. For concreteness, I’ll assume from here on that they take laws to have a form similar to that of structured propositions, and are individuated in a similar manner.23
On Chen and Goldstein’s account, non-probabilistic laws are taken to pick out a set of models, which the world that instantiates that law must be a member of. Probabilistic or chancy laws are given a more ambiguous treatment; Chen and Goldstein discuss five options for how to incorporate chances into their account. Since this is relevant to the discussion to come, it’s worth briefly reviewing what each of these options are.
The first option they consider is pairing their non-Humean account of laws with a Humean account of chances.24 The second option is to take chances to be distinct primitive facts about the world, facts that aren’t directly related to the primitive constraints on nomic possibility imposed by non-probabilistic laws. The third option is to take chances to be primitive facts about the world, and relate these facts to the constraints on nomic possibility imposed by non-probabilistic laws by adopting a gradable notion of constraining. In particular, they suggest understanding both probabilistic and non-probabilistic laws as imposing gradable constraints on nomic possibility—constraints which can take on values between 0 and 1— and taking the non-probabilistic laws to be the ones that only impose extreme values (0 or 1).
The fourth option is to adopt a typicality account of chances. There’s a literature on how to do this for non-dynamical chances, but one might try to extend this interpretation to dynamical chances by positing primitive facts about which dynamical histories are typical. And one can relate these facts to the constraints on nomic possibility imposed by non-probabilistic laws by taking typicality facts to constrain nomic possibility too – i.e., by requiring nomic possibilities to be typical. Unfortunately, because there are are atypical models of laws, this option entails that there are models of laws—prima facie nomic possibilities—that aren’t possible.
The fifth option is similar to the fourth, positing primitive facts about which histories are typical. But this option takes typicality and nomic possibility to be nested modal notions – every typical world is a nomic possibility, but not all nomic possibilities are typical. And both modal notions intuitively constrain the world—the actual world must be both nomically possible and typical. Unlike the fourth option, this option doesn’t immediately rule out atypical nomic possibilities. But it does rule out atypical actualities, and one might be unhappy with this result. For example, if we follow Lewis (1973) in taking actuality to be an indexical property, then every world will be actual with respect to itself. But if so, then to forbid atypical actual possibilities is to forbid atypical possibilities full stop.
One of the main advantages of constraint accounts is that they can accommodate the nomic possibilities taken seriously by the scientific community. In particular, unlike Humean accounts, they can allow that every model of a set of laws corresponds to a nomic possibility. The first, fourth and fifth approaches considered here all give up this advantage, for they require us to deny that every model of a set of laws corresponds to a nomic possibility.25 Thus if we want to keep this advantage intact, we need to adopt either the second or third approach. So while Chen and Goldstein don’t rule out any of these options, I’ll restrict my attention to the second and third options in what follows.26
Chen and Goldstein’s account is neutral with respect to a number of other questions one might ask about laws and chances, and, like Adlam, they take this to be one of the strengths of their account. For example, the account is agnostic about whether there are objective facts about how to divide the complete laws at a world into individual laws. The account is neutral about what kind of ontological category the laws belong to.27 The account leaves open whether the laws are qualitative intrinsic features of the world, which would be preserved by duplication. And so on.
So, to sum up, Chen and Goldstein’s Account posits primitive facts about the world that impose modal constraints on what can be actual, and it identifies laws with these primitive facts.
Meacham’s (2023) proposal offers an account of laws and chances that mirrors certain accounts of quantitative properties suggested by measurement theory. Meacham posits a primitive nomic likelihood relation that tells us, roughly, when one state of affairs is more nomically likely than another. Meacham then characterizes the laws and chances in terms of these nomic likelihood relations.
More formally, let be an ordered triple consisting of a pair of propositions (understood as sets of possible worlds) and , which aren’t themselves about nomic facts, and a world . Meacham’s account begins by positing a fundamental nomic likelihood relation between such triples that satisfies certain constraints. Intuitively, this relation holds when given at is at least as nomically likely as given at . Meacham then provides a representation and uniqueness theorem showing, roughly, that there’s a unique way of assigning to these triples numbers between 0 and 1, and the status of being nomically required or nomically forbidden, in a manner that lines up with these nomic likelihood relations.28
The complete laws of a world are then identified with the property of being a world that bears the same nomic likelihood relations as . (The account doesn’t take there to be any objective facts about how to divide the complete laws at a world into individual laws.) What’s nomically required/forbidden at is determined by the triples that are assigned the status of being nomically required/forbidden. And the chances of are determined by the numbers assigned to those triples. It’s worth noting that on this account the status of being nomically required and the status of having a chance of 1 are distinct. While every triple that is nomically required has a chance of 1, there can be triples with a chance of 1 that aren’t nomically required. Likewise, the status of being nomically forbidden and the status of having a chance of 0 are distinct. While every triple that is nomically forbidden has a chance of 0, there can be triples with a chance of 0 that aren’t nomically forbidden.
So, to sum up, Meacham’s Account posits a fundamental nomic likelihood relation, and identifies laws with the property of being a world at which certain nomic likelihood relations hold.
Although the three constraint accounts described in §3 are similar in many ways, there are also some important differences. In this section I’ll look at some of the differences between them, and discuss why one might take these differences to tell for or against these accounts.29
Before we proceed, let me introduce some terminology. I’ll say that two laws have the same content if they’re compatible with the same worlds (or in the case of probabilistic laws, they assign the same probabilities to the same sets of worlds). I’ll say that two laws have the same non-nomic content if they’re compatible with the same Humean mosaics (or in the case of probabilistic laws, assign the same probabilities to the same sets of Humean mosaics). And I’ll say that two laws are otherwise identical if they have the same non-nomic content.
Consider the complete laws that obtain at a world . Here are some questions one might ask about . How many different fundamental laws obtain at ? What way of dividing the complete laws at into individual laws cuts nature at the joints? Suppose that at the only laws that obtain are two fundamental non-probabilistic laws whose contents are and , respectively. Is distinct from a world with only one fundamental non-probabilistic law whose content is ?
Adlam’s account starts by positing individual fundamental laws, and takes the complete laws to just be the bundle of the individual laws. Thus on Adlam’s account there are non-trivial answers to each of the above questions: there’s a particular number of fundamental laws that obtain at , a correct way of dividing the complete laws at into individual laws, and a difference between a world in which there are two laws and and a world in which there’s only one law .
By contrast, Meacham’s account starts with an account of the complete laws of a world, and takes the question of how to divide the complete laws into individual laws to be a purely pragmatic question. Thus Meacham’s account provides deflationary answers to the above questions: metaphysically speaking, there is no right number of fundamental laws that obtain at a world, there is no correct way to carve the complete laws into individual laws, and there is no difference between a world with two laws and and a world with one law .
Finally, Chen and Goldstein’s account sits in the middle— it’s agnostic about whether there are substantive answers to these kinds of questions.30
Why might one favor one of these approaches over the other? Those with certain inegalitarian intuitions about laws will be attracted to non-deflationary approaches like Adlam’s. For example, some, such as Lange (2009), have argued that some laws (e.g., conservation laws) are more modally robust than others (e.g., force laws). This position is hard to maintain on deflationary approaches, which don’t take there to be metaphysical facts supporting any one way of dividing the complete laws into individual laws over any other.
By contrast, those with certain skeptical intuitions about the structure of laws will be attracted to deflationary approaches like Meacham’s. For example, consider the classical formulation of Maxwell’s Equations in terms of electric and magnetic fields. This yields four equations, and suggests that there are four fundamental laws governing electromagnetic fields. But if we consider the more modern formulation of Maxwell’s Equations in terms of electric and magnetic potentials, we end up with two equations, which suggests that there are two fundamental laws governing electromagnetic fields. And one might be skeptical that there’s a fact of the matter as to which of these ways of counting laws is correct.
We can use a sentence to express the content of a law. But do the laws themselves have something like sentential form? In particular, are laws like sentences in being more fine-grained than the propositions (sets of worlds) expressing their content? This raises a number of interesting questions regarding how fine-grained laws are. For example, are laws fine-grained enough to distinguish between otherwise identical laws which appeal to different properties? Are laws fine-grained enough to distinguish between otherwise identical laws which have different logical forms?
Chen and Goldstein adopt a fine-grained account according to which the laws have something like sentential form. On their account, there are facts about which properties the laws refer to. So otherwise identical laws which appeal to different properties are distinct. Likewise, on their account there are facts about the logical form of the laws— a law corresponding to a simple characterization of the states the laws allow is distinct from a law corresponding to an infinite disjunction of the individual states the laws allow. So otherwise identical laws which have different logical forms are distinct.
By contrast, both Adlam’s and Meacham’s accounts are relatively coarse-grained. On Adlam’s account the laws correspond to functions that assign values to sets of Humean mosaics. Since such functions aren’t sensitive to the properties used to pick out these sets, or the logical form of the claims used to pick out these sets, otherwise identical laws which appeal to different properties or have different logical forms are the same law. Likewise, on Meacham’s account two worlds can only differ with respect to their laws if they differ with respect to their nomic likelihood relations. And since the relata of nomic likelihood relations are propositions which aren’t themselves about nomic facts, and such propositions aren’t sensitive to properties or logical form, nomic likelihood relations aren’t sensitive to such factors either.
Fine-grained accounts are attractive to those who want to capture hyper-intensional distinctions between laws. For example, Adlam (2022b) has us consider the difference between retrocausal laws that have temporally forwards and backwards dynamics, and retrocausal laws that impose an “all at once” constraint on possible world histories.31 On coarse-grained approaches like Adlam’s and Meacham’s, there is no straightforward way to distinguish between such laws—e.g., nothing about a set of Humean mosaics tells us whether these states have been picked out by laws that impose forwards and backwards dynamics instead of laws that impose an all-at-once constraint. By contrast, on fine-grained approaches like Chen and Goldstein’s, we can distinguish between laws that are formulated in terms of dynamical equations of motion (e.g., Newtonian formulations of classical mechanics) and laws which are not (e.g., Lagrangian formulations of classical mechanics).
Coarse-grained accounts are attractive to those who are skeptical about such fine-grained laws. For example, consider a law that says “All emeralds are green and all sapphires are blue.” And consider a law that says “All emeralds first observed before are grue, and all other emeralds are bleen, and all sapphires first observed before are bleen, and all other sapphires are grue.” If one feels that there isn’t a substantive distinction between these two laws (i.e., that they’re just two different ways of expressing the same law) then one will be unhappy with hyper-intensional accounts that take these to be different laws.
At first glance, it seems like probabilistic and non-probabilistic laws are of a kind. Non-probabilistic laws tell us that if one state obtains, then certain other states are required to obtain. Probabilistic laws tell us that if one state obtains, then certain other states have a certain likelihood of obtaining. Both probabilistic and non-probabilistic laws determine how likely certain states are; non-probabilistic laws are just what you get when you turn that likelihood “all the way up.”
How well do the three constraint accounts do at vindicating this thought? On Adlam’s account, laws and chances are the same kind of thing—probability functions over sets of mosaics. Laws and chances differ only in the values they assign—non-probabilistic laws correspond to the functions that only assign values of 0 or 1. Likewise, on Meacham’s account, both probabilistic and non-probabilistic laws represent the nomic likelihoods of some states given others. Non-probabilistic laws just correspond to the special case where these likelihoods are maximized.
On Chen and Goldstein’s account, matters are more complicated – it depends on how we add chances to their account. As noted earlier (§3.2), I’ll restrict my attention here to options 2 and 3. Option 2 takes probabilistic laws to be primitive facts that are distinct from the non-probabilistic laws—facts that have no direct connection to the primitive constraints on nomic possibility imposed by the non-probabilistic laws. So on this option, probabilistic and non-probabilistic laws are not of a kind. Rather, probabilistic laws are distinct primitive facts that are added independently of the primitive constraints that yield non-probabilistic laws.
By contrast, option 3 adopts a gradable notion of constraining that can take values between 0 and 1, and takes probabilistic and non-probabilistic laws to just impose different grades of nomic constraint. On this approach, non-probabilistic laws are the ones which only impose extreme values of nomic constraint (0 or 1), while probabilistic laws are ones which impose some non-extremal values. So on this approach, probabilistic and non-probabilistic laws are of a kind. They’re both nomic constraints, just ones that take different values.
There’s a difference between possibilities that are nomically required and possibilities that have a chance of 1. For while it’s plausible that every possibility that’s nomically required has a chance of 1, not every possibility that has a chance of 1 is nomically required. For example, the chance of at least one of infinitely many fair coin tosses landing heads is 1, but this possibility is not nomically required—it’s nomically possible that all these coins land tails. Likewise, prima facie there’s a difference between possibilities that are nomically forbidden and possibilities that have a chance of 0—for while every possibility that’s nomically forbidden might have a chance of 0, there are possibilities with a chance of 0 that aren’t nomically forbidden (e.g., every one of infinitely many fair coin tosses landing tails).
How well do the constraint accounts do at capturing this distinction? Adlam’s account doesn’t appear able to distinguish between such possibilities. For on this account, the only thing that distinguishes between probabilistic and non-probabilistic laws is whether they assign non-extremal values. And nothing about this account provides us with the tools to distinguish between those chance 1 possibilities that are nomically required and those that are not. Likewise, nothing about the account allows us to distinguish between those chance 0 possibilities that are nomically forbidden and those that are not.
Meacham’s account, on the other hand, can distinguish between such possibilities. The representation and uniqueness theorem in Meacham (2023), coupled with the proposed account of chances and nomic requirements, entails that while every possibility that’s nomically required has a chance of 1, there can be possibilities with a chance of 1 that aren’t nomically required. Likewise, while everything that’s nomically forbidden has a chance of 0, there can be possibilities with a chance of 0 that aren’t nomically forbidden.
On Chen and Goldstein’s account, it will depend on how we decide to incorporate chances into the account. As before, I’ll restrict my attention to option 2 (adding primitive probabilistic facts that aren’t directly related to the non-probabilistic law facts) and option 3 (adopting a degreed notion of constraining, where non-extremal values yield non-trivial chances, and extremal values (1 and 0) yield nomic requirements and forbiddings). Option 2 has the tools to distinguish between having a chance of 1 and being nomically required, since the facts that yield chances and the facts that yield nomic requirements are distinct posits that aren’t reducible to one another. Likewise, this option has tools to distinguish between having a chance of 0 and being nomically forbidden. Option 3, however, seems unable to distinguish between such possibilities. Since having a chance of 1 and being nomically required are identified with the same primitive fact— a constraint with the maximum value – we can’t distinguish between them. Likewise, we can’t distinguish between having a chance of 0 and being nomically forbidden.
This consideration, together with the consideration discussed in §4.3, suggest that proponents of Chen and Goldstein’s account are going to face a choice about which bullet they want to bite. If they adopt option 2 for incorporating chances into the account, then they’ll be able to distinguish between having a chance of 1/0 and being nomically required/forbidden, but won’t be able to vindicate the thought that chances and nomic requirements/forbiddings are of a kind. Whereas if they adopt option 3 for incorporating chances into the account, then they’ll be able to maintain that chances and nomic requirements/forbiddings are of a kind, but won’t be able to distinguish between having a chance of 1/0 and being nomically required/forbidden.
We typically assign numbers to the different outcomes of chance events. For example, given a fair coin toss, we assign 0.5/0.5 to the coin landing heads/tails. And it seems like these assignments are correct; if we were to assign 0.6/0.4 to a fair coin landing heads/tails, we would be doing something wrong. But what makes these first assignments correct and these alternative assignments wrong? How does the structure of chance events vindicate assigning numbers to outcomes in the way that we do?32
The most straightforward way to answer this question would be to posit fundamental relations between the outcomes of chance events and numbers—e.g., between a fair coin landing heads and the number 0.5. But this answer is implausible. It’s a merely conventional choice to assign numbers between 0 and 1; we could assign numbers between 0 and 2 just as well. So what vindicates these numerical assignments can’t be a relation between chance events and particular numbers. What’s more plausible is that chance events have a kind of quantitative structure which is perspicuously represented by numbers. If that’s right, then a satisfactory account of probabilistic laws and chances should tell us what that underlying quantitative structure is, and how it vindicates our assigning numbers to chances in the way that we do.
How well do the different constraint accounts do at satisfying this desideratum? Meacham’s account satisfies this desideratum. The account’s nomic likelihood relations provide the quantitative structure underlying chances. The account’s representation and uniqueness theorem shows that there is a way of assigning numbers in the interval to triples that lines up with the nomic likelihood relation, that this assignment is unique, and that these assignments will satisfy the probability axioms. And since the account identifies chances with these assignments, it vindicates our assigning numbers to chances in the way that we do.
Adlam’s account does not satisfy this desideratum. Adlam’s account takes laws to correspond to probability functions over sets of Humean mosaics, and probabilistic laws to correspond to the subset of such functions that assign non-extremal values. In some places Adlam’s discussion suggests that we might identify the laws with such functions, in other places Adlam merely suggests that the laws “induce” such functions. Let’s consider each of these possibilities in turn. If we identify the laws with such functions, then it’s a primitive fact that these numbers are assigned to the corresponding sets. As we’ve seen, this kind of view is implausible. Alternatively, we might take the laws to merely induce such functions, and take our primitive metaphysical posits to consist of something like fundamental non-numerical relations that encode the quantitative structure of chances. But then the account needs to tell us what these metaphysical posits are, and how they vindicate our numerical assignments to chances. Adlam’s account, as presented, does neither.
For similar reasons, Chen and Goldstein’s account fails to satisfy this desideratum. If we incorporate chances into the account via option 3—adopting a degreed notion of constraining that takes values between 0 and 1—then their account is in a similar position to the first version of Adlam’s account, appearing to ground the chance facts in something like primitive relations to numbers. If we incorporate chances into the account via option 2—adding probabilistic laws as independent primitive facts—then their account can avoid positing implausible relations to numbers by taking these primitive facts to consist of non-numerical relations that ground the quantitative structure of chances. But then their account is in a similar position to the second version of Adlam’s account, needing to tell us what this metaphysical structure is, and how it vindicates our numerical assignments to chances.
It’s natural to wonder whether one could add something like Meacham’s approach to Adlam and Chen and Goldstein’s accounts in order to satisfy this desideratum. As it turns out, it is difficult for these accounts to incorporate something like the representation theorem framework employed in Meacham (2023) without giving up some of the distinctive ways in which these accounts differ from Meacham’s account: taking there to be objective facts about how to carve up the complete laws, and taking laws to be fine-grained. Let’s see why.
First, consider the issue of whether there’s an objective way to carve the complete laws into individual laws. Representation and uniqueness theorems work holistically. These theorems take all of the instances of a certain kind of relational fact (in this case, instantiations of the nomic likelihood relation) and tell us that they can be uniquely represented in a certain way. But these theorems generally won’t work if we only input part of the relevant relational facts. Likewise, they won’t output parts of a representation—they’ll output a single unified representation. So it’s hard to see how a representation theorem approach could yield the division into individual laws that Adlam’s account requires.
Second, consider the issue of whether there are fine-grained laws. The representation theorem Meacham employs, like the variants one can find in the literature on measurement theory, assigns a representation to an algebra defined over a space of elements. So the relevant representational statuses (having a certain chance, being nomically required/forbidden) are assigned to sets of these elements. The straightforward way to interpret this framework is to think of the elements as possible worlds, and take these sets of them to correspond to (coarse-grained) propositions. But this yields objects of representation that are too coarse-grained to capture the kind of fine-grained distinctions that, e.g., Chen and Goldstein want to capture. To get these theorems to apply to fine-grained propositions, the objects of representation would need to correspond to sets of something more fine-grained than possible worlds: impossible worlds. But this replacement renders the theorem uninteresting, since the axioms of these theorems and the representations they yield will no longer mean what we want them to mean.
To see why, consider the most straightforward way of making this replacement. Take some appropriately expressive (interpreted) language and let each set of sentences in this language correspond to an impossible world; intuitively, this set will correspond to all the sentences that are true at that (impossible) world. So while a sentence expressed in terms of green and blue will be true at the same set of possible worlds as the corresponding sentence expressed in terms of grue and bleen, those two sentences will pick out different sets of impossible worlds—there will be some impossible worlds with but not , and some impossible worlds with but not . There aren’t any formal obstacles to applying the axioms to these new elements (impossible worlds), and obtaining a representation regarding sets of these elements. After all, the only thing that’s changed formally is that we’ve replaced one set of elements with another, and nothing about the theorem cares about what the elements we’re working with are.
But this replacement undercuts the interest of the theorem, because the axioms and results will no longer mean what we want them to mean. Consider, for example, the first axiom of Meacham’s approach. Let stand for the set of impossible worlds containing . The first axiom of Meacham’s approach entails (among other things) that if is assigned a likelihood, then is as well. Meacham motivates this axiom by noting that, intuitively, if there’s some nomic likelihood of it raining, then there should also be some nomic likelihood of it not raining. And if corresponds to a set of possible worlds where is true, then the axiom vindicates this intuition, since the set of worlds at which it’s raining ( is true) is the complement of the set of worlds where it’s not raining ( is true). But if corresponds to the set of impossible worlds where it’s true, this motivation no longer works. Because the set of worlds at which it’s raining ( is true) is not the complement of the set of worlds where it’s not raining ( is true); there are impossible worlds containing both and , and impossible worlds containing neither. Thus the complement of the set of worlds where it’s raining doesn’t correspond to anything in particular, and the axiom’s insistence that this complement be assigned a nomic likelihood is unmotivated.
For another example, consider one of the results of the representation theorem, which says that the relevant sets of elements will be assigned values which satisfy the probability axioms, such as additivity. Additivity tells us that if and are disjoint, then the probability of is equal to the sum of the individual probabilities assigned to and . So if these elements are possible worlds, then additivity entails that the probability of it raining or it snowing is equal to the probability of it raining plus the probability of it snowing. But if and correspond to sets of impossible worlds, this is no longer true. For won’t pick out the set of worlds where is true—there are impossible worlds where or is true, but is not. So additivity won’t entail that that probability of it raining or it snowing is equal to the probability of it raining plus the probability of it snowing.
We typically take there to be constraints that require laws to line up with other laws, chances to line up with other chances, and laws and chances to line up with each other. For example, we think that laws that hold at the same world must be compatible with each other—if one law entails that all masses have charge, there can’t be another law that entails that there are charge-less masses. Likewise, we think that chance distributions that hold at the same world must line up with each other in certain ways—if the chance at of the next coin toss landing heads is 1/2, and the chance of the next two coin tosses landing heads is 1/4, and at the first coin toss does land heads, then the chance at of the second coin toss landing heads must be 1/2. And we think that laws and chances must line up with each other as well— an event with a chance of 1/2 can’t be nomically required.
This suggests another desideratum for a satisfactory account of laws and chances: a satisfactory account should yield the kinds of constraints on laws and chances that we typically take to obtain. Since “the constraints that we typically take to obtain” is somewhat vague, this desideratum is somewhat vague as well. But even without a precise list of constraints in mind, we can prefer theories that provide us with plausible restrictions of this kind over those that do not. So how do the constraints views we’ve looked at do at providing us with such restrictions?
Let’s start with Adlam’s account. On Adlam’s account, both probabilistic and non-probabilistic laws correspond to probability functions that assign values to sets of Humean mosaics. In the case of non-probabilistic laws, Adlam’s account requires the actual world to be a member of the intersection of the Humean mosaics the non-probabilistic laws allow. This requires non-probabilistic laws to be consistent with each other.
Turning to probabilistic laws, Adlam tells us to think of the probabilities assigned to sets of mosaics by probabilistic laws as representing the likelihood that each set of mosaics will be “chosen” to contain the actual world. Thus each set that might be chosen for one probabilistic law must be compatible with every set that might be chosen for every other probabilistic law; i.e., the probabilistic laws that obtain at a world must have outcomes that partition each other. Likewise, every outcome of these probabilistic laws must be compatible with the set of mosaics picked out by every non-probabilistic law that holds at that world. So Adlam’s account also imposes some important consistency constraints on probabilistic laws, and on which probabilistic and non-probabilistic laws can obtain at the same world.
That said, Adlam’s account doesn’t tell us as much about probabilistic laws as one might like. One issue is that the account doesn’t tell us how the probabilities of different probabilistic laws are related to one another. Suppose one probabilistic law assigns a chance of 0.5 to and 0.5 to , while another assigns a chance of 0.3 to and 0.7 to . What is the chance of ? Since Adlam’s account doesn’t tell us how to combine the chances of different laws, it doesn’t provide us with an answer to this question.33
Another issue arises when we try to make precise the requirement that each set that might be “chosen” for one probabilistic law must be compatible with each set that might be “chosen” for every other probabilistic law. What does this requirement amount to? A natural thought is that the each set must be assigned a non-zero probability by their law (so that they might be “chosen”), and must have a non-empty intersection (so they’re compatible). But this thought runs into two problems. First, unless we know how the probabilities of each law are related to one another, the fact that each set is assigned a non-zero probability by their respective laws doesn’t entail that their intersection has a non-zero probability; that would require some further assumption, such as that different laws are probabilistically independent. Second, it’s unable to account for probabilistic laws that assign a chance of 0 to each of a continuum of outcomes (e.g., the chance of a dart landing at any particular point on a dart board). For the above thought would entail that no outcome of such a law could be “chosen.”34
A third issue that arises has to do with how Adlam’s account incorporates probabilistic laws. On Adlam’s account, probabilistic laws correspond to probability functions that assign numbers directly to sets of Humean mosaics. This effectively makes them chance distributions with one argument—that argument being the possibility a chance value is assigned to. But this makes it hard to make sense of the kinds of probabilistic laws posited by our physical theories, which posit chance distributions which take two arguments—the possibility the chance value is assigned to, and the background state that picks out the chance distribution.
For example, consider a toy time-dependent chance theory that assigns chances to coin tosses. Suppose that the chance of a coin toss landing heads at is 0.5, and that after that coin lands heads, the chance of that coin toss toss landing heads (at ) is 1. Since these are two different probability distributions, they need to correspond to different laws on Adlam’s account. But these two probability functions seem to assign values that are incompatible, and so seem to correspond to probabilistic laws that can’t obtain at the same world. In a similar vein, statistical mechanical probabilities vary depending on the macrostate they’re assessed with respect to, and the corresponding probability functions will generally assign incompatible values. But if they assign incompatible values, they can’t correspond to probabilistic laws that obtain at the same world.35
Adlam’s account prides itself in being neutral with respect to a number of issues. In some ways this lack of detailed commitments is a strength—it broadens the range of potential buyers. But it’s also a weakness—it leaves the account unable to provide us with the constraints on laws and chances that we typically take to obtain.
Turning to Chen and Goldstein’s account, they require the world that instantiates a (non-probabilistic) law to belong to the set of models generated by that law. This imposes a consistency constraint on the non-probabilistic laws instantiated at a world. But because their account is agnostic about so many other details, it says little else about what kinds of laws and chances there can be. For example, we haven’t been told anything that ensures that different probabilistic laws will make consistent assignments. We haven’t been told how to combine the chance assignments from different probabilistic laws. We haven’t been told anything that requires different chance distributions at a world to line up with each other. And so on. So, like Adlam’s account, Chen and Goldstein’s account does little by way of imposing constraints on laws and chances.
What about Meacham’s account? Because Meacham’s account commits itself to more details than Adlam’s or Chen and Goldstein’s accounts, it’s better placed to provide us with the kinds of constraints on laws and chances that we’d like an account to provide. In particular, from the constraints the account imposes on the nomic likelihood relation, one can derive a number of the constraints on laws and chances that we typically take to obtain. For example, one can derive that if is nomically required then it will have a chance of 1, and if it’s nomically forbidden it will have a chance of 0. One can derive that if is nomically required if some state obtains, and does obtain, then will be true. One can derive that, given certain assumptions, different chance distributions at a world will be related by conditionalization. And so on.36
I think an adequate account of laws and chances should be able to accommodate the kinds of laws and possibilities that physicists have taken seriously. And since all three constraint accounts can do so, and their mainstream competitors cannot, I take these three constraint accounts to be among the most attractive options on the market.
Of these constraint accounts, I (unsurprisingly) find the account in Meacham (2023) to be the most appealing. This is the only one of the three accounts that vindicates our assigning numbers to chances in the way that we do, yields the kinds of constraints on laws and chances that we typically take to obtain, and can both treat laws and chances as a kind and distinguish between nomic requirements/forbiddings and having a chance of 1/0. Of course, this account also holds that there’s no objective way to carve the complete laws into individual laws, and holds that laws are relatively coarse-grained. Whether these features of the account are pros or cons will depend on one’s philosophical sensibilities, but I’m inclined to think these are plausible verdicts.
That said, there are a couple of worries one might raise for the account described in Meacham (2023). First, the account entails that the laws aren’t intrinsic features of the worlds that instantiate them. This is because the laws of a world are determined by the nomic likelihood relations involving that world, and these relations hold between the world and things external to the world, such as other worlds. In a similar vein, this account entails that a perfect duplicate of a world needn’t have the same laws, and that the laws aren’t qualitative features of the world.37 These are implausible consequences.38 Second, by postulating a fundamental relation that holds between worlds and propositions, it adopts an ontological commitment to such things. It would be preferable to have an account of laws without such commitments.
In light of this, it’s worth considering a variant of the account described in Meacham (2023).39 The account in Meacham (2023) focuses on triples consisting of a world and a pair of propositions. It posits a fundamental nomic likelihood relation that holds between such triples (intuitively, the relation that holds when the likelihood of given at world is at least as great as the likelihood of given at world ). And it identifies laws with the property of being a world bearing certain nomic likelihood relations.
This variant starts by positing a number of mutually exclusive fundamental properties of worlds, . It replaces the triples with triples of properties , where A is a property that holds of a world iff that world makes true, C is a property that holds of a world iff that world makes true, and is one of the fundamental properties mentioned above. It posits a fundamental nomic likelihood relation that holds between such triples. And it identifies the different complete laws with the different properties.40 (So on this view, intuitively, the nomic likelihood relation holds when the likelihood of A given C given laws is at least as great as the likelihood of given given laws .)
This variant of the account in Meacham (2023) inherits the advantages of that account.41 But it also allows us to maintain that laws are intrinsic and qualitative features of worlds that are preserved by duplication. And it only commits us to a single ontological category— properties.
The main demerit of this variant is that it needs to posit a collection of fundamental properties. I suspect people will differ on whether they take this variant’s benefits to be worth the cost. But I’m inclined to think they are.
The most popular accounts of laws in the literature share a common problem—they’re unable to accommodate many of laws and possibilities seriously considered by the scientific community. Recently, a number of new accounts of laws— what I’ve called Global Primitive Constraint accounts – have been proposed that don’t have this problem. I’ve presented a version of a constraint account that I think is particularly attractive. Others may find some other constraint account more to their liking. Either way, these constraint accounts are a welcome addition to the literature on laws.
I’d like to thank Emily Adlam, Eddy Chen, Maya Eddon, Mark Steen, three anonymous referees, attendees of the 2022 UMass Brown Bag, and attendees of the 2023 Society for the Metaphysics of Science conference, for helpful comments and discussion.
Adlam, E. (2022a). Determinism Beyond Time Evolution. European Journal for Philosophy of Science, 12(4), 1–36.
Adlam, E. (2022b). Laws of Nature as Constraints. Foundations of Physics, 52(1), 1–41.
Aharonov, Y., P. G. Bergmann, and J. L. Lebowitz (1964). Time Symmetry in the Quantum Process of Measurement. Phys. Rev., 134, B1410–B1416.
Albert, D. Z. (2000). Time and Chance. Harvard University Press.
Armstrong, D. (1983). What is a Law of Nature? Cambridge University Press.
Barrett, J. A. and E. K. Chen (ms). Algorithmic Randomness and Probabilistic Laws. arXiv preprint:2303.01411.
Bigelow, J., B. Ellis, and C. Lierse (1992). The World as One of a Kind: Natural Necessity and Laws of Nature. British Journal for the Philosophy of Science, 43(3), 371–388.
Bird, A. (2007). Nature’s Metaphysics: Laws and Properties. Oxford University Press.
Chen, E. K. and S. Goldstein (2022). Governing Without a Fundamental Direction of Time: Minimal Primitivism About Laws Of Nature. In Y. Ben-Menahem (Ed.), Rethinking the Concept of Law of Nature (21–64). Springer.
Eddon, M. (2007). Armstrong on Quantities and Resemblance. Philosophical Studies, 136(3), 385–404.
Eddon, M. (2013). Quantitative Properties. Philosophy Compass, 8(7), 633–645.
Eddon, M. (2017). Parthood and Naturalness. Philosophical Studies, 174(12), 3163–3180.
Emery, N. (2019). Laws and Their Instances. Philosophical Studies, 176(6), 1535–1561.
Field, H. H. (1980). Science Without Numbers: A Defence of Nominalism. Princeton University Press.
Forge, J. (1999). Laws of Nature as Relations between Quantities? In H. Sankey (Ed.), Causation and Laws of Nature (111–124). Kluwer Academic Publishers.
Gödel, K. (1949). An Example of a New Type of Cosmological Solutions of Einstein’s Field Equations of Gravitation. Rev. Mod. Phys., 21, 447–450.
Goodman, N. (1955). Fact, Fiction, and Forecast. Philosophy, 31(118), 268–269.
Hoefer, C. (2018). Chance in the World: A Humean Guide to Objective Chance. OUP USA.
Ioannidis, S., V. Livanios, and S. Psillos (2021). No Laws and (Thin) Powers In, No (Governing) Laws Out. European Journal for Philosophy of Science, 11(1), 1–26.
King, J. C. (2019). Structured Propositions. In E. N. Zalta (Ed.), Metaphysics Research Lab. Stanford University, Summer 2019 edition.
Lange, M. (2009). Laws and Lawmakers Science, Metaphysics, and the Laws of Nature. Oxford University Press.
Lewis, D. (1983). New Work for a Theory of Universals. In D. H. Mellor and A. Oliver (Eds.), Properties. Oxford University Press.
Lewis, D. (1994). Humean Supervenience Debugged. Mind, 103(412), 473–490.
Lewis, D. K. (1973). Counterfactuals. Blackwell.
Maudlin, T. (2007). The Metaphysics Within Physics. Oxford University Press.
Meacham, C. J. G. (2005). Three Proposals Regarding a Theory of Chance. Philosophical Perspectives, 19(1), 281–307.
Meacham, C. J. G. (2010). Contemporary Approaches to Statistical Mechanical Probabilities: A Critical Commentary - Part i: The Indifference Approach. Philosophy Compass, 5(12), 1116–1126.
Meacham, C. J. G. (2023). The Nomic Likelihood Account of Laws. Ergo: An Open Access Journal of Philosophy, 9(9), 230–284.
Ott, W. (2022). The Metaphysics of Laws of Nature: The Rules of the Game. Oxford University Press, Oxford, UK.
Tooley, M. (1977). The Nature of Laws. Canadian Journal of Philosophy, 7(4), 667–98.
Tooley, M. (1987). Causation: A Realist Approach. Oxford University Press.
Van Fraassen, B. C. (1989). Laws and Symmetry. Oxford University Press.
Vetter, B. (2012). Dispositional Essentialism and the Laws of Nature. In A. Bird, B. Ellis, and H. Sankey (Eds.), Properties, Powers, and Structures: Issues in the Metaphysics of Realism. Routledge.
Wheeler, J. and R. Feynman (1945). Absorber Theory and the Radiation Arrow of Time. Review of Modern Physics, 17, 157–181.
Wilsch, T. (2021). The Governance of Laws of Nature: Guidance and Production. Philosophical Studies, 178(3), 909–933.