In model theory, a branch of mathematical logic, a complete theory T is said to satisfy NIP ("not the independence property") if none of its formulae satisfy the independence property—that is, if none of its formulae can pick out any given subset of an arbitrarily large finite set.
Definition
Let T be a complete L-theory. An L-formula φ(x,y) is said to have the independence property (with respect to x, y) if in every model M of T there is, for each n = {0,1,...,n − 1} < ω, a family of tuples b0,...,bn−1 such that for each of the 2 subsets X of n there is a tuple a in M for which
The theory T is said to have the independence property if some formula has the independence property. If no L-formula has the independence property then T is called dependent, or said to satisfy NIP. An L-structure is said to have the independence property (respectively, NIP) if its theory has the independence property (respectively, NIP). The terminology comes from the notion of independence in the sense of boolean algebras.
In the nomenclature of Vapnik–Chervonenkis theory, we may say that a collection S of subsets of X shatters a set B ⊆ X if every subset of B is of the form B ∩ S for some S ∈ S. Then T has the independence property if in some model M of T there is a definable family (Sa | a∈M) ⊆ M that shatters arbitrarily large finite subsets of M. In other words, (Sa | a∈M) has infinite Vapnik–Chervonenkis dimension.
Examples
Any complete theory T that has the independence property is unstable.
In arithmetic, i.e. the structure (N,+,·), the formula "y divides x" has the independence property. This formula is just
So, for any finite n we take the n 1-tuples bi to be the first n prime numbers, and then for any subset X of {0,1,...,n − 1} we let a be the product of those bi such that i is in X. Then bi divides a if and only if i ∈ X.
Every o-minimal theory satisfies NIP. This fact has had unexpected applications to neural network learning.
Examples of NIP theories include also the theories of all the following structures: linear orders, trees, abelian linearly ordered groups, algebraically closed valued fields, and the p-adic field for any p.
Notes
- See Hodges.
- See Poizat, page 249.
- Pillay and Steinhorn, corollary 3.10 and Knight, Pillay, and Steinhorn, theorem 0.2.
- See Anthony and Bartlett for details.
- See Simon, Appendix A.
References
- Anthony, Martin; Bartlett, Peter L. (1999). Neural network learning: theoretical foundations. Cambridge University Press. ISBN 978-0-521-57353-5.
- Hodges, Wilfrid (1993). Model theory. Cambridge University Press. ISBN 978-0-521-30442-9.
- Knight, Julia; Pillay, Anand; Steinhorn, Charles (1986). "Definable sets in ordered structures II". Transactions of the American Mathematical Society. 295 (2): 593–605. doi:10.2307/2000053. JSTOR 2000053.
- Pillay, Anand; Steinhorn, Charles (1986). "Definable sets in ordered structures I". Transactions of the American Mathematical Society. 295 (2): 565–592. doi:10.2307/2000052. JSTOR 2000052.
- Poizat, Bruno (2000). A Course in Model Theory. Springer. ISBN 978-0-387-98655-5.
- Simon, Pierre (2015). A Guide to NIP Theories. Cambridge University Press. ISBN 9781107057753.