Added to Favorites

Related Searches

Definitions

In mathematics, a quantity that can be expressed as a finite or infinite decimal expansion. The counting numbers, integers, rational numbers, and irrational numbers are all real numbers. Real numbers are used in measuring continuously varying quantities (e.g., size, time), in contrast to measurements that result from counting. The word *real* distinguishes them from the imaginary numbers.

Learn more about real number with a free trial on Britannica.com.

Encyclopedia Britannica, 2008. Encyclopedia Britannica Online.

A real number a is first-order definable in the language of set theory, without parameters, if there is a formula φ in the language of set theory, with one free variable, such that a is the unique real number such that φ(a) holds (in the von Neumann universe V).

For the purposes of this article, such reals will be called simply definable numbers. This should not be understood to be standard terminology.

Note that this definition cannot be expressed in the language of set theory itself.

Assuming they form a set, the definable numbers form a field containing all the familiar real numbers such as 0, 1, π, e, et cetera. In particular, this field contains all the numbers named in the mathematical constants article, and all algebraic numbers (and therefore all rational numbers). However, most real numbers are not definable: the set of all definable numbers is countably infinite (because the set of all logical formulas is) while the set of real numbers is uncountably infinite (see Cantor's diagonal argument). As a result, most real numbers have no description (in the same sense of "most" as 'most real numbers are not rational').

The field of definable numbers is not complete; there exist convergent sequences of definable numbers whose limit is not definable (since every real number is the limit of a sequence of rational numbers). However, if the sequence itself is definable in the sense that we can specify a single formula for all its terms, then its limit will necessarily be a definable number.

While every computable number is definable, the converse is not true: the numeric representations of the Halting problem, Chaitin's constant, the truth set of first order arithmetic, and 0^{#} are examples of numbers that are definable but not computable. Many other such numbers are known.

One may also wish to talk about definable complex numbers: complex numbers which are uniquely defined by a logical formula. However, whether this is possible depends on how the field of complex numbers is derived in the first place: it may not be possible to distinguish a complex number from its conjugate (say, 3+i from 3-i), since it is impossible to find a property of one that is not also a property of the other, without falling back on the underlying set-theoretic definition. Assuming we can define at least one nonreal complex number, however, a complex number is definable if and only if both its real part and its imaginary part are definable. The definable complex numbers also form a field if they form a set.

The related concept of "standard" numbers, which can only be defined within a finite time and space, is used to motivate axiomatic internal set theory, and provide a workable formulation for illimited and infinitesimal number. Definitions of the hyper-real line within non-standard analysis (the subject area dealing with such numbers) overwhelmingly include the usual, uncountable set of real numbers as a subset.

- $forall\; m\; ,\; forall\; n\; ,\; (phi(n,m)ifffrac\{n\}\{m\}).\; math>$

The other sorts of definability thus far considered have only countably many defining formulas, and therefore allow only countably many definable reals. This is not true for ordinal definability, because an ordinal definable real is defined not only by the formula φ, but also by the ordinal γ. In fact it is consistent with ZFC that all reals are ordinal-definable, and therefore that there are uncountably many ordinal-definable reals. However it is also consistent with ZFC that there are only countably many ordinal-definable reals.

- Alan Turing, "On Computable Numbers, With An Application to the Entscheidungsproblem", Proceedings of the London Mathematical Society, 1936 (Turing's original paper distinguishing computable and definable numbers)

Wikipedia, the free encyclopedia © 2001-2006 Wikipedia contributors (Disclaimer)

This article is licensed under the GNU Free Documentation License.

Last updated on Friday September 19, 2008 at 12:14:29 PDT (GMT -0700)

View this article at Wikipedia.org - Edit this article at Wikipedia.org - Donate to the Wikimedia Foundation

This article is licensed under the GNU Free Documentation License.

Last updated on Friday September 19, 2008 at 12:14:29 PDT (GMT -0700)

View this article at Wikipedia.org - Edit this article at Wikipedia.org - Donate to the Wikimedia Foundation

Copyright © 2014 Dictionary.com, LLC. All rights reserved.