Added to Favorites

Related Searches

Nearby Words

In physics the word system has a technical meaning, namely, it is the portion of the physical universe chosen for analysis. Everything outside the system is known as the environment, which in analysis is ignored except for its effects on the system. The cut between system and environment is a free choice, generally made to simplify the analysis as much as possible. An isolated system is one which has negligible interaction with its environment.

Often a system in this sense is chosen to correspond to the more usual meaning of system, such as a particular machine. But physical systems are often more esoteric: an atom, the water in a lake, or indeed the water in the left-hand half of a lake can all be considered as physical systems. In the study of quantum decoherence the "system" may refer to the macroscopic properties of an object (e.g. the position of a pendulum bob), while the relevant "environment" may be the internal degrees of freedom, described classically by the pendulum's thermal vibrations.

If one considers a classical Newtonian ball situation with a number of perfectly moving physical bodies bouncing off the walls of a container, the system-state probability does not change over time. The entropy of the system changes over time, but the probability of the state vector does not change. One can periodically evaluate the complexity of this system, and the complexity of this system does not change.

In a physical system, a lower probability state vector is equivalent to a higher complexity. A self sustaining low probability state vector allows the physical system to remain in a higher complexity state. The study of such systems as applied to our universe is in its infancy and speculative in nature, but it appears that there are some low probability systems that are able to sustain themselves through time.

In mathematical systems, one can consider the complexity of particular states more easily. For example, if one considers a Turing machine that generates random symbols and then utilizes these symbols as an algorithm to create a new series of symbols, the complexity of the final string of symbols is nearly mathematically equivalent to the minimum size of a string required to produce a larger string on a Turing machine as defined by algorithmic information theory.

- the first began with reference to Newtonian mechanics,
- the second with nineteenth century thermodynamics,
- the third throughout the twentieth century with Lorentz, Einstein, and Minkowski,
- and finally with conceptual developments in quantum theory and wave mechanics.

The third and fourth stages eventually led to more open systems concepts, but closed/open system concepts really need be considered in relation to work since that time if one were to define further conceptual developments.

Wikipedia, the free encyclopedia © 2001-2006 Wikipedia contributors (Disclaimer)

This article is licensed under the GNU Free Documentation License.

Last updated on Tuesday September 23, 2008 at 06:58:47 PDT (GMT -0700)

View this article at Wikipedia.org - Edit this article at Wikipedia.org - Donate to the Wikimedia Foundation

This article is licensed under the GNU Free Documentation License.

Last updated on Tuesday September 23, 2008 at 06:58:47 PDT (GMT -0700)

View this article at Wikipedia.org - Edit this article at Wikipedia.org - Donate to the Wikimedia Foundation

Copyright © 2015 Dictionary.com, LLC. All rights reserved.