Used in distinction to digital computers, in which physical or mechanical phenomena are used to construct a finitestate machine[?] which is then used to model the problem being solved. There is an intermediate group, Hybrid computers[?], in which a digital computer is used to control and organize inputs and outputs to and from attached analogue devices; for instance analogue devices might be used to help generate initial values for iterations.
Analog computers operate on real numbers and are differential, whereas digital computers are limited to computable numbers and are algebraic[?]. This means that analog computers have a larger information dimension rate[?] (see Information Theory), or potential computing domain, than do digital computers (see Gödel's incompleteness theorem). This enables analog computers to solve problems that are inextricable on digital computers.
Computer theorists often refer to the analog computer as a real computer[?] (so called because it operates on the set of real numbers), in order to circumvent popular misconceptions of analog computers.
The first analog computer was invented by Charles Babbage in 1835.
Also see signals, set theory, computability theory, differential equation, dynamical systems, chaos theory.
An example:
Examples of analog computers:
Search Encyclopedia

Featured Article
