Parallel computer

From Computer History Wiki
Jump to: navigation, search

A parallel computer is one which uses a parallel implementation for operations on larger units of data (e.g. the addition of two numbers). Essentially all computers are now parallel, since ICs have reduced the cost of transistors, the usual basic functional unit of modern computers, to essentially nothing.

For example, instead of having only a single-bit adder (so that to add two numbers, they are fed into it a bit at a time, one bit on each clock tick, starting with the least significant bit), a parallel computer would have a word-wide adder, which would add two words together in a single clock tick.

Serial computers were more common in the early stages of computing; they are slower, and have more complex control logic, but use fewer components - an acceptable trade-off at that stage, when the technology (e.g. vacuum tubes) was more expensive and physically bulky.

See also