Grosch's law


Grosch's law is the following observation of computer performance, made by Herb Grosch in 1953:
I believe that there is a fundamental rule, which I modestly call Grosch's law, giving added economy only as the square root of the increase in speed — that is, to do a calculation ten times as cheaply you must do it hundred times as fast.

This adage is more commonly stated as
Computer performance increases as the square of the cost. If computer A costs twice as much as computer B, you should expect computer A to be four times as fast as computer B.

Two years before Grosch's statement, Seymour Cray was quoted in Business Week expressing this very same thought:
Computers should obey a square law — when the price doubles, you should get at least four times as much speed.

The law can also be interpreted as meaning that computers present economies of scale: the more costly is the computer, the price–performance ratio linearly becomes better. This implies that low-cost computers cannot compete in the market.
An for computers between 1951 and 1963 by Knight found that Grosch's law held for commercial and scientific operations. In a found that Grosch's law did not apply to computers between 1963-1967.

Debates

asserted in 1997, that "it was never clear whether Grosch's Law was a reflection of how IBM priced its computers or whether it related to actual costs. It provided the rationale that a bigger computer is always better. The IBM sales force used Grosch's rationale to persuade organizations to acquire more computing capacity than they needed. Grosch's Law also became the justification for offering time-sharing services from big data centers as a substitute for distributed computing." Grosch himself has stated that the law was more useful in the 1960s and 1970s than it is today. He originally intended the law to be a "means for pricing computing services".