Home     Overview     Materials     Assignments     Problem Set Submission     Software
CS212CS212 Handouts
Reasoning About Resources

Asymptotic Analysis

When analyzing the running time or space usage of programs, we usually try to estimate the time or space as function of the input size. Thus we might say that the program times-1 (below) runs in time linear in the magnitude of b. By this we mean that the function runs in time at most cb where b is some fixed constant c.

(define (times-1 a b)
    (if (zero? b)
        0
        (+ a (times-1 a (dec b))))))

The asymptotic behavior of a function f refers to the growth of f(n) as n gets large. We typically ignore small values of n, since we are usually interested in estimating how slow the program will be on large inputs. A good rule of thumb is: the slower the asymptotic growth rate, the better the algorithm (although this is often not the whole story).

By this measure, a linear algorithm is always asymptotically better than a quadratic one. That is because for any given (positive) c and d, there is always some n at which the magnitude of cn2 overtakes dn (to be precise, for all n > d/c). For moderate values of n, the quadratic algorithm could very well take less time than the linear one, for example if c is significantly smaller than d. However, the linear algorithm will always be better for sufficiently large inputs. Remember to THINK BIG when working with asymptotic rates of growth.

Worst-Case and Average-Case Analysis

When we say that an algorithm runs in time f(n), we mean that f(n) is an upper bound on the running time that holds for all inputs of size n. This is called worst-case analysis. The algorithm may very well take less time on some inputs of size n, but it doesn't matter. If an algorithm takes n2 steps on only a single input of each size n and only n steps on the rest, it is still a quadratic algorithm.

A popular alternative to worst-case analysis is average-case analysis. Here we do not bound the worst case running time, but try to calculate the expected time spent on a randomly chosen input. This kind of analysis is generally harder, since it involves probabilistic arguments and often requires assumptions about the distribution of inputs that may be difficult to justify.

Order of Growth and Big-O Notation

In estimating the running time of times-1 (or any other procedure) we don't know (or care) what the constant c is. We know that it is a constant of moderate size, but other than that it is not important; we have enough evidence from the asymptotic analysis to know that a logarithmic time algorithm for multiplication (see below) is faster than the linear version, even though the constants may differ somewhat. (This does not always hold, the constants can sometimes make a difference, but it has been a good enough rule-of-thumb that it is quite widely applied.)

We may not even be able to measure the constant c directly. For example, we may know that a given expression of the language, such as if takes a constant number of machine instructions, but we may not know exactly how many. Moreover, the same sequence of instructions executed on a Pentium will take less time than on a 486 (although the difference will be roughly a constant factor). So these estimates are usually only accurate up to a constant factor. For these reasons, we usually ignore constant factors in comparing asymptotic running times.

Computer scientists have developed a convenient notation for hiding the constant factor. We write O(n) (read: ''order n'') instead of ''cn for some constant c.'' Thus an algorithm is said to be O(n) or linear time if there is a fixed constant c such that for all sufficiently large n, the algorithm takes time at most cn on inputs of size n. An algorithm is said to be O(n2) or quadratic time if there is a fixed constant c such that for all sufficiently large n, the algorithm takes time at most cn2 on inputs of size n. O(1) means constant time.

Polynomial time means nO(1), or nc for some constant c. Thus any constant, linear, quadratic, or cubic (O(n3)) time algorithm is a polynomial-time algorithm.

This is called big-O notation. It captures the important differences in the asymptotic growth rates of functions.

One important advantage of big-O notation is that it makes algorithms much easier to analyze, since we can conveniently ignore low-order terms. For example, an algorithm that runs in time

10n3 + 24n2 + 3n log n + 144

is still a cubic algorithm, since

10n3 + 24n2 + 3n log n + 144
<= 10n3 + 24n3 + 3n3 + 144n3
<= (10 + 24 + 3 + 144)n3
= O(n3)
.

Of course, since we are ignoring constant factors, any two linear algorithms will be considered equally good by this measure. There may even be some situations in which the constant is so huge in a linear algorithm that even an exponential algorithm with a small constant may be preferable in practice. This is a valid criticism of asymptotic analysis and big-O notation. However, as a rule of thumb it has served us well. Just be aware that it is only a rule of thumb--the asymptotically optimal algorithm is not necessarily the best one.

Some common orders of growth seen often in complexity analysis are

O(1) constant
O(log n) logarithmic
O(n) linear
O(n log n) "n log n"
O(n2) quadratic
O(n3) cubic
nO(1) polynomial
2O(n) exponential

Here log means log2 or the logarithm to the base 2, although it doesn't really matter since logarithms to different bases differ by a constant factor. Note also that 2O(n) and O(2n) are not the same!

Comparing Orders of Growth

Definition
Let f and g be functions from positive integers to positive integers. We say f = O(g) (read: ''f is order g'') if there exists a fixed constant c such that for all n,

f(n) < cg(n).

Equivalently, f is O(g) if the function f(n)/g(n) is bounded above.

We say f = o(g) (read: ``f is little-o of g'') if for all arbitrarily small real c > 0, for all but perhaps finitely many n,

f(n) < cg(n).

Equivalently, f is o(g) if the function f(n)/g(n) tends to 0 as n tends to infinity.

Here are some examples:

We now introduce some convenient rules for manipulating expressions involving order notation. These rules, which we state without proof, are useful for working with orders of growth:

  1. cnm = O(nk) for any constant c and any m <= k.
  2. O(f(n)) + O(g(n)) = O(f(n) + g(n)).
  3. O(f(n))O(g(n)) = O(f(n)g(n)).
  4. O(cf(n)) = O(f(n)) for any constant c.
  5. c is O(1) for any constant c.
  6. logbn = O(log n) for any base b.

Analyzing Running Times of Procedures

Now we can use these ideas to analyze the asymptotic running time of Scheme procedures. The use of order notation can greatly simplify our task here. We assume that the primitive operations of our language, such as arithmetic operations, tests, and things like pair, head, and tail all take constant time.

Consider the following multiplication routine:

(define (times-1 a b)
    (if (zero? b)
        0
        (+ a (times-1 a (dec b))))))  

What is the order of growth of the time required by times-1 as a function of n, where n is the magnitude of the parameter b? Note that the ``size'' of a number can be measured either in terms of its magnitude or in terms of the number of digits (the space it takes to write the number down). Often the number of digits is used, but here we use the magnitude. Note that it takes only about log10 x digits to write down a number of magnitude x, thus these two measures are very different. Make sure you know which one is being used.

We assume that all the primitive operations in the times-1 procedure (zero?, if, +, and dec) and the overhead for procedure calls take constant time. Thus if n = 0, the procedure takes constant time. If n > 0, the time taken on an input of magnitude n is constant time plus the time taken by the recursive call on n-1. Thus the running time T(n) of times-1 is a solution of

T(n) = T(n-1) + O(1) for n > 0
T(0) = O(1)

This is called a recurrence relation. It simply states that the time to multiply a number a by another number b of size n > 0 is the time required to multiply a by a number of size n-1 plus a constant amount of work (the primitive operations performed). Furthermore, the time to multiply a by zero is a constant (only constant-time primitives are involved). In other words, there are constants c and d such that T(n) satisfies

T(n) = T(n-1) + c for n > 0
T(0) = d

This more specific recurrence relation has a unique closed form solution, namely

T(n) = d + cn

which is O(n), so the algorithm is linear in the magnitude of its second argument. One can obtain this equation by generalizing from small values of n, then prove that it is indeed a solution to the recurrence relation by induction on n.

Now consider the following procedure for multiplying two numbers:

(define (fast-times a b)
  (if (zero? b) 0 
      (if (even? b) (fast-times (double a) (halve b))
          (+ a (fast-times a (dec b))))))

Again we want an expression for the running time in terms of n, the magnitude of the parameter b. We assume that double and halve are constant time as well as the standard primitives (these could be done in constant time using arithmetic shift). The recurrence for this problem is more complicated than the previous one:

T(n) = T(n-1) + O(1) if n > 0 and n is odd
T(n) = T(n/2) + O(1) if n > 0 and n is even
T(0) = O(1)

We somehow need to figure out how often the first versus the second branch of this recurrence will be taken. It's easy if n is a power of two, i.e. if n = 2m for some integer m. In this case, the second branch of will only get taken when n = 1, because 2m is even except when m = 0, i.e. when n = 1. Note further that T(1) = O(1) because T(1) = T(0) + O(1) = O(1) + O(1) = O(1). Thus, for this special case we get the recurrence

T(n) = T(n/2) + O(1) if n > 0 and n is a power of 2
T(0) = O(1)

or

T(n) = T(n/2) + c for n > 0 and n is a power of 2
T(0) = d

for some constants c and d. The closed form solution of this is, for powers of 2,

T(n) = d + c log2 n

which is O(log n).

What if n is not a power of 2? The running time is still O(log n) even in this more general case. Intuitively, this is because if n is odd, then n-1 is even, so on the next recursive call the input will be halved. Thus the input is halved at least once in every two recursive calls, which is all you need to get O(log n).

A good way to handle this formally is to charge the cost of a call to fast-times on an odd input to the recursive call on an even input that must immediately follow it. We reason as follows: on an even input n, the cost is the cost of the recursive call on n/2 plus a constant, or

T(n) = T(n/2) + O(1)

as before. On an odd input n, we recursively call the procedure on n-1, which is even, so we immediately call the procedure again on (n-1)/2. Thus the total cost on an odd input is the cost of the recursive call on (n-1)/2 plus a constant. In this case we get

T(n) = T((n-1)/2) + O(1)

In either case,

T(n) <= T(n/2) + O(1)

which has the solution O(log n). This approach is more or less the same as explicitly unwinding the if clause that handles odd inputs:

(define (fast-times a b)
    (if (zero? b) 0
          (if (even? b) (fast-times (double a) (halve b))
              (+ a (fast-times (double a) (halve (dec b)))))))

then analyzing the rewritten program, without actually doing the rewriting.

Charging one operation to another (bounding the number of times one thing can happen by the number of times that another thing happens) is a common technique for analyzing the running time of complicated algorithms.

Order notation is a useful tool, and should not be thought of as being just a theoretical exercise. For example, the practical difference in running times between the logarithmic fast-times and the linear times-1 is noticeable even for moderate values of n. With the a version of Scheme on an old machine, (times-1 10 10) takes 0.25 seconds and (fast-times 10 10) takes 0.22 seconds--not a big difference. However, when n = 10,000, times-1 takes 8.37 seconds whereas fast-times takes only 0.3 seconds. This is already quite a big difference--about a factor of 30. When n = 100,000, times-1 runs out of stack space before finishing, because there are too many deferred operations. The fast-times procedure still only requires about 0.3 seconds even for this large value of n.

The key points in this handout are:




Valid HTML 4.0!CS212 Home Page
© 1999 Cornell University Computer Science