A machine executes 1 instruction per microsecond (that is, 1,000,000 instructions per second).
Algorithm k (for k = 1, 2, 3, 4) takes Tk(N) microseconds for input of size N.
What is the largest input size, N, that each algorithm can complete in 1 second?
T1(N) = log(N) (base 2)
T2(N) = N
T3(N) = N2
T4(N) = N3