Apex Institute of Technology Department of Computer Science & Engineering
Apex Institute of Technology Department of Computer Science & Engineering
• Like all recursive functions, it has one or more recursive cases and one or more base cases.
• Example:
• The portion of the definition that does not contain T is called the base case of the recurrence relation; the portion that
contains T is called the recurrent or recursive case.
• Recurrence relations are useful for expressing the running times (i.e., the number of basic operations executed) of
recursive algorithms
• The specific values of the constants such as a, b, and c (in the above recurrence) are important in determining the exact
solution to the recurrence. Often however we are only concerned with finding an asymptotic upper bound on the solution.
We call such a bound an asymptotic solution to the recurrence.
Forming Recurrence Relations
• For a given recursive method, the base case and the recursive case of its recurrence relation correspond directly to the base case and the
recursive case of the method.
• Example 1: Write the recurrence relation for the following method:
public void f (int n) {
if (n > 0) {
System.out.println(n);
f(n-1);
}
}
• The base case is reached when n = = 0. The method performs one comparison. Thus, the number of operations when n = = 0, T(0), is
some constant a.
• When n > 0, the method performs two basic operations and then calls itself, using ONE recursive call, with a parameter n – 1.
• Therefore the recurrence relation is:
T(0) = a for some constant a
• In General, T(n) is usually a sum of various choices of T(m ), the cost of the recursive
T(n) = b + T(n – 1) for some constant b
subproblems, plus the cost of the work done outside the recursive calls:
T(n ) = aT(f(n)) + bT(g(n)) + . . . + c(n)
where a and b are the number of subproblems, f(n) and g(n) are subproblem sizes, and
c(n) is the cost of the work done outside the recursive calls [Note: c(n) may be a constant]
Forming Recurrence Relations (Cont’d)
• Example 2: Write the recurrence relation for the following method:
• The base case is reached when n == 1. The method performs one comparison and one return statement. Therefore,
T(1), is some constant c.
• When n > 1, the method performs TWO recursive calls, each with the parameter n / 2, and some constant # of basic
operations.
• The base case is reached when n == 1 or n == 2. The method performs one comparison and one return statement.
Therefore each of T(1) and T(2) is some constant c.
• When n > 2, the method performs TWO recursive calls, one with the parameter n - 1 , another with parameter n – 2,
and some constant # of basic operations.
• The base case is reached when n == 0 or n == 1. The method performs one comparison and one return statement. ThereforeT(0) and
T(1) is some constant c.
• At every step the problem size reduces to half the size. When the power is an odd number, an additional multiplication is involved. To
work out time complexity, let us consider the worst case, that is we assume that at every step an additional multiplication is needed.
Thus total number of operations T(n) will reduce to number of operations for n/2, that is T(n/2) with seven additional basic operations
(the odd power case)
• There are five methods to solve recurrence relations that represent the running time of
recursive methods:
Iteration method (unrolling and summing)
Substitution method (Guess the solution and verify by induction)
Recursion tree method
Master theorem (Master method)
Using Generating functions or Characteristic equations
• In this course, we will use the Iteration method and a simplified Master theorem.
Solving Recurrence Relations - Iteration method
• Steps:
Expand the recurrence
Express the expansion as a summation by plugging the recurrence back into itself until
you see a pattern.
Evaluate the summation
• In evaluating the summation one or more of the following summation
formulae may be used:
• Special Cases of Geometric Series:
• Arithmetic series:
Geometric Series:
Solving Recurrence Relations - Iteration method (Cont’d)
Harmonic Series:
Others:
Analysis Of Recursive Factorial method
Example1: Form and solve the recurrence relation for the running time of
factorial method and hence determine its big-O complexity:
long factorial (int n) {
if (n == 0)
return 1;
else
return n * factorial (n – 1);
}
T(0) = c (1)
T(n) = b + T(n - 1) (2)
= b + b + T(n - 2) by subtituting T(n – 1) in (2)
= b +b +b + T(n - 3) by substituting T(n – 2) in (2)
…
= kb + T(n - k)
The base case is reached when n – k = 0 k = n, we then have:
T(n) = nb + T(n - n)
= bn + T(0)
= bn + c
Therefore the method factorial is O(n)
Analysis Of Recursive Selection Sort
public static void selectionSort(int[] x) {
selectionSort(x, x.length - 1);
}
• The recurrence relation for the running time of the method is:
T(1) = a if n = 1 (one element array)
T(n) = T(n / 2) + b if n > 1
Analysis Of Recursive Binary Search (Cont’d)
Without loss of generality, assume n, the problem size, is a multiple of 2, i.e., n = 2 k
Expanding:
T(1) = a (1)
T(n) = T(n / 2) + b (2)
= [T(n / 22) + b] + b = T (n / 22) + 2b by substituting T(n/2) in (2)
= [T(n / 23) + b] + 2b = T(n / 23) + 3b by substituting T(n/22) in (2)
= ……..
= T( n / 2k) + kb
public static void hanoi(int n, char from, char to, char temp){
if (n == 1)
System.out.println(from + " --------> " + to);
else{
hanoi(n - 1, from, temp, to);
System.out.println(from + " --------> " + to);
hanoi(n - 1, temp, to, from);
}
}
• The recurrence relation for the running time of the method hanoi is:
T(n) = a if n = 1
T(n) = 2T(n - 1) + b if n > 1
Analysis Of Recursive Towers of Hanoi Algorithm (Cont’d)
Expanding:
T(1) = a (1)
T(n) = 2T(n – 1) + b if n > 1 (2)
= 2[2T(n – 2) + b] + b = 22 T(n – 2) + 2b + b by substituting T(n – 1) in (2)
= 22 [2T(n – 3) + b] + 2b + b = 23 T(n – 3) + 22b + 2b + b by substituting T(n-2) in (2)
= 23 [2T(n – 4) + b] + 22b + 2b + b = 24 T(n – 4) + 23 b + 22b + 21b + 20b by substituting
T(n – 3) in (2)
= ……
= 2k T(n – k) + b[2k- 1 + 2k– 2 + . . . 21 + 20]
T(n) = c if n = 1 or n = 2 (1)
T(n) = T(n – 1) + T(n – 2) + b if n > 2 (2)
• If T(n) is interpreted as the number of steps needed to execute an algorithm for an input of size n, this recurrence corresponds to a
“divide and conquer” algorithm, in which a problem of size n is divided into a sub-problems of size n / b, where a, b are positive
constants:
Divide-and-conquer algorithm:
• divide the problem into a number of subproblems
• conquer the subproblems (solve them)
• combine the subproblem solutions to get the solution to the original problem
Example: Merge Sort
• divide the n-element sequence to be sorted into two n/2- element sequences.
• conquer the subproblems recursively using merge sort.
• combine the resulting two sorted n/2-element sequences by merging
Simplified Master Theorem
• The Simplified Master Method for Solving Recurrences:
log ba
T(n) O(n )
if a > b c
c
T(n) O(n log n)
if ac=) bc
T(n) O(n
• Note: Since k and h do not affect the result, they are sometimes not included
in the above recurrence
if a < b c
Simplified Master Theorem (Cont’d)
Example1: Find the big-Oh running time of the following recurrence. Use the Master
Theorem:
Example3: Find the big-Oh running time of the following recurrence. Use the Master Theorem:
T(1) = 1
T(n) = 4T(n / 2) + kn3 + h where k ≥ 1 and h 1
Solution: a = 4, b = 2, c = 3 a < bc Case 3
Hence T(n) is O(n3)
References
• Fundamentals of Computer Algorithms 2nd Edition (2008) by Horowitz, Sahni
and Rajasekaran
• Introduction to Algorithms 3rd Edition (2012) by Thomas H Cormen, Charles E
Lieserson, Ronald
22
THANK YOU
For queries
Email: [email protected]
03/07/2022 23