Lecture 9. Conclusion.
Transcription
Lecture 9. Conclusion.
Algorithms (IX) Guoqiang Li School of Software, Shanghai Jiao Tong University Q: What we have learned in Algorithm? Algorithm Design Algorithm Design Algorithm Design • Basic methodologies: Algorithm Design • Basic methodologies: • Divide and Conquer (Chapter 2) Algorithm Design • Basic methodologies: • Divide and Conquer (Chapter 2) • Master Theorem Algorithm Design • Basic methodologies: • Divide and Conquer (Chapter 2) • Master Theorem • Recursion (Chapter 0 & 1) Algorithm Design • Basic methodologies: • Divide and Conquer (Chapter 2) • Master Theorem • Recursion (Chapter 0 & 1) • Advanced topics: Algorithm Design • Basic methodologies: • Divide and Conquer (Chapter 2) • Master Theorem • Recursion (Chapter 0 & 1) • Advanced topics: • Dynamic Programming (Chapter 6) Algorithm Design • Basic methodologies: • Divide and Conquer (Chapter 2) • Master Theorem • Recursion (Chapter 0 & 1) • Advanced topics: • Dynamic Programming (Chapter 6) • Filling in Squares. Algorithm Design • Basic methodologies: • Divide and Conquer (Chapter 2) • Master Theorem • Recursion (Chapter 0 & 1) • Advanced topics: • Dynamic Programming (Chapter 6) • Filling in Squares. • Complexity. Algorithm Design • Basic methodologies: • Divide and Conquer (Chapter 2) • Master Theorem • Recursion (Chapter 0 & 1) • Advanced topics: • Dynamic Programming (Chapter 6) • Filling in Squares. • Complexity. • Greedy Algorithm (Chapter 5) Algorithm Design • Basic methodologies: • Divide and Conquer (Chapter 2) • Master Theorem • Recursion (Chapter 0 & 1) • Advanced topics: • Dynamic Programming (Chapter 6) • Filling in Squares. • Complexity. • Greedy Algorithm (Chapter 5) • Linear Programming (Chapter 7) Algorithm Design • Basic methodologies: • Divide and Conquer (Chapter 2) • Master Theorem • Recursion (Chapter 0 & 1) • Advanced topics: • Dynamic Programming (Chapter 6) • Filling in Squares. • Complexity. • Greedy Algorithm (Chapter 5) • Linear Programming (Chapter 7) • Simplex. Algorithm Design • Basic methodologies: • Divide and Conquer (Chapter 2) • Master Theorem • Recursion (Chapter 0 & 1) • Advanced topics: • Dynamic Programming (Chapter 6) • Filling in Squares. • Complexity. • Greedy Algorithm (Chapter 5) • Linear Programming (Chapter 7) • Simplex. • System Design. Algorithm Design • Basic methodologies: • Divide and Conquer (Chapter 2) • Master Theorem • Recursion (Chapter 0 & 1) • Advanced topics: • Dynamic Programming (Chapter 6) • Filling in Squares. • Complexity. • Greedy Algorithm (Chapter 5) • Linear Programming (Chapter 7) • Simplex. • System Design. • Duality. Algorithm Design • Basic methodologies: • Divide and Conquer (Chapter 2) • Master Theorem • Recursion (Chapter 0 & 1) • Advanced topics: • Dynamic Programming (Chapter 6) • Filling in Squares. • Complexity. • Greedy Algorithm (Chapter 5) • Linear Programming (Chapter 7) • Simplex. • System Design. • Duality. • Randomized Algorithm Algorithm Design • Basic methodologies: • Divide and Conquer (Chapter 2) • Master Theorem • Recursion (Chapter 0 & 1) • Advanced topics: • Dynamic Programming (Chapter 6) • Filling in Squares. • Complexity. • Greedy Algorithm (Chapter 5) • Linear Programming (Chapter 7) • Simplex. • System Design. • Duality. • Randomized Algorithm • Approximation Algorithm Algorithm Analysis Algorithm Analysis Algorithm Analysis • Basic Knowledge Algorithm Analysis • Basic Knowledge • Big-O Notation (Chapter 0) Algorithm Analysis • Basic Knowledge • Big-O Notation (Chapter 0) • Recursion and Loop (Chapter 1) Algorithm Analysis • Basic Knowledge • Big-O Notation (Chapter 0) • Recursion and Loop (Chapter 1) • geometrical Analysis (Chapter 2) Algorithm Analysis • Basic Knowledge • Big-O Notation (Chapter 0) • Recursion and Loop (Chapter 1) • geometrical Analysis (Chapter 2) • Advanced Methodology: Algorithm Analysis • Basic Knowledge • Big-O Notation (Chapter 0) • Recursion and Loop (Chapter 1) • geometrical Analysis (Chapter 2) • Advanced Methodology: • Probability Analysis (Chapter 1) Algorithm Analysis • Basic Knowledge • Big-O Notation (Chapter 0) • Recursion and Loop (Chapter 1) • geometrical Analysis (Chapter 2) • Advanced Methodology: • Probability Analysis (Chapter 1) • Amortized Analysis (Chapter 5) Algorithm Analysis • Basic Knowledge • Big-O Notation (Chapter 0) • Recursion and Loop (Chapter 1) • geometrical Analysis (Chapter 2) • Advanced Methodology: • Probability Analysis (Chapter 1) • Amortized Analysis (Chapter 5) • Competition Analysis (Chapter 9) Standard Algorithm Standard Algorithms Standard Algorithms • Arithmetic Algorithm Standard Algorithms • Arithmetic Algorithm • Sorting Standard Algorithms • Arithmetic Algorithm • Sorting • Searching • Strongly Connected Components Standard Algorithms • Arithmetic Algorithm • Sorting • Searching • Strongly Connected Components • Finding shortest paths in graphs Standard Algorithms • Arithmetic Algorithm • Sorting • Searching • Strongly Connected Components • Finding shortest paths in graphs • Minimum spanning trees in graphs Standard Algorithms • Arithmetic Algorithm • Sorting • Searching • Strongly Connected Components • Finding shortest paths in graphs • Minimum spanning trees in graphs • Matchings in bipartite graphs Standard Algorithms • Arithmetic Algorithm • Sorting • Searching • Strongly Connected Components • Finding shortest paths in graphs • Minimum spanning trees in graphs • Matchings in bipartite graphs • Maximum flows in networks Advanced Data Structure Advanced Data Structure Advanced Data Structure • Trees, Graphs Advanced Data Structure • Trees, Graphs • Kripke Structure, Automata Advanced Data Structure • Trees, Graphs • Kripke Structure, Automata • Priority Queue (Chapter 4) Advanced Data Structure • Trees, Graphs • Kripke Structure, Automata • Priority Queue (Chapter 4) • Disjoint Set (Chapter 5) What Is Important in Algorithm under My Concern? I. Proof and Witness Proof and Witness • Mathematical Induction. Proof and Witness • Mathematical Induction. • Proof of iff, =, ⇔ Proof and Witness • Mathematical Induction. • Proof of iff, =, ⇔ • Deduction. Examples log n! = Θ(n log n) Examples log n! = Θ(n log n) A directed graph has a cycle if and only if its depth-first search reveals a back edge. Examples log n! = Θ(n log n) A directed graph has a cycle if and only if its depth-first search reveals a back edge. An undirected graph is bipartite if and only if it contains no cycles of odd length. II. Reductions Bipartite Matching BOYS GIRLS Al Alice Bob Beatrice Chet Carol Dan Danielle Bipartite Matching Al Alice Bob Beatrice Chet Carol Dan Danielle s t Reductions • We want to solve Problem P. • We already have an algorithm that solves Problem Q. • If any subroutine for Q can also be used to solve P, we say P reduces to Q. • Often, P is solvable by a single call to Q’s subroutine, which means any instance x of P can be transformed into an instance y of Q such that P(x) can be deduced from Q(y). III. Divide-and-Conquer Vs. Dynamic Programming Master Theorem If T (n) = aT (dn/be) + O(nd ) for some constants a > 0, b > 1 and d ≥ 0, then d if d > logb a O(n ) d T (n) = O(n log n) if d = logb a O(nlogb a ) if d < logb a The Proof of the Theorem Proof: • Assume that n is a power of b. • The size of the subproblems decreases by a factor of b with each level of recursion, and therefore reaches the base case after logb n levels - the the height of the recursion tree. • Its branching factor is a, so the k-th level of the tree is made up of ak subproblems, each of size n/bk . ak × O( n d a ) = O(nd ) × ( d )k bk b • k goes from 0 to logb n, these numbers form a geometric series with ratio a/bd , comes down to three cases. The Proof of the Theorem • The ratio is less than 1. Then the series is decreasing, and its sum is just given by its first term, O(nd ). • The ratio is greater than 1. The series is increasing and its sum is given by its last term, O(nlogb a ) • The ratio is exactly 1. In this case all O(log n) terms of the series are equal to O(nd ). The Problem • When a spell checker encounters a possible misspelling, it looks in its dictionary for other words that are close by. S S − U N O W N N − Cost: 3 Y Y − S S U N O W N − − Cost: 5 − N Y Y The Problem • When a spell checker encounters a possible misspelling, it looks in its dictionary for other words that are close by. • Q: What is the appropriate notion of closeness in this case? S S − U N O W N N − Cost: 3 Y Y − S S U N O W N − − Cost: 5 − N Y Y The Problem • When a spell checker encounters a possible misspelling, it looks in its dictionary for other words that are close by. • Q: What is the appropriate notion of closeness in this case? • A natural measure of the distance between two strings is the extent to which they can be aligned, or matched up. S S − U N O W N N − Cost: 3 Y Y − S S U N O W N − − Cost: 5 − N Y Y The Problem • When a spell checker encounters a possible misspelling, it looks in its dictionary for other words that are close by. • Q: What is the appropriate notion of closeness in this case? • A natural measure of the distance between two strings is the extent to which they can be aligned, or matched up. • Technically, an alignment is simply a way of writing the strings one above the other. S S − U N O W N N − Cost: 3 Y Y − S S U N O W N − − Cost: 5 − N Y Y The Problem • The cost of an alignment is the number of columns in which the letters differ. S S − U N O W N N − Cost: 3 Y Y − S S U N O W N − − Cost: 5 − N Y Y The Problem • The cost of an alignment is the number of columns in which the letters differ. • And the edit distance between two strings is the cost of their best possible alignment. S S − U N O W N N − Cost: 3 Y Y − S S U N O W N − − Cost: 5 − N Y Y The Problem • The cost of an alignment is the number of columns in which the letters differ. • And the edit distance between two strings is the cost of their best possible alignment. • Edit distance is so named because it can also be thought of as the minimum number of edits S S − U N O W N N − Cost: 3 Y Y − S S U N O W N − − Cost: 5 − N Y Y The Problem • The cost of an alignment is the number of columns in which the letters differ. • And the edit distance between two strings is the cost of their best possible alignment. • Edit distance is so named because it can also be thought of as the minimum number of edits • insertions, deletions, and substitutions of characters needed to transform the first string into the second. S S − U N O W N N − Cost: 3 Y Y − S S U N O W N − − Cost: 5 − N Y Y A Dynamic Programming Solution • When solving a problem by dynamic programming, the most crucial question is, A Dynamic Programming Solution • When solving a problem by dynamic programming, the most crucial question is, What are the subproblems? A Dynamic Programming Solution • When solving a problem by dynamic programming, the most crucial question is, What are the subproblems? • Our goal is to find the edit distance between two strings x[1, . . . , m] and y[1, . . . , n] . A Dynamic Programming Solution • When solving a problem by dynamic programming, the most crucial question is, What are the subproblems? • Our goal is to find the edit distance between two strings x[1, . . . , m] and y[1, . . . , n] • For every i, j with 1 ≤ i ≤ m and 1 ≤ j ≤ n, let . A Dynamic Programming Solution • When solving a problem by dynamic programming, the most crucial question is, What are the subproblems? • Our goal is to find the edit distance between two strings x[1, . . . , m] and y[1, . . . , n] • For every i, j with 1 ≤ i ≤ m and 1 ≤ j ≤ n, let • E(i, j): the edit distance between some prefix of the first string, x[1, . . . , i], and some prefix of the second, y[1, . . . , j]. . A Dynamic Programming Solution • When solving a problem by dynamic programming, the most crucial question is, What are the subproblems? • Our goal is to find the edit distance between two strings x[1, . . . , m] and y[1, . . . , n] • For every i, j with 1 ≤ i ≤ m and 1 ≤ j ≤ n, let • E(i, j): the edit distance between some prefix of the first string, x[1, . . . , i], and some prefix of the second, y[1, . . . , j]. • E(i, j) = min{1 + E(i − 1, j), 1 + E(i, j − 1), diff(i, j) + E(i − 1, j − 1)}, where diff(i, j) is defined to be 0 if x[i] = y[j] and 1 otherwise. . An Example Edit distance between EXPONENTIAL and POLYNOMIAL, subproblem E(4, 3) corresponds to the prefixes EXPO and POL. The rightmost column of their best alignment must be one of the following: O − or − L O L or Thus, E(4, 3) = min{1 + E(3, 3), 1 + E(4, 2); 1 + E(3, 2)}. (b) (a) j−1 j n i−1 i m GOAL E X P O N E N T I A L 0 1 2 3 4 5 6 7 8 9 10 11 P 1 1 2 2 3 4 5 6 7 8 9 10 O 2 2 2 3 2 3 4 5 6 7 8 9 L 3 3 3 3 3 3 4 5 6 7 8 8 Y 4 4 4 4 4 4 4 5 6 7 8 9 N 5 5 5 5 5 4 5 4 5 6 7 8 O 6 6 6 6 5 5 5 5 5 6 7 8 M 7 7 7 7 6 6 6 6 6 6 7 8 I 8 8 8 8 7 7 7 7 7 6 7 8 A 9 9 9 9 8 8 8 8 8 7 6 7 L 10 10 10 10 9 9 9 9 9 8 7 6 Chapter by Chapter Chapter 0 Prologue Big-O Notation • Multiplicative constants can be omitted: 14n2 becomes n2 . • na dominates nb if a > b, for instance, n2 dominates n • an dominates bn if a > b, for instance, 3n dominates 2n • Any exponential dominates any polynomial: 3n dominates n5 • Any polynomial dominates any logarithm: n dominates (log n)3 . This also means, for example, that n2 dominates n log n. An Example 0.1. In each of the following situations, indicate whether f = O(g), or f = Ω(g), or both (in which case f = Θ(g)). f (n) g(n) (a) n − 100 n − 200 (b) n1/2 n2/3 (c) 100n + log n n + (log n)2 (d) n log n 10n log 10n (e) log 2n log 3n (f) 10 log n log(n2 ) (g) n1.01 n log2 n (h) n2 / log n n(log n)2 (i) An Example (i) (j) (k) (l) (m) (n) (o) (p) (q) n0.1 log n (log √ n) n n1/2 n2n 2n n! log n (log Pn n) k i i=1 (log n)10 n/ log n (log n)3 5log2 n 3n 2n+1 2n 2 2(log2 n) nk+1 Chapter 1 Algorithms with Numbers Recursion Analysis MODEXP(x, y, N); Two n-bit integers x and N, and an integer exponent y; if y = 0 then return 1; z=MODEXP (x, by/2c, N); if y is even then return z2 mod N; else return x · z2 (mod N); end Another formulation: ( (x by/2c )2 (mod N) if y is even x y (mod N) = by/2c 2 x · (x ) (mod N) if y is odd Arithmetic Algorithm What is the GCD of 42 and 96? Can you prove it? Chapter 2. Divide-and-Conquer An Example Foo(x); if x > 1 then Printf(“Hello World”); Foo(x/4); Foo(x/4); Foo(x/4); end Chapter 3. Decompositions of Graphs Our Tools and Weapons Our Tools and Weapons • DFS (Explore) Our Tools and Weapons • DFS (Explore) • DAG, Linearization Our Tools and Weapons • DFS (Explore) • DAG, Linearization • SCC Exercise 1 3.16. Suppose a CS curriculum consists of n courses, all of them mandatory. The prerequisite graph G has a node for each course, and an edge from course v to course w if and only if v is a prerequisite for w. Find an algorithm that works directly with this graph representation, and computes the minimum number of semesters necessary to complete the curriculum (assume that a student can take any number of courses in one semester). The running time of your algorithm should be linear. Exercise 2 3.22. Give an efficient algorithm which takes as input a directed graph G = (V, E), and determines whether or not there is a vertex s ∈ V from which all other vertices are reachable. Exercise 3 A bipartite graph is a graph G = (V , E) whose vertices can be partitioned into two sets (V = V1 ∪ V2 and V1 ∩ V2 = ∅) such that there are no edges between vertices in the same set (for instance, if u, v ∈ V1 , then there is no edge between u and v). (a) Give a linear-time algorithm to determine whether an undirected graph is bipartite. (b) Prove that an undirected graph is bipartite if and only if it contains no cycles of odd length. (c) At most how many colors are needed to color in an undirected graph with exactly one odd length cycle? Exercise 4 Prove that given a connected undirected graph G = (V , E), there exists a node v ∈ V , such that G is still connected after deleting v. Exercise 5 Given a directed graph, work out an efficient algorithm, to find all cross edges in DFS searching. Chapter 4. Path in Graphs Our Tools and Weapons Our Tools and Weapons • BFS Our Tools and Weapons • BFS • Shortest Path Our Tools and Weapons • BFS • Shortest Path • Shortest Path in DAG Exercise 1 4.14. You are given a strongly connected directed graph G = (V, E) with positive edge weights along with a particular node v0 ∈ V . Give an efficient algorithm for finding shortest paths between all pairs of nodes, with the one restriction that these paths must all pass through v0 . Exercise 2 Give an efficient algorithm that takes as input a directed acyclic graph G = (V , E), and two vertices s, t ∈ V , and outputs the number of different directed paths from s to t in G. Exercise 3 Give an efficient algorithm to solve: Input Given an undirected graph G = (V , E),every edge is assigned a weight le > 0, and given a specific edge e ∈ E. Output the length of that shortest circle that contains e. Exercise 4 Given an undirected graph with n nodes and the degree of each node is ≥ 1, there must exists a circle, and the number of nodes with degree ≥ 3 is less than 2 · dlog2 ne Chapter 5. Greedy Algorithms Our Tools and Weapons Our Tools and Weapons • MST Our Tools and Weapons • MST • Cut Theorem Chapter 6. Dynamic Programming Exercise 1 6.2. You are going on a long trip. You start on the road at mile post 0. Along the way there are n hotels, at mile posts a1 < a2 < · · · < an , where each ai is measured from the starting point. The only places you are allowed to stop are at these hotels, but you can choose which of the hotels you stop at. You must stop at the final hotel (at distance an ), which is your destination. You’d ideally like to travel 200 miles a day, but this may not be possible (depending on the spacing of the hotels). If you travel x miles during a day, the penalty for that day is (200 − x)2 . You want to plan your trip so as to minimize the total penalty—that is, the sum, over all travel days, of the daily penalties. Give an efficient algorithm that determines the optimal sequence of hotels at which to stop. Exercise 2 Given two strings x = x1 x2 . . . xn and y = y1 y2 . . . ym , we wish to find the length of their longest common subsequence. Show how to do this in time O(mn). Chapter 7. Linear Programming An Example max x + 2y − 1 2x + y ≤ 3 x + 3y ≤ 5 x, y ≥ 0