Professor: Armin Jamshidpey | Term: Winter 2025
Lecture 1
Syllabus
- Divide-and-conquer, master theorem
- Breadth-first and depth-first search
- Greedy algorithms
- Dynamic Programming
- NP-completeness
Cost of Algorithms
Inputs are parameterized by an integer , called the size.
Runtime of an instance: runtime on input
Worst-case Runtime:
Average Runtime:
We will sometimes use more than one parameter (i.e. number of rows and columns in a matrix)
Asymptotic Notation - CS240 Review
Consider two functions with values in
We say that if
We say that if
This is equivalent to
We say that if
This is equivalent to and
In particular, true if for some
We say that if
This is equivalent to
We say that if
This is equivalent to
This is also equivalent to
Example
means (at most) polynomial in
is in and
Question
?
Yes,
Question
?
.
Thus,
Consider with values in
Less strict definition: such that for and
Computational model: word RAM
Rough definition:
- Memory locations contain integer words of bits each
- Assume for input size
- Random Access Memory: Can access any memory location at unit cost
- Basic operations on words have unit costs
Question
What is the cost of the following?
Sum(A[1...n])
s = 0
for i = 1, ..., n
s ← s + A[i]
Question
What is the cost of the following?
Product(A[1..n])
s = 1
for i = 1, ..., n
s = s x A[i]
- Big- is only an upper bound
Typical example: is in and is in . Try to give ‘s if possible
- Big-anything hides constants
A will beat a algorithm eventually
- We use a simplified model
We focus on operations, forget memory requirements.
Lecture 2
Question
Given an array , find a contiguous subarray that maximizes the sum
Example
Given , the subarray is the maximal sum.
Brute force algorithm
BruteForce(A)
opt = 0
for i = 0 to n do
for j = i to n do
sum = 0
for k = i to j do
sum = sum + A[k]
if sum > opt
opt = sum
return opt
Runtime:
We can improve the brute force algorithm (since we are recomputing the same sum many times)
BetterBruteForce(A)
opt = 0
for i = 1 to n do
sum = 0
for j = 1 to n do
sum = sum + A[j]
if sum > opt
opt = sum
return opt
Runtime:
Divide-and-conquer: solve the problem twice in size (we assume is a power of ).
The optimal subarray
- Is completely in the left half
- Or is completely in the right half
- Or contains both and
To find the optimal subarray in case , write
More abstractly: for in and in
To maximize , maximize and independently.
Maximize half-sums
MaximizeLowerHalf(A)
opt = A[n/2]
sum = A[n/2]
for i = n/2 -1, ..., 1 do
sum = sum + A[i]
if sum > opt
opt = sum
return opt
operations
MaximizeUpperHalf(A)
opt = A[n/2 + 1]
sum = A[n/2 + 1]
for i = n/2 + 2, ..., n do
sum = sum + A[i]
if sum > opt
opt = sum
return opt
Same thing as above, just for the right half of the array.
Main algorithm
DivideAndConquer(A[1...n])
if n = 1 return max(A[1], 0)
opt_lo = DivideAndConquer(A[1..n/2])
opt_hi = DivideAndConquer(A[n/2+ 1 .. n])
opt_middle = MaximizeLowerHalf(A) + MaximizeUpperHalf(A)
return max(opt_lo, opt_hi, opt_middle)
Runtime: so
This is the same as MergeSort.
Dynamic Programming
Idea: Solve the problem in subarrays of sizes
- Is either a subarray of
- Or contains
Translation: write sum for subarrays of . Then
with sum for subarrays of , that include .
Each cell of this array keeps the answer of its cell.
Question
How can we compute ?
Idea: As before, the optimal subarray that contains
- is of the form , for some
- or is exactly
should be simple. Must be .
Translation:
Can eliminate recursive calls, and write as a loop.
M = A[1]
for i = 2, ... n do
M = A[i] + max(M, 0)
DP(A)
Mbar = A[1]
M = max(Mbar, 0)
for i = 2, ..., n do
Mbar = A[i] + max(Mbar, 0)
M = max(M, Mbar)
return M
Runtime:
Design Idea for MergeSort
Input: Array of integers
- Split into two subarrays: consists of the first elements in and consists of the last elements in
- Recursively run MergeSort on and
- After and have been sorted, use a function Merge to merge them into a single sorted array.
The following is a sloppy recurrence.
We can make it simpler.
Exact and sloppy are identical when is a power of . We solve the sloppy recurrence when using the recursion three method.
Tree Method:
- Start with a one-node tree, , which receives the value
- Grow two children of . These children, and receive the value of , and the value of is updated to be .
- Repeat this process recursively, terminating when a node receives the value
- Sum the values on each level of the tree, and then compute the sum of all these sums; the result is .
Note, this is not a proof. If a proof is required, it is necessary to use induction.
The Master Theorem provides a formula for the solution of many recurrence relations typically encountered in the analysis of divide-and-conquer algorithms.
Theorem
Suppose that and . Consider the recurrence in sloppy or exact form. Denote . Then
Suppose that are integers and
Lecture 3
Computing using the Master Theorem
and
Let
Our total is
We know that
Proof:
Case 1: (e.g. ): Balanced
Total:
Case 2: (e.g. : Heavy Top
Total:
Case 3: (e.g. ): Heavy Leaves
Total:
The exist other methods - the substitution method
To solve a recurrence, do the following:
- Guess the solution (the form of the solution)
- Use induction to prove it
Example
Want to show that
So, we are trying to show that for where , .
Induction Hypothesis: Assume the statement holds for
. Not a good base case.
Let and
A general algorithmic paradigm
- Divide: Split a problem into several subproblems
- Conquer: Solve the subproblems (recursively) applying the same algorithm
- Combine: Use subproblem results to derive final result
Question
When do we use Divide and Conquer?
- Original problem is easily decomposable into subproblems
- Combining solutions is not too costly
- Subproblems are not overly unbalanced
Counting Inversions: Given an unsorted array , find the number of inversions in it
Definition
is an inversion if and
Example
, we get
We show the indices were inversions occur.
Easy algorithm (two nested loops) in
Idea:
- = number of inversions in
- = number of inversions in
- = number of transverse inversions with and
- Return
From previous array, .
and are done recursively.
Question
What about ?
Goal: How many pairs with ?
This number does not change if both sides are sorted.
Example: Starting from , we get
Number of ‘s greater than + number of ‘s greater than + number of ‘s greater than + number of ‘s greater than .
Option 1:
Take each and binary-search its position in the right-hand side
- This is per , so total
- Plus another for sorting left and right
- Recurrence:
- Gives
Sketchy Proof:
Lecture 4
// Copy A into a new array S; c=0
i = 1; j = n/2 + 1;
for (k = 1; k <= n; k++) {
if (i > n/2) A[k] = S[j++]
else if (j > n) A[k] = S[i++]; c = c + n/2
else if (S[i] < S[j]) A[k] = S[i++]; c = c + j - (n/2 + 1)
else A[k] = S[j++]
}
Multiplying polynomials.
Goal: Given and , compute
for i = 0, ..., n-1 do
for j = 0, ..., n-2 do
h_{i+j} = h_{i+j} + f_{i}g_{j}
This is slow. How can we break up a polynomial into two.
can be turned into .
Example
Divide-and-conquer idea
, . Then
Analysis:
- 4 recursive calls in size
- : additions to compute and etc.
Recurrence:
- so
Not better than naive algorithm.
Karatsuba’s algorithm
Idea: Use the identity
Analysis:
- 3 recursive calls in size
- additions to compute and
- Multiplications by and are free
- additions and subtractions to combine the results
Recurrence:
- so
Remark: Key idea a formula for degree-1 polynomials that does 3 multiplications
Toom-Cook:
FFT:
Multiplying Matrices:
Goal: Given and compute
Input and output size , easy algorithm in
Divide and Conquer: Divide the matrices into 4. More multiplications, but the size is smaller.
Naively: 8 recursive calls in size additions
This didn’t get better.
Strassen’s Algorithm: 7 multiplications and we can construct the entries of the matrix.
Analysis: 7 recursive calls in size additions . The constant hiding is large. Only makes sense to use on galactic algorithms.
Generalization: An algorithm that does multiplications for matrices of size gives
Closest Pairs
Goal: Given points () in the plane, find a pair that minimizes the distance
Equivalent to minimize
All ‘s are pairwise distinct.
Idea: Separate the points into two halves at the median value (can use quick select to find this in )
- all points with
- all points with
- The closest pair is either between points in , or between points in , or transverse.
Set
We only need to consider transverse pairs with and .
Lecture 5
For the transverse case, we only need to search from to . Assume the points are ordered by the coordinate and only find distance upwards (don’t look downwards).
For any , enough to look at points with . It is enough to check distances for in the rectangle.
Claim
There are at most 8 points from our initial set (including ) in the rectangle.
Initialization: Sort the points, with respect to
Cost: , before recursive calls.
Note: Merge based on the coordinate so that the result is sorted in coordinate.
Then: Recursion
- Finding median is easy:
- For the next recursive calls, split the sorted lists:
- Remove the points at distance from the splitting line:
- Inspect all remaining points in increasing order. For each, compute the distance to the next points and keep the min:
We are splitting by value, and then merging by value.
Runtime: so
Median of medians
Median: Given , find the entry that would be at index if was sorted
Selection: Given and in , find the entry that would be at index if was sorted.
Known results: Sorting in , or a simple randomized algorithm in expected time
Question
How do we find a good pivot for quick select such that both and are not too large.
Sketch of the algorithm:
- Divide into groups, of size .
- Find the medians of each group:
- Pivot is the median of :
With this choice of , the indices and are at most .
Proof:
- Half of the ‘s are greater than ()
- For each , there are 3 elements in greater than or equal to
- So at least elements greater than
- So at most elements less than
- So is at most . Same thing for
The runtime satisfies
This gives
We show that for a large enough and all
We know that if is large enough then for all
Also choose a constant to write as
Assume the statement is correct for all . We have
Show that is
We will have
, if
Now if , then we have and we can choose
Note:
Lecture 6
- Basics on undirected graphs
- Undirected BFS and applications (shortest paths, bipartite graphs, connected components)
- Undirected DFS and applications
- Basics on directed algorithms
- Directed DFS and applications
Definition
Undirected Graphs: A graph is pair :
- is a finite set, whose elements are called vertices
- is a finite set, whose elements are unordered pairs of distinct vertices, and are called edges
Convention: is the number of vertices, is the number of edges.
Data structures:
- Adjacency list: An array s.t. is the linked list of all edges connected to . list cells (an edge is in 2 lists), total size , but testing if an edge exists is not
- Adjacency matrix: A matrix of size , with is an edge. Size is , but testing if an edge exists is
Adjacency list of above:
Array with corresponding linked list of edges.
Adjacency Matrix
Definition
Path: A sequence of vertices, with in for all . is okay
Definition
Connected graph: such that for all in , there is a path
Definition
Cycle: A path , with and ‘s pairwise distinct
Definition
Tree: A connected graph without any cycle
Definition
Rooted tree: A tree with a special vertex called root
Definition
Subgraph of : A graph , where
- , with all edges joining vertices from
Definition
Connected component of
- A connected subgraph of
- That is not contained in a larger connected subgraph of .
Let be the connected components of .
- The ‘s are a partition of , with .
- The ‘s are a partition of , with
Question
Assume we are looking for a person in a social network and we don’t want to use the usual search. What is a possible strategy?
BFS(G, s):
G: a graph with n vertices, given be adjacency list
s: a vertex from G
let Q be an empty queue
let visited be an array of size n, with all entries set to false
enqueue(s, Q) // add s to Q
visited[s] = true
while Q not empty do: // s is inside Q
v = dequeue(Q)
for all w neighbours of v do:
if visited[w] is false:
enqueue(w, Q)
visited[w] = true
A vertex is enqueued at most once. Nodes are dequeued at most once. Each adjacency list is read at most once.
For all , write number of neighbours of length of degree of .
Then total cost at step 7 is
The adjacency array has 2 cells (handshaking lemma from MATH239).
Total:
Question
For all vertices , there is a path in if and only if is true at the end
Claim (first)
For all vertices , if is true at the end, there is a path in .
Proof: Let be the vertices for which is set to true, in this order. We prove: for all , there is a path , by induction.
This is OK for .
Suppose it is true for .
Since is set to true, we are examining the neighbours of a certain .
By assumption, there is a path because is in , there is a path
Claim (second)
For all vertices , if there is a path in , is true at the end.
Proof: Let be a path . We prove is true for all , by induction.
is true.
If is true, we will examine all neighbours of so at the end of Step 7, all will be true, for neighbour of in particular, will be true.
Lemma: For all vertices , there is a path in if and only if is true at the end.
Applications:
- Testing if there is a path
- Testing if is connected in
For a connected graph, . This can help us simplify our runtime.
BFS(G, s)
Q is an empty queue
parent is an array of size n, all entries set to NIL
level is an array of size n, with all entries to infinity
enqueue(s, Q)
parent[s] = s
level[s] = 0
while Q not empty do:
v = dequeue(Q)
for all w neighbours of v do:
if parent[w] is NIL:
enqueue(w, Q)
parent[w] = v
level[w] = level[v] + 1
Since was added first, we dequeue first and update its neighbours. Then once that happens we dequeue .
Definition
The BFS tree is the subgraph made of:
- all such that NIL
- all edges , for as above (except )
Claim
The BFS tree is a tree
Proof: By induction on the vertices for which is not NIL.
When we set , only one vertex, no edge.
Suppose true before we set . was in before, was not, so we add one vertex and one edge to so remains a tree
Remark: We make it a rooted tree by choosing as a root.
Lecture 7
Shortest path from the BFS tree.
Sub-claim 1
The levels in the queue are non-decreasing
Sub-claim 2
For all vertices , if there is an edge , then
Proof:
If we dequeue before ,
If we dequeue before , the parent of is either , or was dequeued before . In any case, , but , so OK.
Claim
For all in :
- There is a path in there is a path in
- If so, the path in is a shortest path and
Proof of second item:
(follow the path on ). is the shortest path.
For all , for all , if there is a path of length , then
- True for
- Suppose true for and take , a path of length and let be the vertex before . Induction assumption: , so
So
Since
Definition
A graph is bipartite if there is a partition such that all edges in and one end in
Claim
Suppose connected, run BFS from any , and set
- vertices with odd level
- vertices with even level
Then is bipartite if and only if all edges have one end in and one end in (testable in )
Proof: obvious
For , let be a bipartition. Because paths alternate between W_2$:
- ( vertices with odd level) is included in
- ( vertices with even level) is included in
So and
Computing the connected components
Idea: add an outer loop that runs BFS on successive vertices
Each pass of BFS , if is the th connected component.
Total
Going depth-first
- Travel as deep as possible, as long as you can
- When you can’t go further, backtrack
DFS implementations are based on stacks, either implicitly (recursion) or explicitly (as with queues for BFS).
DFS(G):
// G: a graph with n vertices, given by adjacency lists
// Let visited be an array of size n, with all entries set to false
for all v in G
if visited[v] is false
explore(v)
explore(v):
visited[v] = true
for all w neighbour of v do
if visited[w] = false
explore (w)
Claim
”White path lemma” - When we start exploring a vertex , any that can be connected to by an unvisited path will be visited during
Proof: Let be a path , with not visited. We prove all ‘s are visited before $explore(v) is finished.
True for . Suppose true for . When we visit , is not finished, and is one of its neighbours.
- If is true when we reach step 3 (OK, means we visited it)
- Else: We will visit it just now (OK, will be done before is finished)
Claim
If is visited during , there is a path
Proof: Same as BFS
Previous properties: After we call explore at in DFS, we have visited exactly the connected components containing
Shortest paths: No
Runtime: Still
Previous observation:
- gives a partition of into vertex-disjoint rooted trees (DFS forest)
Definition
Suppose the DFS forests is and let be two vertices.
- is an ancestor of if they are on the same and is on the path root
- equivalent: is a descendant of
Claim
All edges in connect a vertex to one of its descendants or ancestors (statement does not hold for BFS)
Proof: Let be an edge, and suppose we visit first.
Then we visit is an unvisited path between and , so will become a descendant of (white path lemma)
Definition
A back edge is an edge in connecting an ancestor to a descendant, which is not a tree edge.
Observation: All edges are either tree edges or back edges.
Set a variable to 1 initially, create two arrays start and finish, and change explore:
explore(v)
visited[v] = true
start[v] = t
t++
for all w neighbour of v do
if visited[w] = false
explore(w)
finish[v] = t
t++
Intervals are either completely inside each other or completely disjoint.
Lecture 8
Observation:
if , then either or .
Proof:
Let . If , we are done.
Assume then we have:
Cut vertices
Definition
For connected, a vertex in is a cut vertex if removing (and all edges that contain it) makes disconnected.
Also called articulation points.
Find the cut vertices ( connected)
Setup: We start from a rooted DFS tree , knowing parent and level.
Warm-up: The root is a cut vertex if and only if it has more than one child
Proof:
If has one child, removing leaves connected. So is not a cut vertex.
Suppose has subtrees .
Key property: No edge connecting to for . So removing creates connected components.
Definition
For a vertex , let
is the lowest level of an edge to .
is the lowest level of any edge from the subtree rooted at
For any (except the root), is a cut vertex if and only if it has a child with . Intuitively: We can never go above from any child of .
Proof:
Take a child of , let be the subtree at . Let also be the subtree at .
If , then there is an edge from to a vertex above . After removing , remains connected to the root. Number of connected components remains the same.
If , then all edges originating from end in .
Proof: any edge originating from a vertex in ends at a level at least , and connects to one of its ancestors or descendants (key property).
If has children , then
Conclusion:
- Computing is where degree of
- Knowing all , we get in
- So all values can be computed in (remember when connected
- Testing the cut-vertex condition at is
- Testing all is
Directed Graphs
Definition
as in the the undirected case, with the difference that edges are (directed) pairs
- edges also called arcs
- is the source node, is the target A path is a sequence of vertices, with in for all . is OK. A cycle is a path A directed acyclic graph (DAG) is a directed graph with no cycle.
Definition
The in-degree of is the number of edges of the form
Info
The out-degree of is the number of edges of the form
Data structures
- Adjacency lists
- Adjacency matrix (not symmetric anymore)
The DFS and BFS algorithms work without any change. We will focus on DFS.
Still true:
- We obtain a partition of into vertex-disjoint tees
- When we start exploring a vertex , any with an unvisited path becomes a descendant of (white path lemma)
- Properties of start and finish times
But there can exist edges connecting the trees .
Suppose we have a DFA forest. Edges of are one of the following:
- Tree edges
- Back edges: from descendant to ancestor
- Forward edges: from ancestor to descendant (but not tree edge)
- Cross edges: all others
explore(v):
visited[v] = true
start[v] = t, t++
for all w neighbour of v do
if visited[w] = false
explore(w)
finish[v] = t, t++
For a given edge we have four cases for start and finish times while we explore :
- starts and finishes, and then starts and finishes. This is not possible when we explore is unvisited and by white path lemma will be visited before the exploration of is over.
- If we visit while we are checking neighbours of and this happens directly from , then becomes the parent of , and is a tree edge. Otherwise we have visited through another neighbour and this means that is a descendant of . So, is a forward edge.
- (same drawing as above but different and are swapped) was visited but not finished. This implies that we are visiting , through a neighbour of and this means that becomes a descendant of . Hence is a back edge.
- In this case was visited and finished before we start . In this case or are not descendants of each other and belong to two different subtrees. Hence is a cross edge.
So overall, if was visited:
- If not finished, back edge
- else if , forward edge
- else , cross edge
Lecture 9
Claim
has a cycle if and only if there is a back edge in the DFS forest
First direction. If has a back edge, then there has a cycle.
Suppose there is a back edge . Then is a descendant of , so there is a path , and a cycle
Other direction: If has a cycle, then it has a back edge.
Suppose there is a cycle . WLOG, we find first in the DFS.
Starting from , we will reach (white path lemma). We check the edge , and is not finished. So back edge.
Acyclicity test in .
Topological ordering
Definition
Suppose is a DAG. A topological order is an ordering of such that for any edge , we have .
A DAG will always have an in-degree 0 node.
The start time from a DFS forest is not helpful. The finish times are helpful, but we have to reverse their order.
Claim
Suppose that is ordered using the reverse of the finishing order: . This is a topological order.
Proof: Have to prove: For any edge .
If we discover before , will become a descendant of (white path lemma), and we finish exploring it before we finish .
If we discover before , because there is no path ( is a DAG), we will finish before we start .
Consequence: Topological order in .
Definition
A directed graph is strongly connected if for all in , there is a path (and thus a path ).
Observation: is strongly connected there exists such that for all , there are paths and .
Proof:
- is obvious
- For , take vertices . We have paths and , so . Same thing with .
Testing strong connectivity:
- We call explore twice, starting from a same vertex .
- Edges reversed the second time.
Correctness:
- First run tells whether for all , there is a path
- Second one tells whether for all , there is a path in the reverse graph (which is a path in ).
Consequence: Test in
Definition
A strongly connected component of is
- A subgraph of
- Which is strongly connected
- But not contained in a larger connected subgraph of .
A directed graph can be seen as a DAG of disjoint strongly connected components.
Kosaraju’s algorithm for strongly connected components
Definition
For a directed graph , the reverse (or transverse) graph is the graph with same vertices, and reversed edges.
SCC(G)
run DFS on G and record finish times
run DFS on G^T, with vertices ordered in decreasing finish time
return the trees in the DFS forest of G^T
Complexity: .
We want to prove: and belong to the same SCC of and belong to the same tree in DFS forest of .
Proof of () (Order of vertices does not matter)
Assume is the SCC containing and .
When we run DFS on , there is a first node of which we visit. Call it .
All vertices on this path are in .
Otherwise we could enlarge . This is a white path from to . So, by white path lemma, becomes a descendant of . So, and belong to the same tree in DFS forest of . With a similar argument for , we get belong to the same tree.
Proof of ( )
Show that for eery vertex (including and ), and are in the same connected component of .
Since there is a path from to in (which is in ), there exists a path from in .
Now it is enough to show that is a descendant of in the DFS forest of . Which means that there is a path from to .
Inductive argument: Assume the statement is true for and show that it is true for children of .
Let be a child of .
- is a descendant of , so, ( starts after starts and finishes before finishes)
- We visited before here and based on our ordering in the first run of DFS we had
Two cases:
- (Good) We start after we start , and we finish before we finish .
- (Bad) The finish time of is before the start time of . starts after starts and finishes before finishes.
Bad case is not possible:
, so . By the white path lemma we visit before the exploration of is over (since is a white path). Hence we have ( starts after starts and finishes before ends. starts after ends. starts after starts and finishes before ends).
Lecture 10
a directed graph with a weight function:
The weight of path is:
Question
Do shortest path exists in any directed weighted graph?
No, we have negative-weight cycles, can keep looping.
Assuming has no negative-weight cycles, then there does exist a shortest path.
The shortest path weight from to :
Single-Source Shortest Path Problem
Input: and a source
Output: A shortest path from to each
True/False: If is a shortest path from to , then is a shortest path from to for any .
This is true.
Dijkstra’s algorithm: Explanation.
Dijkstra’s algorithm is a greedy algorithm.
Input: A weighted directed graph with non-negative wedge weights.
For all vertices, maintain quantities:
- : a shortest-path estimate from to
- : predecessor in the path (a vertex or NIL)
Initialize , repeat the following until :
- Add with smallest -value to
- Update -values of vertices with with :
- Relaxation:
- Update if is changed
Question
Which Abstract Data Type (ADT) should we use for vertices?
We will use a priority queue.
Cost of operations of a binary min-heap (of size ):
- Insert:
- Extract-Min:
- Update-key:
Dijkstra’s Algorithm: Pseudocode
Dijkstra(G, w, s)
for each vertex v \in V[G]
d[v] = infty
pi[v] = NIL
d[s] = 0
C = \emptyset
Q = V[G]
while Q != \emptyset:
u = Extract-Min(Q)
C = C union {u}
for each vertex v \in Adj[u]
if d[v] > d[u] + w(u,v)
d[v] = d[u] + w(u,v)
pi[v] = u
Complexity Analysis:
Entirely array implementation
-
1-5:
-
6:
-
7: iterations
-
8-9:
-
10: iterations
-
11-13:
-
Total:
-
1-5: Array Imp: . Heap Imp:
-
6: Array Imp: . Heap Imp:
-
7: Array Imp: iterations. Heap Imp: iterations
-
8-9: Array Imp: . Heap Imp:
-
10: Array Imp: iterations. Heap Imp: iterations
-
11-13: Array Imp: . Heap Imp:
-
Total: Array Imp: . Heap Imp:
Performance depends on density/sparsity of graph.
Proof of Correctness:
Claim
For each vertex , we have at the time when is added to set . To prove the claim, we follow these steps.
- Proof by contradiction: Assume the claim is not correct and is the first vertex for which when it is added to
- Time : Beginning of the iteration in which is added to .
- Use a shortest path , from to (needs justification)*
On , find (with predecessor ), the first vertex in .
At the time , we have (needs justification)**
Fact 1:
Fact 2: At time the algorithm chose . Hence
From equations facts 1 and 2, we conclude:
Justification 1: Use a shortest path from to .
Proof:
- , since
- If there is no path from to , then and , so
- There exists a path between and . So there exists a shortest path. Name it .
Justification 2: At time , we have
Proof:
- was added to before . So, .
- When was added to , its edges were relaxed, including
- is on the shortest path from to , so the part of the path from to is a shortest path. Hence:
- For relaxing is compared to which is the minimum value for . So, at the time of relaxation
Lecture 11
Definition
is a connected graph. A spanning tree in is a tree of the form , with a subset of .
In other words: a tree with edges from that covers all vertices. Examples: BFS tree, DFS tree.
Suppose edges have weights . The goal is to find a spanning tree with minimal weight.
Kruskal’s algorithm
GreedyMST(G)
A = []
// sort edges by increasing weight - slow
for k = 1, ..., m do
if e_k does not create a cycle in A then
append e_k to A
Claim
Let be a connected, and let be a subset of the edges of . If has no cycle and , then one can find an edge not in such that still has no cycle.
Proof:
- In any graph, vertices connected components edges
- For , this gives so
- Take any edge on a path that connects two components.
In :
Which means that has at least two connected components which belong to .
Take to be an edge on a path connecting two connected components of in . does not create a cycle.
Claim
If the output is then is a spanning tree (and )
Proof:
Of course has no cycle. Suppose is not a spanning tree. Then, there exists an edge not in , such that still has no cycle.
Case 1: ). Impossible, since is the element with the smallest weight.
Case 2: . Impossible: At the moment we inserted , we decided not to include . This means that created a loop with .
Case 3: . Impossible: We would have included it in , since there is no loop in .
Claim
Let and be two spanning trees, and let be an edge in but not in . Then there exists an edge in but not in such that is still a spanning tree. Bonus: is on the cycle that creates in .
Proof:
Question
If we add to , what happens?
will have a cycle. Let’s say the cycle is in .
splits into two subtrees after removing .
starts in , crosses over to , so it contains another edge between and .
We have that and
is a spanning tree.
Idea: Start with any spanning tree and make it step-by-step closer to while we make sure that at each step we get such that
Correctness: Exchange argument
- Let be the output of the algorithm
- Let be any spanning tree
- If let be an edge in but not in
- So there is an edge in but not in such that is a spanning tree, and is on the cycle that creates in .
- During the algorithm we considered but rejected it, because it created a cycle in
- All other elements in this cycle have smaller (or equal) weight
- So
- So has weight , and one more common element with .
Lecture 12
Disjoint-set data structure. Stores partition of a set into disjoint subsets. We have operations to add sets, merge sets, and finding members of sets. This is used for Kruskal’s algorithm.
Operations on disjoint sets of vertices:
- Find: Identify which set contains a given vertex
- Union: Replace two sets by their union
T = []
U = {{v_1}, ..., {v_n}}
// sort edges by increasing weight
for k = 1,..., m do
if U.Find(e_k.1) != U.Find(e_k.2) then
U.Union(U.Find(e_k.1), U.Find(e_k.2))
append e_k to T
We can use an array of linked lists.
Union process is straightforward. Combine linked list.
For find, we add an array of indices, set that contains .
In the worst case:
- Find is
- Union traverses one of the linked lists, updates corresponding entries of , concatenates two linked lists. Worst case
Kruskal’s algorithm:
- Sorting edges
- Find
- Union
Worse case
We can modify Union
- Each set in keeps track of its size
- Only traverse the smaller list
- Also add a pointer to the trail of the lists to concatenate in
Key observation: Worst case for one union is still , but better total time.
For any given vertex , the size of the set containing at least doubles when we update . So is updated at most times. So the total cost of union per vertex is
Conclusion: for all unions and total.
Greedy Algorithms
We are trying to solve a combinatorial optimization problem:
- Have a large, but finite, domain
- Want to find an element in minimizes/maximizes a cost function
Greedy strategy:
- Build step by step
- Don’t think ahead
- Hard to prove correctness
Review from CS240: Huffman Tree
- Given frequencies for characters
- We build a binary tree for the whole code
Greedy strategy: Build the tree bottom up
- Create many single-letter trees
- Define the frequency of a tree as the sum of the frequencies of the letters in it
- Build the final tree by putting together the smaller trees.
Claim
This minimizes {length of encoding of }
Interval Scheduling Problem
Input: intervals
Output: A maximal subset of disjoint intervals. By disjoint intervals we mean
Example
A car rental company has the following requests for a given day:
- : 2pm to 8pm
- : 3pm to 4pm
- : 5pm to 6pm Answer is
Greedy strategies
- Consider the earliest starting time (Choose the interval with )
- Consider shortest interval (choose the interval with )
- Consider minimum conflicts (choose the interval that overlaps with the minimum number of other intervals)
- Consider earliest finishing time (Choose the interval with )
Only the last one works.
Algorithm: Interval Scheduling
- Sort the intervals such that
- For from 1 to do
- If Interval , has no conflicts with intervals in
- add to
- If Interval , has no conflicts with intervals in
- return
Checking conflicts in , we only need to check the last element in . Thus it is possible in for part 3.
Thus our total is .
Correctness: The Greedy Algorithm Stays Ahead
Assume is an optimal solution. Our goal is to show .
- Suppose are the intervals in in the order they were added to by the greedy algorithm
- Similarly, let the intervals in be denoted by
- Assume that the intervals in are ordered in the order of the start and finish times.
- We prove that
Lemma: For all indices we have that
Proof: We use induction
For the statement is true. The finish time of .
Suppose and the statement is true for . We will show that the statement is true for .
By induction hypothesis we have .
By the order on we have .
Hence we have that
Thus at the time the greedy algorithm chose , the interval was a possible choice.
The greedy algorithm chooses an interval with the smallest finish time. So,
For the sake of contradiction, assume that the output is not optimal, then .
is the last interval in and must have an interval .
Apply the previous lemma with , and we get .
We have .
So, was a possible choice to add to by the greedy algorithm. This is a contradiction.
Lecture 13
Interval Colouring Problem
Input: intervals
Output: Use the minimum number of colours to colour the intervals, so that each interval gets one colour and two overlapping intervals get two different colours.
Algorithm: Interval Colouring
Sort the intervals by starting time:
For from 1 to do:
Use the minimum available colour to colour the interval . (i.e. use the minimum number to colour the interval so that it doesn’t conflict with the colours of the intervals that are already coloured).
Can be done in (and even ).
If the greedy algorithm uses colours, then is the minimum number of colours needed. In other words, no other algorithm can solve the problem with colours.
Idea: Show that there exists a time such that it is contained in intervals.
Assume is the first interval which uses colour . This means that other colours have been used.
Since we have an increasing order on start times:
On the other hand, all intervals have overlap with , hence:
So is the time which is contained in intervals, so no colouring with less than colours.
Minimizing Total Completion Time Problem
Input: jobs, each requiring processing time
Output: An ordering of the jobs such that the total completion time is minimized.
Note: The completion time of a job is defined as the time when it is finished.
Example
, processing times [2, 8, 1, 10, 5]
2 | 8 | 1 | 10 | 5 | |
---|---|---|---|---|---|
2 | 2 | 2 | 2 | 2 | |
8 | 8 | 8 | 8 | ||
1 | 1 | 1 | |||
10 | 10 | ||||
5 | |||||
2 | 10 | 11 | 21 | 26 | 70 |
Total completion time total processing time
Algorithm: Order the jobs in non-decreasing processing times.
1 | 2 | 5 | 8 | 10 | |
---|---|---|---|---|---|
1 | 1 | 1 | 1 | 1 | |
2 | 2 | 2 | 2 | ||
5 | 5 | 5 | |||
8 | 8 | ||||
10 | |||||
1 | 3 | 8 | 16 | 26 | 54 |
Correctness: Exchange Argument
Assume that there is an optimal solution with a different ordering in comparison to the solution by the greedy algorithm.
Let is an optimal solution and is not in a non-decreasing order of processing time. So there exists such that . Sum of the completion times of is .
The contribution of and is .
After switching and , we get another ordering, call it and the contribution of and in is:
Dynamic Programming
Examples: Interval scheduling, longest increasing subsequence, longest common subsequence, all pairs shortest path.
Definition
Fibonacci Numbers
- for
Slow recursive algorithm
Fib(n):
if n == 0 return 0
if n == 1 return 1
return Fib(n-1) + Fib(n-2)
Assuming we count additions at unit cost, runtime is , T(n-1) + T(n-2) + 1$
This give , so ,
flowchart TD
Fn --- Fn-1
Fn-1 --- c[Fn-2]
c --- g[Fn-3]
c --- h[Fn-4]
Fn-1 --- d[Fn-3]
Fn --- Fn-2
Fn-2 --- e[Fn-3]
Fn-2 --- f[Fn-4]
Redundant computation.
To compute , we only need the values of . The algorithm recomputes them many, many times.
Improved recursive algorithm
let T = [0, 1, x, y, ...]
Fib(n)
if T[n] = x
T[n] = Fib(n-1) + F(n-2)
return T[n]
Iterative version
Fib(n)
let T = [0,1, x, y, ...]
for i = 2, ..., n
T[i] = T[i-1] + T[i-2]
return T[n]
These are DAGs.
Iterative version (enhanced)
Fib(n)
(u,v) = (0,1)
for i = 2, ..., n
(u,v) = (v, u+v)
return v
All these improved versions use additions.
Main features: Solve “subproblems” bottom up, and store solutions if needed.
Recipe for designing a DP algorithms
- Identify the subproblem
- Establish DP-recurrence
- Set values for the base cases
- Specify the order of computation
- Recovery of solution (if needed)
Key features:
- Solve problems through recursion
- Use a small (polynomial) number of nested subproblems
- May have to store results for all subproblems
- Can often be turned into one or more loops
Dynamic programming vs divide-and-conquer
- Dynamic programming usually deals with all input sizes
- Divide and conquer may not solve subproblems
Lecture 14
Interval Scheduling Problem
Input: intervals where each interval has weight
Output: A choice of intervals that do not overlap and maximizes . Greedy algorithm in the case .
Example
A car rental company has the following requests for a given day:
Answer is
Basic idea: Either we choose or not.
- Then the optimum is the max of two values:
- , if we choose , where are the intervals that do not overlap with
- , if we don’t choose
In general, we don’t know what look like.
Goal:
- Find a way to ensure that are of the form , for some (and so on for all indices )
- Then it suffices to optimize over all
Sort ‘s by increasing order of end time:
Define for interval to be the largest index less than such that the interval and are disjoint.
Claim: for all , the set of intervals that do not overlap is of the form for some ( if no such interval)
The algorithm will need the ‘s
- If earliest finish time
- If
Let be a permutation of such that
FindPj(A, s_1,...,s_n,f_1,...,f_n)
f_0 = -infty
i = 1
for k = 0, ..., n
while i <= n and f_k <= s_{A[i]} < f_{k+1}
p_{A[i]} = k
i++
Runtime: (sorting) and (loops)
Main procedure:
Definition
is the maximal weight we can get with intervals
Recurrence: and for
Runtime: (sorting twice) and (finding the ‘s)
0/1 Knapsack Problem
Input: Items with weights and values and a capacity .
Output: A choice of items that satisfies the constraint and maximizes the value
Lecture 15
Continuing with the 0/1 Knapsack problem.
Basic Idea: Either we choose item or not
Then the optimum is the max of two values
- , if we choose (and )
- if we don’t choose
maximum value achievable using a knapsack of capacity and items .
Initial conditions: for any and for any .
// 01KnapSack(v_1, ..., v_n, w_1 ..., w_n, W):
// initialize an array O[0..W,0..n]
for i = 1,...,n
for w = 1,...,W
if w_i > w:
O[w,i] = O[w, i-1]
else:
O[w,i] = max(v_i + O[w-w_i, i-1], O[w, i-1])
The runtime of this algorithm is
This is called a pseudo-polynomial algorithm. In our word RAM model, we have been assuming all ‘s and ‘s fit in a word. So input size is words. But the runtime also depends on the values of the inputs.
01-knapsack is NP-complete, so we don’t really except to do much better.
The next problem is the longest increasing subsequence problem.
Input: An array ] of integers.
Output: A longest increasing subsequence of (or just its length) (does not need to be contiguous).
Example
gives
Remark: There are subsequences (including an empty one, which doesn’t count).
Tentative subproblems
Attempt 1:
- Subproblems: The length of a longest increasing subsequence of
- For the example above, . Not enough to deduce
Attempt 2:
- Subproblems: The length of a longest increasing subsequence of , together with its last entry
- Example: , with last element . This is OK if we can add , but what if not?
Attempt 3:
- Let be the length of a longest increasing subsequence of that ends with , for . So
Idea:
A longest increasing subsequence ending at looks like
is a longest increasing subsequence ending at (or it is empty). Don’t know , but we can try all for which .
Iterative Algorithm
// LongestIncreasingSubsequence(A[1..n])
L[1] = 1
for i = 2, ..., n do
L[i] = 1
for j = 1, ..., i-1 do
if A[j] < A[i] then
L[i] = max(L[i], L[j] + 1)
return the maximum entry in L
The runtime is
Remark: The algorithm does not return the sequence itself, but could be modified to do so.
Lecture 16
Longest Common Subsequence Problem
Input: Arrays and of characters.
Output: Maximum length of a common subsequence to and (not necessarily contiguous).
Example
“blurry”, “burger”, longest common subsequence is “burr”
0 | 1 | |||||
---|---|---|---|---|---|---|
0 | 0 | 0 | 0 | 0 | ||
1 | 0 | |||||
0 | * | |||||
0 | ** |
* = length of the longest common subsequence. Thus, the longest common subsequence exists in cell **.
In and 3 cases may happen:
- does not appear in the longest common subsequence. Then the sequence is contained in and
- doesn’t appear in the longest common subsequence. Then the sequence is contained in and
- appears in the sequence, and the rest of the sequence is contained in and
Definition
Let be the longest subsequence between and
- for all
- for all
- is the max of up to three values
- (don’t use )
- (don’t use )
- if
Algorithm computes all using two nested loops, so runtime
1 | 2 | 3 | 4 | 5 | 6 | |
---|---|---|---|---|---|---|
A | b | l | u | r | r | y |
B | b | u | r | g | e | r |
0 | 1 | 2 | 3 | 4 | 5 | 6 | |
---|---|---|---|---|---|---|---|
0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
1 | 0 | 1 | 1 | 1 | 1 | 1 | 1 |
2 | 0 | 1 | 1 | 1 | 1 | 1 | 1 |
3 | 0 | 1 | 2 | 2 | 2 | 2 | 2 |
4 | 0 | 1 | 2 | 3 | 3 | 3 | 3 |
5 | 0 | 1 | 2 | 3 | 3 | 3 | 4 |
6 | 0 | 1 | 2 | 3 | 3 | 3 | 4 |
Italics represent the path taken.
Walk back from the solution at , each diagonal arrow adds a character.
Bellman-Ford Algorithm: Single source shortest path (with negative weights).
- If no negative cycle, computes all distances
- Can detect negative cycles
Definition
For set length of the shortest path with at most edges if no such path,
This sets us up for a dynamic programming approach for shortest paths.
- This gives and for .
- If there is no negative cycle, .
- In any case, for all and for all .
Assume there are no negative cycles.
If , this is obvious. If is a finite number, it means a path from to . So, is either the weight of that path or a shorter path. Hence, .
If , then it is obvious. If is a finite number, it is the weight of a shortest path from to . We claim that has at most edges. If our claim is not true, then has at least edges, which means that one of the nodes appears at least twice in the path. Hence contains a cycle. The cycle cannot be negative by our assumption. Hence, it is a positive cycle which contradicts with being a shortest path. (remove the cycle from , and we get a shortest path).
Recurrence:
Assume . This means that is the length of a shortest path with at most edges which has a finite weight. The path either has exactly edges or at most edges. If it has at most edge, we use . If it has exactly edges, we decompose the path to one last edge and a path with edges to .
So we use .
However, for finding the shortest path we need to consider all edges as the path may pass through any neighbours of . Hence:
// BellmanFord(G,s)
d_0 = [0, \infty, ..., \infty] // (s is the first index)
parent = [s, x, ..., x] // (s is the first index)
for i = 1, ..., n-1 do
for all v in V do
d_1[v] = d_{i-1}[v]
for all (u,v) in E do
if d_{i-1}[u] + w(u,v) < d_i[v] then
d_i[v] = d_{i-1}[u] + w(u,v)
parent[v] = u
Correctness: , so if no negative cycles.
//BellmanFord2.0
d = [0, \infty, ..., \infty] // (s is the first index)
parent = [s, x, ..., x] // (s is the first index)
for i = 1, ..., n-1 do
for all (u,v) in E do
if d[u] + w(u,v) < d[v] then
d[v] = d[u] + w(u,v)
parent[v] = u
Runtime is:
Lecture 17
Plan to show correctness of the improved version. Show that for to after iteration.
If this is true, then after iterations:
Claim
For all , after iteration , we have for all
Proof: By induction
True for , so we suppose true at index and prove true at . At the beginning of the loop, for all , . can only decrease, so this stays true throughout the loop. is replaced by , where . So at the end of iteration , .
The operation is called a relaxation.
Claim
can only decrease through a relaxation, and if and before relaxation, then post relaxation.
Proof: First item is obvious.
Second item: (triangle inequality). So . But also, .
Consequence: If all satisfy , and we apply any number of relaxations, all inequalities stay true.
Claim
For , after iteration , for all .
Proof: We have proved this via the previous two claims.
Summary:
If there is no negative cycle
- At the end, for all
- In particular, for any edge ,
If there is a negative cycle there exists an edge such that .
Proof: There exists a negative cycle: and
Assume for all ‘s we have sum the previous inequality around the cycle:
Since we have (as ), we conclude which is a contradiction.
So, if there is a negative cycle with
- Claim: There must be an edge with
- Else: for all ; sum and derive a contradiction.
Conclusion: For extra , can check the presence of a negative cycle.
Floyd-Warshall Algorithm: No fixed source, computes all distances
Negative weights are OK, but no negative cycles. Simple pseudo-code, but slower than other algorithms. Doing Bellman-Ford from all takes
Subproblems for dynamic programming
- Bellman-Ford uses paths with fixed number of steps
- Floyd-Warshall restricts which vertices can be used
Bellman-Ford considers number of edges.
Floyd-Warshall considers intermediate vertices.
is the length of a shortest path from to with no intermediate vertices.
is the length of a shortest path from to with
- if there is an edge
- otherwise.
Claim
Proof: Either the shortest path does not go through , or it does (and if it does, it’s only once).
DP recurrence for Floyd-Warshall.
Consider a shortest path from to with all intermediate vertices in and call it .
Case 1: is not on . Then the intermediate path vertices on are in so we can use
Case 2: is on . Then we can decompose to and . is a shortest path from to . is a shortest path from to . In both and , all intermediate paths are in . So in both cases we can use
//FloydWarshall(G)
// set up D_0 above
for i = 1, ..., n do
for j = 1, ..., n do
for k = 1, ..., n do
D_i[v_j, v_k] = min(D_{i-1}[v_j, v_k], D_{i-1}[v_j, v_i] + D_{i-1}[v_i, v_k])
Runtime and memory:
Lecture 18
Edit Distance
Input: Arrays and of characters.
Output: Minimum number of add, delete, change operations that turn into .
Example
A = “snowy”, B = “sunny” There are multiple answers: 3C; 1A, 1C, 1D; 2A, 1C, 2D
There are three cases.
Case 1:
We write on top of . The last column will add 1 to the cost if and adds zero if . The edit distance of the remaining columns is based on the edit distance between and
Case 2:
In this case, the edit distance is 1 plus the edit distance between and
Case 3:
In this case, the edit distance is 1 plus the edit distance between and
Generalize: Let be an array of sizes .
ED between and
Definition
Let be the edit distance between and
- for all (add characters)
- for all (delete characters)
- is the min of three values
- (if ) or (otherwise)
- (delete and match with )
- (add and match with )
The algorithm computes all , using two nested loops, so runtime is
Optimal binary search trees
Input:
- Integers
- Probabilities of access with
Output:
- An optimal BST with keys
- Optimal: Minimizes (depth) expected number of tests for a search.
Example
flowchart TD
a(3) --- b(2)
b --- c(1)
a --- d(5)
d --- e(4)
flowchart TD
a(1) --- b(2)
b --- c(3)
c --- d(4)
d --- e(5)
A greedy strategy does not work here.
Assume it does, and we put the key with the highest probability in the root.
1 | 2 | 3 | 4 | 5 | |
---|---|---|---|---|---|
flowchart TD
a(5) --- b(3)
b --- c(2)
b --- d(4)
c --- e(1)
However, there is a better cost.
flowchart TD
a(3) --- b(2)
a --- c(5)
b --- d(1)
c --- e(4)
Define to be the minimal cost for the items to . Take an item, say , between and and put it in the root.
We chose just a between and and the cost will be
Recurrence
An independent set of a graph , is if there are no edges between elements of .
Given a tree as the input, either the root appears in a maximum independent set or not.
Case 1:
Since , none of the children of can be in . So, the remaining part of should come from subtrees rooted at grandchildren of .
Case 2:
In this case all elements of are from subtrees rooted at children of .
Lecture 19
Decision Problem: Given a problem instance , answer a certain question “yes” or “no”.
Problem Instance: Input for the specified problem.
Problem solution: Correct answer (“yes” or “no”) for the specified problem instance. is a yes-instance if the correct answer for the instance is “yes”. is a no-instance if the correct answer for the instance is “no:.
Size of the problem instance: is the number of bits required to specify (or encode) the instance .
The goal is to find a polynomial time algorithm to convert (transform) inputs of a problem to inputs of a problem .
Polynomial Time Reduction: A decision problem is polynomial time reducible to a decision problem , if there is a polynomial time algorithm that transforms any instance of to an instance of such that is a YES-instance of is a YES-instance of .
We write , if such a polynomial time reduction exists. We also write if and .
Assume and are decision problems and we are given:
- An algorithm , which solves in polynomial time,
- A polynomial reduction , which gives
Design a polynomial time algorithm to solve .
Input: An instance of
Output: Whether is a YES-Instance
Use to transform into , which is an instance of the problem
: Return
If the running time of is , then the output of which is has the property: (Since the algorithm has to return ). The total running time: If has runtime then the total running time is
Transitivity of Polynomial time reductions
Lemma: and
There exists a polynomial time algorithm which maps instance of to
There exists a polynomial time algorithm which maps instances of to .
Question
What does (Composition of and ) do?
maps the instances of to in polynomial time. It also respects the property on yes/no instances.
Assume problem is known to be impossible to be solved in polynomial time.
It is true that cannot be solved in polynomial time.
Assume the contrary statement. can be solved in polynomial time with an algorithm . Then we use for solving be reducing it to , which is a contradiction.
Now looking at the notation again: . is at least as hard as , in other words, if we could solve in polynomial time, then we could solve in polynomial time.
The following three problems are equivalent in terms of polynomial-time solvability.
Maximum Clique:
is a clique if, for all .
Input: and an integer
Output: Is there a clique in with at least vertices.
Maximum Independent set (IS):
is an independent set if, for all .
Input: and an integer
Output: Is there an independent set in with at least vertices.
Minimum Vertex Cover (VC):
is a vertex cover if, for all
Input: and an integer
Output: Is there a vortex cover in with at most vertices.
Clique IS
For the Clique problem, we must find a set of size at least of vertices with all edges in between.
For the IS problem, we must find a set of at least vertices with no edges in between.
Idea: Change edges into no edges (Find the complement of ). is the algorithm to construct the complement of input for the clique. Clearly it can be done in polynomial time and one can check that is a clique in if and only if is an independent set in . Hence is a YES-instance for clique if and only if is a YES-instance for IS.
Assume is a graph.
Lemma: is a vertex cover is an independent set in .
() Assume is a vertex cover in . If is not an IS in , then there must exist such that . By definition of a vertex cover, one of or must be in which is a contradiction
( ) Assume is an IS in . If is not a vertex cover in , then there exists an edge such that and . This means that and there is an edge between them, which is a contradiction.
We want to show that VC IS and IS VC
The previous lemma shows that: has a vertex cover of size at worst has an IS of size at least .
So our reduction algorithm maps for CV to for IS.
It runs in polynomial time and maps yes/no instances of VC to yes/no instances for IS respectively. The above results and transitivity shows that Clique IS VC.
More Simple Reductions
Hamiltonian Cycle (HC):
A cycle is a Hamiltonian Cycle if it touches every vertex exactly once.
Input: Undirected graph
Output: Does have a Hamiltonian Cycle.
Hamiltonian Path (HP):
A path is a Hamiltonian Path if it touches every vertex exactly once.
Input: Undirected graph
Output: Does have a Hamiltonian path.
Proposition: HP HC
Proof of HP HC
Given for HP we have to transform it to for HC.
We construct in the following way:
- Add a vertex to :
- Add edges for all
Claim: has a HP has a HC
() Assume is a HP in , with endpoints and . Then is a HC in
() Assume is a HC in . There must be two incident edges on , call them and . Then by removing and from , we get which is a HP in .
Lecture 20
Proof of HC HP
Assume is given for HC. We construct in the following way:
Choose an arbitrary vertex
- Add a duplicate of
- Add vertices and with degree one and edges and
It is a polynomial time computation.
Claim: has a HC has a HP
() Assume there a HC in
() Assume there is a HP in , it should have the form:
The two end points must be and (degree 1 nodes). has two neighbours, and a node called . We also know that is an edge in as is a copy of . So, by removing and adding , we get a H which is in .
Review:
where ‘s are Boolean variables. A literal term is either or . A clause is a disjunction of distinct literals, , where . We say that the clause is of length .
An assignment satisfies a clause if it causes to evaluate to true. A conjunction of a finite set of clauses, is called a formula in CNF CS245.
3-SAT Problem
Input: A CNF-formula in which each clause has at most three literals.
Output: Is there a truth assignment to the variables that satisfies all the clauses.
Theorem
3-SAT IS
A CNF with at most 3 literals per clause is given.
For each clause we form a triangle with vertices labeled as the literals of the clause
Graph with edges
A clause with 2 literals will be an edge joining the literals.
A clause with one literal is easy to deal with, we don’t include them.
To force exactly one choice from each clause, we set to the number of clauses.
We have to make sure that we are not choosing opposite literals in different clauses.
So, we put an edge between any two vertices that correspond to opposite literals.
This construction takes polynomial time.
Claim
Suppose the formula has clauses. Then the formula is satisfiable there is an independent set (IS) of size in the graph.
Note: We only have two types of edges:
- Type 1: Edges in the clauses
- Type 2: Edges between two clauses
Proof:
() Assume the formula is satisfiable. Then we have at least one literal per clause which is set to true. For each clause choose one literal which is true and add its corresponding node to the set .
Since we picked one node per clause . The set is an independent set. Between any two nodes in , there are no type 1 edges (by construction). Also we don’t choose nodes with type 2 edges between them.
() Now assume there is an IS of size in . Any independent set can have at most one node in each clause, since there are edges between all nodes in a clause.
As has only clauses, our IS has exactly one vertex from each clause.
Moreover, because of the second type edges, we do not have and at the same time in our IS.
For each in the IS we set to be true. Otherwise we set to be false.
This assignment satisfies every clause
For a problem , we represent an instance of as a binary string .
Definition
A problem is in NP, if there is a polynomial time verification algorithm such that the input is a yes-instance there is a proof (certificate) which is a binary string of length so that returns yes.
Vertex Cover
- here is an input graph and an integer
- here is a subset of of with : go through all and check if covers the edges and .
3-SAT
- here is a 3-SAT formula
- here is a truth assignment : check whether satisfies all clauses
Clique, IS, HC, HP, Subset-Sum are all in NP.
Question
Are all problems in NP?
No.
Definition
co-NP is the set of decision problems whose no-instances can be certified in polynomial time.
Every polynomial time solvable decision problem is in NP.
Definition
is the set of decision problems that can be solved in polynomial time.
P NP.
The most important open problem in theoretical computer science.
Question
Is P equal to NP?
The name NP comes from Non-deterministic Polynomial time. A non-deterministic machine, has the power to correctly guess a solution.
Informally, we say a problem is NP-complete if it is the hardest problem in NP.
Definition
A problem NP is NP-complete if for all NP
Fact
P NP an NP-complete problem can be solved in polynomial time.
Theorem (Cook-Levin)
3-SAT is NP-complete
Lecture 21
If we can prove 3-SAT , then is NP-complete. For example IS NPC, since 3-SAT IS
To prove that a problem NP is NP-complete, we just need to find a NP-complete problem , and prove that
Circuit-SAT
- Instance: A circuit DAG with labels on the vertices
- Inputs labelled by boolean variables
- Internal vertices labelled by and, or, not
- There is a marked vertex for the output
- Problem: Is there a choice of boolean that makes true?
Plan to prove Cook-Levin Theorem
- Show that Circuit-SAT is NP-complete
- Show that Circuit-SAT 3-SAT
Circuit-SAT is NP-complete (sketch)
Given: instance of NP
Want: proof (certificate) such that true
Verification algorithm can be turned into a circuit with as input. We call Circuit-SAT to find
Example
Problem : IS Instance complete graph with 3 vertices, Certificate : 3 bits
All possible subsets:
Circuit for computes the formula
We have shown that all problems in NP are reducible to Circuit-SAT. In order to be NP-complete, we must also show that this problem is in NP (not hard).
We will now show that Circuit-SAT is reducible to 3SAT.
From logic we know that we have the following
We add a variable () for the output.
gives and