Theory of Computing

Computer Science 461 - Spring 2025

Jump to: CS 461 Homepage, Week 1, Week 2, Week 3, Week 4, Week 5, Week 6, Week 7, Week 8, Week 9, Week 10, Week 11, Week 12, Week 13, Week 14

Week 1 Notes

Day Section Topic
Mon, Jan 13 1.3 Proof techniques
Wed, Jan 15 2.1 - 2.2 Notation & encodings
Fri, Jan 17 3.1 - 3.7 Boolean circuits

Mon, Jan 13

Today we talked about proof techniques, particularly proof by induction. We looked at these two examples:

  1. Prove that if a set AA has nn elements, then AA has 2n2^n subsets.

The second example had to do with the famous Tower’s of Hanoi puzzle( see https://youtu.be/SMleU0oeGLg).

  1. Use induction to prove that it is always possible to move the disks from one peg to another by moving one disk at a time without breaking the rules.

You can also translate many induction arguments into recursive algorithms.

  1. Write a Python function sublists(A) that inputs a list A and returns a list containing every sublist of A.

    We did this in class and we came up with something like this:

     def sublists(A):
         if len(A) == 0:
             return [A]
         last = A[-1]
         B = A[0:-1]
         sublists_without_last = sublists(other_elements)
         sublists_with_last = [s + [last] for s in sublists_without_last]
         return sublists_without_last + sublists_with_last
  2. Suppose you have three stack variables (implemented as Python lists):

    stack1 = [10 - k for k in range(10)]
    stack2 = []
    stack3 = []

    Write a function move_disks(n, start_stack, temp_stack, final_stack) that recursively moves n disks from the start_stack to the final stack.

Wed, Jan 15

Today we reviewed mathematical notation, including some new notation we will be using. We defined alphabets which are sets of symbols we can use to represent things. The most common alphabet in computer science is the binary alphabet Σ={0,1}\Sigma = \{0, 1\}. We use the notation Σ*\Sigma^* to denote the set of all possible finite length strings constructed from Σ\Sigma.

A set SS can be represented or encoded using an alphabet Σ\Sigma if there is a 1-to-1 function E:SΣ*E: S \rightarrow \Sigma^*.

Theorem. Any countable set SS can be encoded with a 1-to-1 function E:S{0,1}*E:S \rightarrow \{0,1\}^*.

We discussed specific encodings such as how to encode the integers \mathbb{Z} and the rationals \mathbb{Q}.

We also observed another fact:

Theorem. If we can encode a set SS using Σ\Sigma, then we can encode tuples in SS using Σ\Sigma.

At the end we considered the set {0,1}={f:f:{0,1}}\{0,1\}^\infty = \{f : f : \mathbb{N}\rightarrow \{0,1\} \} which you can think of as the set of all infinitely long strings of zeros and ones. We finished by proving

Theorem. There is no one-to-one function from {0,1}\{0,1\}^\infty into {0,1}*\{0,1\}^*.

A corollary of this theorem is that {0,1}\{0,1\}^\infty is uncountable.

Fri, Jan 17

We talked about Boolean circuits which are formed by AND, OR, and NOT gates. These can be used to implement any Boolean expression formed by the \wedge (AND), \vee (OR), and ¬\neg (NOT) operators.

XKCD 2497
  1. Write a Boolean expression that takes three Boolean inputs, x,y,zx, y, z and returns true if x,y,zx, y, z are all the same, and false otherwise.

  2. Show that you can construct the function IF x THEN y ELSE z for any Boolean inputs x, y, z using AND, OR, and NOT gates.

    (xy)(¬xz)(x \wedge y) \vee (\neg x \wedge z)

  3. Use mathematical induction (and the previous result) to prove the following:

Theorem. Every function f:{0,1}n{0,1}f:\{0,1\}^n \rightarrow \{0,1\} can be represented by a Boolean circuit.

We say that the Boolean operations (\wedge, \vee, ¬\neg) are a universal set of operations since every Boolean function on {0,1}n\{0,1\}^n can be constructed using those operations.

We finished by talking about how the NAND operation is universal all by itself. Using just NAND gates, you can construct any Boolean function on {0,1}n\{0,1\}^n.

Theorem. Every function f:{0,1}n{0,1}f:\{0,1\}^n \rightarrow \{0,1\} can be represented using NAND operations.

  1. Show that you can use a NAND gate to implement a NOT gate.

  2. Show that you can use two NAND gates to implement an AND gate.

  3. Use a truth table to confirm that (NOT x) NAND (NOT y) is equivalent to x OR y.

Week 2 Notes

Day Section Topic
Mon, Jan 20 MLK day, no class
Wed, Jan 22 Impossible computer programs
Fri, Jan 24 2.1 Intro to finite automata

Wed, Jan 22

Fri, Jan 24

Last time we saw that some functions f:{0,1}*{0,1}f: \{0,1\}^* \rightarrow \{0,1\} cannot be computed by a computer program. We saw two proofs that seemed very different. One used a cardinality argument and the other used proof by contradiction to show that the program check_if_program_returns_one() is impossible. It turns out that both proofs are actually closely related.

Suppose we make a list of all possible programs on the left column of an infinite table, with a list of all possible input strings at the top of the table. The values in the body of the table are the output of each program on each input (including the possibility that there is no output). The table might look like this:

Input 1 Input 2 Input 3 \ldots
Program 1 1 1 1 \ldots
Program 2 0 0 na \ldots
Program 3 0 na 1 \ldots
\vdots \vdots \vdots \vdots \ddots

We can construct a function f:{0,1}*{0,1}f:\{0,1\}^* \rightarrow \{0,1\} that cannot correspond to any program by using the same diagonalization argument that proves that there are uncountably many infinite binary sequences. That’s essentially what we did when we tried to construct the function check_if_program_returns_one() last time. We created a function reverse_check() that corresponds to reversing the values on the main diagonal of the table above. So we were using a diagonalization argument, which is the same argument used to prove that there are uncountably many functions f:{0,1}*{0,1}f:\{0,1\}^* \rightarrow \{0,1\}.

After that, we introduced finite automata. We started with this example: An automatic door at a grocery store has sensors on the front and rear. It opens if the sensor on the front is active. It won’t close until neither of the sensors are active. You can model this with the following state diagram:

This is an example of a finite state machine, also known as a deterministic finite automata (DFA).

Definition. A deterministic finite automata (DFA) consists of

  1. A finite set of states QQ.
  2. A finite alphabet Σ\Sigma of possible input signals.
  3. A transition function δ:Q×ΣQ\delta: Q \times \Sigma \rightarrow Q.
  4. An initial or start state q0Qq_0 \in Q.
  5. A set of final or accepting states FQF \subseteq Q.
  1. What are the sets QQ and Σ\Sigma for the automatic door example above?

  2. Make a table showing the values of the transition function δ\delta for the automatic door.

  3. An automatic toll booth accepts nickles, dimes, and quarters. The gate won’t open until it receives 25 cents. Draw a state diagram for the toll booth. What are the sets QQ and Σ\Sigma?

  4. A combination lock (like the ones at the campus post office) can be modeled as a finite state machine. What are the states and what are the input signals?

  5. Describe a DFA that can compute the function f:{0,1}*{0,1}f: \{0,1\}^* \rightarrow \{0,1\} which returns 1 when the input string has an odd number of 1’s and 0 otherwise.

Week 3 Notes

Day Section Topic
Mon, Jan 27 2.2 - 2.3 Regular languages
Wed, Jan 29 2.4 Nondeterministic finite automata
Fri, Jan 31 2.6 NFAs and regular languages

Mon, Jan 27

Today we started with these questions about DFAs:

  1. For the DFA shown below:
    1. What is the transition function?
    2. Describe the function f:{0,1}*{0,1}f:\{0,1\}^* \rightarrow \{0,1\} that this DFA computes.
  2. Draw the state diagram for a DFA that computes whether a binary string contains 011.

Definition. For any finite alphabet Σ\Sigma, a language is a subset of Σ*\Sigma^*. A language LΣ*L \subseteq \Sigma^* is regular if there is a DFA MM such that L={wΣ*:M accepts w}L = \{w \in \Sigma^* : M \text{ accepts } w \}.

Conceptually, when we think about a DFA, we understand that it can do two things:

It’s important to understand that these two interpretations of what a DFA does are equivalent, because there is a simple bijection between the set of all possible languages (the power set 2Σ*2^{\Sigma^*} of Σ*\Sigma^*) and the set of all possible functions f:Σ*{0,1}f: \Sigma^* \rightarrow \{0,1\}.

  1. Construct a bijection from the power set 2Σ*2^{\Sigma^*} to {f:f:Σ*{0,1}}\{f: f: \Sigma^* \rightarrow \{0,1\}\}.

Our solution ended up involving the indicator function for a set LΣ*L \subseteq \Sigma^*, where fL(w)={1 if wL,0 otherwise.f_L(w) = \begin{cases} 1 & \text{ if } w \in L, \\ 0 & \text{ otherwise.} \end{cases}

Wed, Jan 29

We started by constructing some more examples of DFAs.

  1. Find a DFA that computes whether a binary string ends with 011. (This is similar, but not the same as a DFA we constructed last time).

  2. How many states would a DFA need if you wanted to check whether a binary string has a 1 in the third position from the last?

Then we talked about these properties of regular languages.

Theorem (Finite languages are regular). If Σ\Sigma is a finite alphabet, then any finite LΣ*L \subset \Sigma^* is a regular language.

  1. How would you prove this? If I give you a finite set of strings LL, how could you turn that into a DFA that recognizes LL?

Theorem (Closure properties of regular languages). Let A,BΣ*A, B \subseteq \Sigma^* be regular languages. Then

We say that the set of regular languages over Σ\Sigma is closed under the three operations: union, concatentation, and Kleene-star.

  1. Suppose that A={big,small,pet}A = \{ \text{big}, \text{small}, \text{pet} \} and B={cat,dog}B = \{ \text{cat}, \text{dog} \}. What are the elements of ABA \circ B and B*B^*?

We proved the first part of the theorem above (regular languages are closed under unions) in class by thinking about how to construct a new DFA MM that accepts ABA \cup B using DFAs MAM_A and MBM_B that accept AA and BB respectively. To prove this, we answered these questions:

  1. If the machine MM is built by running both MAM_A and MBM_B simultaneously, what are the possible states of MM?

  2. What are the initial states for MM?

  3. What is the transition function for MM?

  4. What are the accepting states of MM?

Fri, Jan 31

  1. We started by asking whether regular languages are closed under intersections. Can we adapt the proof we gave last time for unions to intersections?

    Yes, regularly languages are closed under intersections. The proof is almost exactly the same as the proof for unions, the only difference is that the DFA to recognize ABA \cap B has final states FA×FBF_A \times F_B while the DFA to recognize ABA \cup B has final states (FA×QB)(QA×FB)(F_A \times Q_B) \cup (Q_A \times F_B), that is, any pair of states from QA×QBQ_A \times Q_B where at least one of the two is final.

To prove that regular languages are closed under concatenation, we introduced a new idea: non-deterministic finite automata (NFAs).

We looked at this example:

An NFA can have more than one arrow exiting a state with the same input. When that happens, the machine splits into multiple copies, one for each possible next state.

If an NFA enters a state that has exiting arrows labeled with ϵ\epsilon, then the NFA immediately splits into multiple copies, one where the current state stays the same, and one where the current state moves along each branch labeled ϵ\epsilon. This can happen several times, if there is a sequence of branches labeled ϵ\epsilon.

If there is no arrow leaving a state for a given input signal, that means that the current copy of the machine is a dead end and it does not need to be continued.

The NFA accepts a string if any of the parallel computations is in an accepting state after the string is read. We answered these questions:

  1. Which states are active at each step as we read the input string 01010?

  2. Does this NFA accept the string 01010?

  3. Describe the set of all strings in {0,1}*\{0,1\}^* that this NFA will accept.

Here is the technical definition of an NFA.

Definition. A non-deterministic finite automata (NFA) consists of

  1. A finite set of states QQ.
  2. A finite alphabet Σ\Sigma of possible input signals.
  3. A transition function δ:Q×Σ2Q\delta: Q \times \Sigma \rightarrow 2^Q.
  4. An initial or start state q0Qq_0 \in Q.
  5. A set of final or accepting states FQF \subseteq Q.

The only change from the definition of a DFA is that the transition function for an NFA can return a set of states at each step (including, possibly, the empty set). Think of these multivalued transitions as creating multiple branching computations that run in parallel. If the transition function returns the empty set, that means that that branch of the parallel computation is a dead end and does not continue.

  1. Let A,BΣ*A, B \subseteq \Sigma^* be regular languages recognized by DFAs MAM_A and MBM_B respectively. Describe an NFA that uses the MAM_A and MBM_B to check if a string is in the union ABA \cup B. Note that every DFA is an NFA that has a transition function that only ever returns a single state at a time.

  2. Describe an NFA that can check whether or not the 3rd to last digit of a binary string is 1 using fewer states than the DFA we described last time. Hint: What if the NFA “guesses” every time it sees a 1 that that might be the third to last entry. What should it’s states be from that point on?

Week 4 Notes

Day Section Topic
Mon, Feb 3 2.5 NFAs and DFAs are equivalent
Wed, Feb 5 2.7 Regular expressions
Fri, Feb 7 2.8 Regular expressions and languages

Mon, Feb 3

Today we continued talking about NFAs. We’ve been following the textbook by Meshwari & Smid pretty closely this week, so I recommend reading skimming Section 2.6 which is what we covered today.

  1. We described an NFA that recognizes the concatentation ABAB of two regular languages by running DFAs MAM_A and MBM_B in parallel. We answer these questions about the NFA:

    1. What are the states of the NFA?
    2. What are the final states of the NFA?
  2. We also described an NFA that recognizes the Kleene star A*A^* of a regular language. The idea is similar to the previous construction, but you need a way to accept the empty string (if it isn’t part of AA).

To complete the proof that regular languages are closed under these two operations, we need to prove this:

Theorem (Equivalence of NFAs and DFAs). If there is an NFA that recognizes a language AΣ*A \subseteq \Sigma^*, then there is a DFA that also recognizes AA.

The proof is surprisingly simple. An NFA with states QQ, transition function δ:Q×Σ2Q\delta: Q \times \Sigma \rightarrow 2^Q can be turned into a DFA that has states 2Q2^Q and transition function δ2:2Q×Σ2Q\delta_2: 2^Q \times \Sigma \rightarrow 2^Q.

  1. Describe specifically what δ2(S,σ)\delta_2(S, \sigma) returns for any subset SQS \subseteq Q using δ\delta.

  2. If FF denotes the accepting states of the NFA, then what are the corresponding accepting states of the DFA?

  3. If qQq \in Q is the initial state of the NFA, what is the initial state of the DFA?

We applied the idea of this proof to Exercise 5 from Homework 3.

Wed, Feb 5

We introduced regular expressions (regex).

Definition. A regular expression over an alphabet Σ\Sigma is a string ee with symbols from the extended alphabet Σ{(,),*,|}\Sigma \cup \{ ~(~ , ~)~ , ~*~ , ~|~ \} that has one of the following forms:

  1. A symbol eΣe \in \Sigma is a regular expression.
  2. A concatenation e=e1e2e = e_1e_2 where e1e_1 and e2e_2 are regular expressions.
  3. A union e=e1|e2e = e_1|e_2 where e1e_1 and e2e_2 are regular expressions.
  4. A grouping e=(e1)e = (e_1) where e1e_1 is a regular expression.
  5. A star e=e1*e = e_1* where e1e_1 is either a single symbol or grouping.

We also accept the empty set \varnothing and the empty string ϵ\epsilon as regular expressions.

Regular expressions are used to match sets of strings (i.e., languages over Σ\Sigma).

  1. Describe what the following regular expressions represent:

    1. (0|1)0*
    2. Σ*1Σ*
    3. (0|1)*
  2. Write a regular expression that recognizes the base-10 decimal form of any integer that is a multiple of 5.

Most programming languages use Perl compatible regular expressions which have several additional features to make it easier to work with regular expressions. Here are some examples we talked about in class.

Special Symbols
\s whitespace (tabs, spaces, etc.)
\w alphanumeric (letters & digits)
\d digits (0-9)
. wildcard matches any single character
\ escape (for special characters)
Extra Operators
(e)+ at least one
(e)? maybe one (0 or 1)
Square Brackets
Can match one or more ranges of characters
[a-z] any lower case letter
[a-zA-Z] any letter
[^a-z] anything except a-z

Here is a simple Python regex example:

import re 

s = "My pet dog likes to go to the park."
expr = "pet\s(dog|cat|bird)"
match = re.search(expr, s)
print(match)
  1. Write a regular expression that would match any e-mail address of the form name@domain.extension.

  2. Find a regular expression over the alphabet Σ={0,1}\Sigma = \{0,1\} that matches all strings that start with a 1, end with a 1, and have an even number of zeros between.

We stated, but did not prove the following theorem.

Theorem (Equivalence of regular expressions and regular languages). A language AΣA \subseteq \Sigma is regular if and only if there is a regular expression that describes it.

  1. Find an NFA that recognizes the same language as the regular expression (ab|a)*. Use the ideas from the previous classes about how to build NFAs to find the union, concatenation, and Kleene star of languages.

Fri, Feb 7

Today we talked about the proof of the theorem from last time that regular languages and regular expressions are equivalent. One direction is easy: every language described by a regular expression is regular. This is because we already know that regular languages are closed under the union, concatenation, and Kleene-star operations and we also know that any finite set of strings is a regular language.

Therefore, we should be able to solve this example, which we didn’t have time for on Wednesday.

  1. Find an NFA that recognizes the same language as the regular expression (ab|a)*. Use the ideas from the previous classes about how to build NFAs to find the union, concatenation, and Kleene star of languages.

To prove the converse, we described an algorithm for converting NFAs to regular expressions. Note that the Maheshwwari & Smid textbook describes a different approach to convert an NFA into a regular expression in Section 2.8. Our approach is to use a generalized NFA:

Definition. A generalized non-deterministic finite automata (GNFA) is just like an NFA, except for the following changes:

  1. There is only one accept state and it is not the start state.

  2. No transitions enter the start state or leave the accept state.

  3. Transitions can be triggered by input strings that match a regex (not just single symbols).

To finish the proof, we make two observations:

First, you can convert any NFA to a GNFA by adding a special start state that has an epsilon transition to the original start state and a new accept state that has an epsilon transition from any old final state.

Second, if you have a GNFA with n>2n > 2 states, there is an equivalent GNFA with n1n-1 states, because you can remove any state (other than the initial and final states) and replace incoming/outgoing transitions with transitions described by a regex.

We did the following example to illustrate.

  1. Let Σ={A,C,G,T}. Convert the following NFA to a GNFA, and then remove states until there are only 2 (the start and accept states).
  1. What regular expression is equivalent to the NFA above?

We finished by talking about languages which are not regular. One simple language that is not regular is the following:

A={w{0,1}*:w has an equal number of 0s and 1s}.A = \{w \in \{0,1\}^* : w \text{ has an equal number of } 0 \text{s and } 1 \text{s} \}.

Week 5 Notes

Day Section Topic
Mon, Feb 10 2.9 Non-regular languages & the pumping lemma
Wed, Feb 12 2.9 More non-regular languages
Fri, Feb 14 Review

Mon, Feb 10

Theorem (Pumping Lemma). If LΣ*L \subseteq \Sigma^* is a regular langauage, then LL has a length p1p \ge 1 (called the pumping length of LL) such that if a string wLw \in L is longer than pp, then w=xyzw = xyz where x,y,zΣ*x, y, z \in \Sigma^* are substrings such that

  1. The middle part yy is not the empty string (i.e., |y|1|y| \ge 1),
  2. The first two parts have length |xy|p|xy| \le p, and
  3. The middle part can be “pumped”, that is xynzLxy^nz \in L for all n0n \ge 0.

Here is a picture that helps to understand what is going on. It also gives a clue about why this theorem is true.

The proof is really simple. If a finite state machine reads a long enough string, eventually it has to loop back to the same state twice. The input signals that caused the machine to take that loop make up the substring yy in the theorem.

The pumping lemma is a useful tool to prove that some languages are not regular. You can use the pumping lemma to give a proof by contradiction. Be careful, the pumping lemma cannot be used to prove that a language is regular.

Steps to prove a language is not regular

  1. Temporarily assume the language LL is regular and has pumping length pp.

  2. Try to construct a string in LL so that:

    1. The first part of the string is longer than pp.
    2. No substring inside the first part can be pumped.
  3. If you can’t pump the string, that contradicts the pumping lemma, so LL can’t be regular.

We applied this technique to prove that the following languages are not regular.

  1. L={w:w has an equal number of 0 and 1’s}L = \{w : w \text{ has an equal number of 0 and 1's}\}
  1. L={0n2:n}L = \{0^{n^2} : n \in \mathbb{N}\}.

Here is a nice meme proof using the pumping lemma from Barak textbook.

Wed, Feb 12

Today we looked at more examples of regular and non-regular languages.

  1. L={w{0,1}*:w is a palindrome}L = \{w \in \{0, 1\}^* : w \text{ is a palindrome} \}.

  2. L={w:w contains an equal number of 01 and 10 substrings}L = \{w : w \text{ contains an equal number of } 01 \text{ and } 10 \text{ substrings}\}.

The converse of the pumping lemma is not true. So you can’t use the pumping lemma to prove that a language is regular. In fact, here is a language that is non-regular, but every string in the language is pump-able!

L={aibjck:i,j,k0 and if i=1 then j=k}.L = \{a^i b^j c^k : i, j, k \ge 0 \text{ and if } i = 1 \text{ then } j = k\}.

  1. Explain why there is a pumping number pp (in fact p=1p=1 works) such that any string wLw \in L can be “pumped”.

  2. Despite this, explain why LL is not a regular language. Hint: if there was a DFA that recognizes LL, what other DFAs could you construct?

Many programming languages, including Python & Javascript allow backreferences to previous groups in a regular expression. A group is a part of the regular expression inside parentheses. The special symbol \1 refers to anything matched by the first group in the regular expression. Similarly \2 refers back to anything matched by the second group, and so on. For example: the regular expression "(\w\w*) \1" would match any repeated word like “word word” or “dog dog”. I tried to demonstrate this in class, but I got tripped up because Python requires the regular expression to be entered as a raw string.

import re

regex = r"(\w\w*) \1" # need the r before the quotes to indicate that this is a raw string

print(re.search(regex, "Did I type the the same word twice?"))
  1. Explain why regular expressions with backreferences are not really regular expressions (at least not according to our definition). Show that they can match non-regular languages.

Fri, Feb 14

Today we went over some of the problems from homework 4 and the midterm 1 review problems.

Week 6 Notes

Day Section Topic
Mon, Feb 17 Midterm 1
Wed, Feb 19 3.1 - 3.2 Context-free grammars
Fri, Feb 21 3.3 Parsing and parse-trees

Wed, Feb 19

Class canceled because of snow.

Fri, Feb 21

Today we introduced context-free grammars.

Definition. A context-free grammar (CFG) consists of

  1. A set of variables VV
  2. A set of symbols Σ\Sigma (disjoint from VV)
  3. A set of rules of the form ABA \rightarrow B where AVA \in V and B(ΣV)*B \in (\Sigma \cup V)^*
  4. A start state SVS \in V.

  1. We explained how the context free grammar

    S         →  aSb
    S         →  ε

    generates the (non-regular) language L={anbn:n}.L = \{a^n b^n : n \in \mathbb{N}\}.

  2. We looked at the example of a CFG and used it to identify the parts (V,Σ,R,S)(V,\Sigma, R, S) in the formal definition.

    S         →  <subject> <verb> <object>
    <subject> →  <article> <noun>
    <article> →  a | the
    <noun>    →  boy | girl | dog
    <verb>    →  runs | walks | talks
    <object>  →  <prep> <subject>
    <prep>    →  to |from

    Draw a parse tree that shows how to derive the sentence “The boy talks to the girl.”

  3. Find a CFG for the language of matched parentheses.

    Seth suggested this CFG:

         S         →  (S)
         S         →  ()S
         S         →  ε

    This isn’t correct, but it is a good first guess. Explain why this grammar cannot be used to construct the string (())().

  4. Find a context-free grammar for the language L={w{0,1}*:w has an equal number of 0’s and 1’s}.L = \{w \in \{0,1\}^* : w \text{ has an equal number of } 0\text{'s and } 1 \text{'s} \}.

We didn’t get to this example, but it is good practice too.

  1. Describe how to generate the string (a + a) * a using the context free grammar below:

    <expr>    →  <expr> + <term> | <term>
    <term>    →  <term> * <factor> | <factor>
    <factor>  →  (<expr>) | a

We finished by looking at this example that gives the idea of how a programming language can be though of as a context-free language.

Week 7 Notes

Day Section Topic
Mon, Feb 24 3.3 Parsing and parse-trees
Wed, Feb 26 3.8 The pumping lemma for CFGs
Fri, Feb 28 3.5 - 3.6 Pushdown automata

Mon, Feb 24

Last time we used parse trees, but didn’t define them.

Definition (Parse Tree). Given a context free grammar (V,Σ,S,R)(V, \Sigma, S, R) and a string wΣ*w \in \Sigma^* that can be generated by the grammar, a parse tree for the string is a labeled ordered tree with nodes labeled by elements in VΣV \cup \Sigma such that the arrows leaving a node labeled by an element in VV correspond to one of the rules in RR and the nodes labeled by elements in Σ\Sigma are terminal (no arrows leave them).

Today we looked at some more examples of Context Free Grammars. Lisp is a family of programming languages that have a particularly simple syntax. Lisps are built around S-expressions, which are defined by the following CFG rules:

S → (L) | A
L → S | S L
A → <number> | <string> | <symbol>

For the variables above, S represents any S-expression, L represents a list, and A represents an atom, which can be a number, a string, or a symbolic expression like a variable name or operator (among other things).

  1. Draw a parse tree for the S-expression (* (+ 5 6) (- 4 2)). Each node in the parse tree should be labeled with a variable (or symbol if it is a terminal node).

  2. Is it possible to draw a different parse tree for this expression? Why or why not?

The S-expression grammar is unambigious, every expression has only one possible parse tree. Unfortunately, many CFGs are ambiguous, i.e., they allow strings with more than one parse tree. Here is a simple example.

E → E + E 
E → E * E 
E → (E) 
E → a
  1. Use this CFG to derive the string a+a*a two different ways.

In some (but not all cases) it is possible to find two CFGs that generate the same set of strings, but one is ambiguous and the other isn’t. For example, if we change the grammar above to distinguish between expressions that are terms (T), or factors (F) or sums of terms (E), then we can make an equivalent CFG that is unambiguous:

E  →  E + T | T
T  →  T * F | F
F  →  (E) | a

We say that a language is context-free if it can be generated by a CFG. We use the shorthand CFL to denote a context-free language.

Theorem (Regular Languages are Context-Free). Every regular language is context-free.

We’ll prove this later by showing that context-free grammars correspond to a special kind of automata with a (potentially infinte) stack called a pushdown automata. But today, we talked about how you can prove this theorem using the three operations (union, concatenation, and Kleene star) for regular expressions.

  1. If AA and BB are both context-free languages, explain why the union ABA \cup B is context free.

  2. If AA and BB are both context-free languages, explain why the concatenation AAA \circ A is context free.

  3. Explain why the Kleene star of a CFL is context-free.

Wed, Feb 26

Theorem (Pumping Lemma for CFLs). If LΣ*L \subseteq \Sigma^* is a context-free language, then LL has a length p1p \ge 1 (called the pumping length of LL) such that if a string wLw \in L is longer than p,p, then w=uvxyzw = uvxyz where u,v,x,y,zΣ*u, v, x, y, z \in \Sigma^* are substrings such that

  1. At least one of vv or yy is not empty (i.e., |vy|1|vy| \ge 1),
  2. The middle part has length |vxy|p|vxy| \le p, and
  3. Both vv and yy can be “pumped”, that is uvnxynzLuv^nxy^nz \in L for all n0n \ge 0.

This is a little more complicated than the pumping lemma for regular languages. We drew some pictures of parse trees to give an intuitive explanation for why this new pumping lemma is true (see page 126 in Maheshwari & Smid).

We started by looking at two examples:

  1. Show that the language L={anban:n}L = \{a^n b a^n : n \in \mathbb{N}\} is context-free but not regular.

  2. What is the pumping length for the language LL above?

  3. Show that L={w=w:w{a,b}*}L = \{w = w : w \in \{a, b\}^* \} is not context-free.

  1. Show that L={an2:n}L = \{a^{n^2} : n \in \mathbb{N}\} is not context-free.

We didn’t have time for this last problem, so I left it as an exercise to consider:

  1. A={anbncn:n}A = \{a^n b^n c^n : n \in \mathbb{N}\} is not context-free. What about the language B{a,b,c}*B \subset \{a,b,c\}^* that consists of all strings with an equal number of a’s, b’s, and c’s? Is that context-free?

Fri, Feb 28

Today we talked about pushdown automata (PDA) which are basically just NFAs with a stack. Note that the book calls these nondeterministic pushdown automata. There are also deterministic pushdown automata, but those turn out to be more complicated, and only the nondeterministic pushdown automata are actually equivalent to CFGs. So, we will always assume that our PDAs are nondeterministic, unless specifically mentioned otherwise. We started with this example:

This PDA accepts the language L={anbn:n}L = \{a^n b^n : n \in \mathbb{N}\}. Notice that each arrow has three parts, for example the first loop is labeled a,ϵ/aa, \epsilon/a which means you can take this loop if you read an aa from the input string, and pop ϵ\epsilon (i.e., nothing) from the stack, and then push aa onto the stack. We will always use this notation (read, pop/push) for each arc.

PDAs work just like NFAs, except they have this extra rule: You only accept a string if you finish in an accepting state and the stack is empty.

Some things to note:

  1. Since the machine is nondeterministic, you can have more than one state (and more than one stack) branching off whenever you have a choice about which arrow to take.

  2. If you are in a state and there is no valid arrow to take, then that state dies.

Here is the formal definition.

Definition. A nondeterministic pushdown automata (PDA) consists of

  1. A finite set of states QQ.
  2. A finite input alphabet Σ\Sigma.
  3. A finite stack alphabet Γ\Gamma.
  4. A transition function δ:Q×(Σ{ϵ})×(Γ{ϵ})2Q×(Γ{ϵ})\delta: Q \times (\Sigma \cup \{\epsilon\}) \times (\Gamma \cup \{ \epsilon \}) \rightarrow 2^{Q \times (\Gamma \cup \{\epsilon\})}.
  5. A start state q0Qq_0 \in Q.
  6. A set of accepting states FQF \subseteq Q.

Note: Some alternative definitions allow PDAs to push more than one symbol from Γ\Gamma onto the stack in one step.

We looked at these questions:

  1. For the PDA below, describe the language that it accepts.

  2. Describe an PDA that accepts the language L={anbm:nm}L = \{ a^n b^m : n \ge m \}.

  3. Describe an PDA with just one state that accepts the language of balanced parentheses.

  4. How many distinct PDAs with 5 states, and Σ=Γ={0,1}\Sigma = \Gamma = \{0,1\} are possible (assuming you only push at most one variable onto the stack at each step), according to the definition?

Week 8 Notes

Day Section Topic
Mon, Mar 3 3.7 Pushdown automata & context-free grammars
Wed, Mar 5 3.4 Chomsky normal form
Fri, Mar 7 4.1 - 4.2 Definition of Turing machines

Mon, Mar 3

Today we talked about the equivalence between CFGs and PDAs.

Theorem (Equivalence of CFGs and PDAs). A language is context-free if and only if there is a pushdown automata that recognizes it.

We sketched the proof of one direction: that if you have a CFG, then there is an PDA that accepts the same language. The other direction is harder to prove, so we skipped it, but you can find the details in more advanced theory of computing textbooks.

The idea of the proof for the forward direction is to create an PDA with three states. In state q0q_0 you immediately transition to q1q_1 reading nothing and push S$ on the stack (where $ is a special symbol to mark the bottom of the stack and S is the symbol for the start variable). Then in state q1q_1 you have two different types of transitions that loop back to q1q_1:

Finally, you have a final transition to the accepting state q2q_2 which involves reading nothing and popping the $ off the stack. If you have finished reading the input when you do this, you will accept the string.

We illustrated how this idea works using the grammar and the input string aabbbb.

    S →  AB
    A →  aAb | ε
    B →  bB | b

After this example, we talked briefly about normal forms for context-free grammars. One important normal form is the Chomsky normal form, which we will talk about next time.

For a context-free grammar GG, we write L(G)L(G) to denote the language generated by GG.

  1. A variable AA in a context-free grammar GG is useless if there are no strings in L(G)L(G) that have a parse tree with AA as one of the nodes. Prove that if GG has any useless variables, then there is another context-free grammar HH that generates the same language, but does not have any useless variables.

    Let HH be the CFG obtained by removing all useless variables from GG and all rules that reference those useless variables. We have to show that HH and GG both generate the same language. If wL(G)w \in L(G), then there is a parse tree for ww that doesn’t contain any useless variables. Therefore the same parse tree is a parse tree for generating ww using HH, so wL(H)w \in L(H). If wL(H)w \in L(H), then all of the variables and rules for the parse tree to generate ww are also variables and rules of GG. Therefore wL(G)w \in L(G) which proves that L(H)=L(G)L(H) = L(G). \square

Wed, Mar 5

Definition (Chomsky normal form). A context free grammar G=(V,Σ,R,S)G = (V, \Sigma, R, S) is in Chomsky normal form if all rules are in one of the following forms:

  1. SϵS \rightarrow \epsilon
  2. ABCA \rightarrow BC
  3. AaA \rightarrow a

where A,B,CVA, B, C \in V, aΣa \in \Sigma, and neither BB nor CC is the start variable.

One nice feature of Chomsky normal form is that it makes it easier to check whether a string can be generated by a grammar.

Definition (Equivalent grammars). Two context-free grammars are equivalent if they generate the same language.

It turns out that there is no general algorithm to check whether or not two CFGs are equivalent. But all CFGs can be converted to Chomsky normal form.

Theorem (Chomsky). Every CFG is equivalent to a CFG in Chomsky normal form.

The proof is the following algorithm to convert any CFG into Chomsky normal form.

At the end of this process your grammar will be in Chomsky normal form.

  1. Convert the following CFG to Chomsky normal form.

    S → ABa
    A → aab
    B → Ac
  2. Convert the following CFG to Chomsky normal form.

    S → aSa | A
    A → abA | b
  3. Explain why the following algorithm won’t work to check if any two CFGs are equivalent:

We didn’t have time for it today, but here is one more practice example:

  1. Convert the following CFG to Chomsky normal form.

    S → ASA | A | ε
    A → aa | ε

Fri, Mar 7

Today we introduced Turing machines (TMs) which are the most general model of computation that we will discuss. Turing machines are deterministic finite state machines, with an infinite tape that the machine can move along, and both read and write from. Like DFAs and PDAs, Turing machines can accept or reject an input string, but Turing machines also have a third possibility: they might loop forever and neither accept nor reject.

Here are some differences between finite automata and Turing machines:

We looked at these examples of TMs:

  1. Describe an algorithm that a Turing machine could implement to check if a string in {a,b,=}*\{a,b,=\}^* is in the language {w=w:w{a,b}*}\{w=w : w \in \{a,b\}^* \}. Use the tape alphabet {a,b,=,x,}\{a,b,=,x, □\}, where you can use an xx to mark symbols that have been dealt with. Hint: How could you check that the first character of the string matches the first character after the equals sign?

  2. Draw a state diagram for the Turing machine above. Use the notation readwrite,move\text{read} \rightarrow \text{write}, \text{move} to label transitions. The move can be either LL or RR, and you can use regular expressions for the character to read. The character to write is optional, you can move without writing anything.

  3. Describe an algorithm that a TM could use to accept the language {a2n:n0}.\{a^{2^n} : n \ge 0\}. Hint: Try moving from left to right, crossing out every other aa as you go. How will you know if there were an even number of a’s? Could you repeat the process? When should you stop?

We didn’t get to it in class, but here is a state diagram for a TM that can accept the language L={a2n:n0}.L = \{a^{2^n}: n \ge 0\}.

Week 9 Notes

Day Section Topic
Mon, Mar 17 4.2 Turing computable functions
Wed, Mar 19 4.3 Multi-tape Turing machines
Fri, Mar 21 4.4 Church-Turing thesis

Mon, Mar 17

Here is the formal definition of a Turing machine.

Definition. A Turing machine consists of

  1. A finite set of states QQ.
  2. A finite input alphabet Σ\Sigma that doesn’t include the blank symbol □.
  3. A finite tape alphabet Γ\Gamma such that ΣΓ\Sigma \subset \Gamma and Γ□ \in \Gamma.
  4. A transition function δ:Q×ΓQ×Γ×{L,R}\delta: Q \times \Gamma \rightarrow Q \times \Gamma \times \{L, R\}.
  5. A start state q0Qq_0 \in Q.
  6. An accept state qacceptQq_{\text{accept}} \in Q.
  7. A reject state qrejectQq_{\text{reject}} \in Q.

Because Turing machines can loop forever, there is a difference between recognizing/accepting a language and deciding a language.

Definition. A Turing machine accepts (or recognizes) a language LΣ*L \subseteq \Sigma^* if the set of strings that it accepts is exactly LL.

A Turing machine that only recognizes a language LL might loop forever when you give it a string that is not in LL.

Definition. A Turing machine decides a language LΣ*L \subseteq \Sigma^* if it not only recognizes LL, but also rejects every string that is not in LL.

Both of the example Turing Machines above actually decide their languages, since they will successfully reject any input that doesn’t match a valid string (they won’t get stuck looping forever without halting). We’ll see examples later of Turing machines that accept but don’t decide a language.

Another thing that Turing Machines can do is compute functions by leaving the output on the tape when it halts.

  1. Construct a Turing machine that reads an input string w{a}*w \in \{a\}^* and leaves wwww on the tape when it halts.

This wasn’t to hard to describe the algorithm, but it got a little complicated when we made the state diagram. Once we did that, we defined Turing computable functions.

Definition. A function f:Σ*Σ*f: \Sigma^* \rightarrow \Sigma^* is Turing computable if there is a Turing machine that halts with f(w)f(w) on the tape whenever it is given a tape with wΣ*w \in \Sigma^* as input.

  1. Explain why the function f(w)=wwwwf(w) = wwww is Turing computable.

Theorem (Compostions are computable). If f,g:Σ*Σ*f, g : \Sigma^* \rightarrow \Sigma^* are both Turing computable functions, then so is the composition fgf \circ g.

  1. Explain why the function f(w)={1 if length of w is even.0 otherwise.f(w) = \begin{cases} 1 & \text{ if length of } w \text{ is even.} \\ 0 & \text{ otherwise.} \end{cases} is Turing computable. Hint: Every regular language is Turing decidable.

We finished by describing how this function (which comes up in the Collatz-Conjecture) is Turing computable: f(n)={n/2 if n is even,3n+1 if n is odd.f(n) = \begin{cases} n/2 & \text{ if } n \text{ is even,} \\ 3n+1 & \text{ if } n \text{ is odd.} \end{cases}

Wed, Mar 19

Today we looked at multi-tape Turing machines.

Theorem. Any k-tape Turing machine has an equivalent single-tape Turing machine.

Proof idea. For every tape symbol of the k-tape TM, we need a “dotted” symbol and we also need a special separator symbol (#). Then we can put the contents of the k tapes onto a single tape if we add separators and dotted symbols to mark the which symbols are currently pointed to by the TM.

# a a # b b a # a a a #

For each step of the k-tape TM, the 1-tape TM follows these rules:

Another variation on a standard TM is a nondeterministic Turing machine (NTM). At each step, these can split into more than one possible next state. The transition function has form: δ:Q×Γ2Q×Γ×{L,R}.\delta: Q \times \Gamma \rightarrow 2^{Q \times \Gamma \times \{L, R \}}. The nondeterministic TM accepts a language as soon as any of its parallel versions reaches an accept state.

Theorem. Any language that can be accepted by a nondeterministic Turing machine can also be accepted by a regular Turing machine.

We talked about how you can use a 3-tape TM to used a breadth first search through the branching possibilities of a NTM to see if any accept.

We also talked about other things that a TM can do:

  1. If f:Σ*{0,1}f: \Sigma^* \rightarrow \{0,1\}, and g,h:Σ*Σ*g, h: \Sigma^* \rightarrow \Sigma^* are Turing computable, then so is the function

    IF f(x) THEN g(x) ELSE h(x).

  2. What about a while-loop?

    while(f(x)):
        x = g(x)

We didn’t have time for this last example:

  1. Can a Turing machine have a for-loop, where it repeats some function nn times? Something like this Python example where function ff is applied nn times?

    for i in range(n): 
        x = f(x)

Fri, Mar 21

We started with this example:

  1. Suppose that f:Σ*Σ*f: \Sigma^* \rightarrow \Sigma^* is a Turing computable function. Describe a 2-tape TM that inputs a string xΣ*x \in \Sigma^* on one tape and a number nn (encoded with nn 1’s) on the second tape, that computes fk(x):=f(f(f(k-timesx)))).f^{k}(x) := \underbrace{f(f(\cdots f(}_{k\text{-times}}x)))).

In mathematics, there are many algorithms, which are step-by-step procedures that can be carried out to solve problems. Prior to the work of Turing and other computer scientists, no one ever tried to define exactly what an algorithm is. It turns out to be hard to give a precise definition of an algorithm. Computer scientists have settled on the following:

Definition (Church-Turing thesis). An algorithm is a computational task that can be carried out by a Turing machine.

The idea is that Turing machines can do all of the things (like looping, and conditionals, storing variables) that we think of algorithms as doing, so we use TMs to give a precise definition for algorithms.

There are many other models of computing that can be proven to be Turing complete, that is, they can compute the same functions as Turing machines. But it is believed that there is no effective method of computation that can solve problems that Turing machines cannot solve.

We finished by discussing Hilbert’s 10th problem: Is there an algorithm to determine whether or not a multivariable polynomial with integer coefficients has an integer root? We looked at some example polynomials like

p(x,y)=3x22xy+y22p(x,y) = 3x^2 - 2xy + y^2 - 2 and q(x,y)=x2+y2+1.q(x,y) = x^2 + y^2 + 1. The pair (1,1)(1,1) is a root for the first polynomial, since p(1,1)=0p(1,1) = 0, but the second polynomial has no integer roots.

  1. Describe an algorithm that we could implement using a TM or computer program to try to find out if a polynomial has a root.

The algorithm we described will accept any polynomial that has a root. But it doesn’t reject polynomials that don’t, instead it loops forever looking for a root. It wasn’t until 1970 that Hilbert’s 10th problem was solved, when it was proved that there is no Turing machine that can decide the language of integer multivariable polynomials with integer roots. We’ll look at other examples of undecidable languages after the midterm.

Week 10 Notes

Day Section Topic
Mon, Mar 24 Review
Wed, Mar 26 Midterm 2
Fri, Mar 28 5.5 - 5.7 Universal Turing machines

Mon, Mar 24

Today we reviewed context-free languages for the midterm exam. We also talked about the Chomsky hierarchy of languages.

Fri, Mar 28

Today we introduced universal Turing machines. We used them to consider this language:

Accept={M,w:M is a Turing machine and w is a string that M accepts}.\text{Accept} = \{ \langle M, w \rangle : M \text{ is a Turing machine and } w \text{ is a string that } M \text{ accepts}\}.

  1. Explain why Accept is Turing recognizable.

Then we gave a diagonalization proof that Accept is not Turing decidable. Suppose there was a decider for Accept. Then we could make a table with rows for each Turing machine MiM_i and columns every Turing machine encoding Mj\langle M_j\rangle and entries 1 or 0 depending on whether MiM_i accepts Mj\langle M_j \rangle.

  1. What would happen if we created a TM NN inputs any TM encoding M\langle M \rangle and then does the opposite of what DD does on M,M\langle M, \langle M \rangle \rangle. What would NN do when you input N\langle N \rangle into NN?

  2. Why is that a contradiction?

We finished by observed that there are only a countably infinite number of possible Turing machines, but there is an uncountable infinite number of languages, so there must be lots of languages that are not decidable!

Week 11 Notes

Day Section Topic
Mon, Mar 31 5.1-5.3 The halting problem & Rice’s theorem
Wed, Apr 2 6.1 The running time of algorithms
Fri, Apr 4 6.1 The running time of algorithms

Mon, Mar 31

We started by explaining why our proof from last time that the language Accept is undecidable was a diagonalization argument. Suppose there were a Turing machine DD (call it a decider) that can decide whether or not any other TM will accept any string.
D(M,w)={1 if M accepts w.0otherwise.D(\langle M, w \rangle) = \begin{cases} 1 & \text{ if }M \text{ accepts } w. \\ 0 & \text{otherwise}. \end{cases}

Imagine we make a table with columns corresponding to different Turing machine encodings and rows corresponding to different Turing machines showing the output of DD on each pair.

DD M1\langle M_1 \rangle M2\langle M_2 \rangle M3\langle M_3 \rangle \ldots
M1M_1 1 0 1
M2M_2 0 1 0
M3M_3 0 0 0
\vdots

But what happens if we create a new TM NN that always returns the opposite of DD? Why is there a contradiction if we try to look up the diagonal entry where NN is given itself and its encoding string as inputs?

By proving that no TM can decide the language Accept, we get the following corollary about the halting problem.

Theorem (The Halting Problem is Undecidable). The language
Halt={M,w:M is a TM that halts on w}\text{Halt} = \{ \langle M, w \rangle : M \text{ is a TM that halts on }w \}

is undecidable.

We gave a proof by contradiction for this theorem.

It turns out that not only is the halting problem undecidable, but lots of other questions about Turing machines are undecidable. A powerful theorem known as Rice’s theorem says that every nontrivial semantic property of a Turing machine is undecidable.

Fix an encoding method and let LTML_{TM} be the set of all strings that are encoded Turing machines, i.e., LTM={M:M is a Turing machine}.L_{TM} = \{ \langle M \rangle : M \text{ is a Turing machine} \}. For any Turing machine MM, let L(M)={wΣ*:M accepts w}.L(M) = \{w \in \Sigma^* : M \text{ accepts } w \}.

Definitions. A property of Turing machines is a subset PLTMP \subseteq L_{TM}. We say that a Turing machine MM has property PP if MP\langle M \rangle \in P.

Here are some examples of properties. For each example, determine whether it is a syntactic or semantic property.

  1. INFINITE={M:L(M) is infinite}.\text{INFINITE} = \{\langle M \rangle : L(M) \text{ is infinite} \}.
  2. REGULAR={M:L(M) is a regular language}.\text{REGULAR} = \{\langle M \rangle : L(M) \text{ is a regular language} \}.
  3. HAS-THREE-STATES={M:M has 3 states}.\text{HAS-THREE-STATES} = \{\langle M \rangle : M \text{ has 3 states} \}.

Rice’s Theorem. Every nontrivial semantic property is undecidable.

We proved Rice’s theorem using a proof by contradiction (see below). Our book also has a different proof in Section 5.3 if you are interested.

Proof. Suppose M1M_1 is a Turing machine with property PP and M2M_2 is a Turing machine that doesn’t have PP. We can assume without loss of generality that L(M2)=L(M_2) = \varnothing. Then for any Turing machine MM and string ww, construct a new Turing machine TMwT_{Mw} that starts by running MM on ww until that halts, then it runs M1M_1 on the actual input, returning whatever M1M_1 returns. So L(TMw)={L(M1) if M halts on wL(M2)= if M doesn’t halt on w.L(T_{Mw}) = \begin{cases} L(M_1) & \text{ if } M \text{ halts on } w \\ L(M_2) = \varnothing & \text{ if } M \text{ doesn't halt on } w. \end{cases} Therefore TMwT_{Mw} has property PP if and only if MM halts on ww.

Wed, Apr 2

Today we introduced big-O notation and used it to talk about running times of algorithms.

Definition (Big-O notation). Let f,g:[0,)f, g: \mathbb{N}\rightarrow [0, \infty). We say that fO(g)f \in O(g) if there are constants C,n0>0C, n_0 > 0 such that f(n)Cg(n)f(n) \le C g(n) for all nn0n \ge n_0. So O(g)O(g) is the set of all functions that are eventually bounded by CgC g for some constant CC.

Note that there are lots of important examples in Barak Chapter 12 of algorithms and their running times.

We proved the following facts about big-O notation:

Theorem (Only Dominant Terms Matter). If f:[0,)f: \mathbb{N}\rightarrow [0,\infty) is a sum of nonnegative functions f1,,fkf_1, \ldots, f_k, then there is at least one function fjf_j such that O(f)=O(fj)O(f) = O(f_j). We call that function a dominant term for ff.

Theorem (Constants Don’t Matter). For any f:[0,)f: \mathbb{N}\rightarrow [0,\infty) and constant c>0c > 0, O(cf)=O(f)O(cf) = O(f) and O(f±c)=O(f)O(f \pm c) = O(f).

Theorem (Products of Functions). If f1O(g1)f_1 \in O(g_1) and f2O(g2)f_2 \in O(g_2), then f1f2O(g1g2)f_1f_2 \in O(g_1 g_2).

We get this hierarchy of functions:

logarithmsbase doesn’t matterpolynomialsdominant power mattersexponentialsbase mattersfactorials\underbrace{\text{logarithms}}_{\text{base doesn't matter}} \ll \underbrace{\text{polynomials}}_{\text{dominant power matters}} \ll \underbrace{\text{exponentials}}_{\text{base matters}} \ll \text{factorials}

We used these ideas to find the dominant term for these expressions:

  1. n4+n3logn+(logn)5n^4 + n^3 \log n + (\log n)^5.

  2. n100+2nn^{100} + 2^n.

So why do we use big-O notation? It makes it easier to talk about a lot of formulas. For example, an algorithm that requires 7(n+3)n!k!(nk)!+1677(n+3) \dfrac{n!}{k!(n-k)!}+167 steps can be described simply as being O(nk+1)O(n^{k+1}).

Caution 1. Big-O notation only gives an upper bound for how fast functions grow. Functions in O(f)O(f) don’t have to grow at the same rate as ff, they can also grow much slower.

Caution 2. You should always write the simplest possible expression inside big-O notation. For example O(5n3+logn)=O(n3)O(5n^3 + \log n) = O(n^3), so just write O(n3)O(n^3).

We finished by talking about how we use big-O notation to describe the runtime of algorithms. We did this example:

Week 12 Notes

Day Section Topic
Mon, Apr 7 6.2 The complexity class P
Wed, Apr 9 6.3 The complexity class NP
Fri, Apr 11 6.5 The SAT problem

Week 13 Notes

Day Section Topic
Mon, Apr 14 6.5 Hamilton path problem
Wed, Apr 16 6.5 NP-Complete languages
Fri, Apr 18

Week 14 Notes

Day Section Topic
Mon, Apr 21 Review
Wed, Apr 23 Midterm 3
Fri, Apr 25 TBA
Mon, Apr 28 TBA