# Graph Isomorphism on graphs of bounded treewidth

The Graph Isomorphism problem is interesting for many reasons, one being the fact that its complexity status is still unclear - it is probably not NP-complete, but not known to be in P either. In the past decades, researchers identified a large number of special graph classes for which GI is polynomial-time solvable, including planar graphs, graphs of bounded maximum degree, and graphs excluding a fixed minor, among others.

A closer look at the known algorithms for graphs of bounded degree or graphs excluding a fixed minor reveals that the dependency in the running time bound on the "forbidden pattern" is quite bad:  in the first case or  in the second case, for some function . Seeing such dependencies, a natural question is: do there exist fixed-parameter algorithms, parameterized by the "forbidden pattern"? In other words, does the exponent in the polynomial factor in  need to depend on the pattern, on can we obtain an algorithm with running time ?

This question remains widely open for all aforementioned cases of bounded degree or excluded minor. In our recent FOCS 2014 paper, we have answered the question affirmatively for the special case of bounded treewidth graphs, proving that isomorphism of two -vertex graphs of treewidth  can be tested in time .

As a starting point, observe that comparing (testing isomorphism) of two graphs together with their given tree decompositions (that is, we want to check if the whole structures consisting of a graph and its decomposition are isomorphic) can be done easily in time polynomial in the size of the graphs, and exponential in the width of the decompositions. This is a relatively standard, but tedious, exercise from the area of designing dynamic programming algorithms on tree decompositions: if you have seen a few of these, I guess you can figure out the details, but if you don't, this is probably one of the worst examples to start with, so just assume it can be done. Alternatively, you may use the recent results of Otachi and Schweitzer that say that one needs only a set of bags for the decomposition, even without the structure of the decomposition tree itself.

With this observation, the task of comparing two bounded-treewidth graphs reduces to the following: given a graph  of treewidth , we would like to compute in an isomorphic-invariant way a tree decomposition of $G$ of near-optimal width. Here, isomorphic-invariant means that we do not want to make any decisions depending on the representation of the graph in the memory (like, "take arbitrary vertex"), and the output decomposition should be the same (isomorphic) for isomorphic input graphs. Given such an isomorphic-invariant treewidth approximation algorithm, we can now compute the decompositions of both input graphs, and compare given pairs (graph, its tree decomposition), as described in the previous paragraph. Note that for this approach we do not need to output exactly one tree decomposition; we can output, say,  of them, and compare every pair of decompositions for two input graphs for the Graph Isomorphism problem; we only need that the set of output decompositions is isomorphic invariant.

To cope with this new task, let us look at the known treewidth approximation algorithms; in particular, the arguably simplest one - the Robertson-Seymour algorithm. This algorithm provides a constant approximation in  time, where  is the treewidth of the input -vertex graph . It is a recursive procedure that decomposes a part  of the input graph, given a requirement that a set  should be contained in the top bag of the output decomposition. Think of  as an interface to the rest of the graph; we will always have , but for technical reasons the set  could be larger. During the course of the algorithm we maintain an invariant .

In the leaves of the recursion we have , and we output a single bag . If  is small, say, , then we can add an arbitrary vertex to  and recurse. The interesting things happen when  grows beyond the threshold .

As the graph is of treewidth , but , in an optimal decomposition of  the set  is spread across many bags. A simple argument shows that there exists a bag , , such that every component of  contains at most  vertices of . The algorithm takes  as a root bag (note that ), and recurses on every connected component of ; formally, for every connected component  of , we recurse on  with . Since every component of  contains at most  vertices of , in every recursive call the new set  is of size at most .

In the above algorithm, there are two steps that are very deeply not isomorphic-invariant: "take an arbitrary vertex to " in the case of small , and "take any such " in the case of large . We start fixing the algorithm in the second, more interesting case.

Before we start, let us think how we can find such a set  - the argument provided two paragraphs ago only asserted its existence. One of the ways to do it is to iterate through all disjoint sets  of size exactly  each, and check if the minimum separator between  and  in  is of size at most  (such a separator can contain vertices of  or , these are deletable as well). If such a separator is found, it is a good candidate for the set  - it does not necessarily have the property that every connected component of  contains at most half of the vertices of , but the claim that in the recursive calls the size of the set  decreased is still true. Of course, in this approach we have two not isomorphic-invariant choices: the choice of  and , and the choice of the minimum separator .

Luckily, an (almost) isomorphic-invariant way to make the second choice has already been well understood: by submodularity of cuts, there exists a well-defined notion of the minimum cut closest to , and the one closest to . But how about the choice of  and ?

The surprising answer is: it works just fine if we just throw into the constructed bag all separators for every choice of  and . That is, we start with the tentative bag , we iterate over all such (ordered) pairs , and throw into the constructed bag  the minimum cut between  and  that is closest to , as long as it has size at most . We recurse on all connected components of  as before; the surprising fact is that the sizes of the sets  in the recursive calls did not grow. To prove this fact, we analyse what happens to the components of  when you add  cuts one-by-one; a single step of this induction turns out to be an elementary application of submodularity. Furthermore, note that if the size of  is bounded in terms of , so is the set of the bag : .

Having made isomorphic-invariant the "more interesting" case of the Robertson-Seymour approximation algorithm, let us briefly discuss the case of the small set : we need to develop an isomorphic-invariant way to grow this set. Here, we need some preprocessing: by known tricks, we can assume that (a) the input graph does not contain any clique separators, and (b) if , then the minimum cut between  and  is of size at most . Furthermore, observe that in our algorithm, contrary to the original Robertson-Seymour algorithm, we always maintain the invariant that  and  is connected. Thus,  is never a clique in , and, if , we can throw into a bag the minimum cut between  and  that is closest to , for every ordered pair  where , .

In this way, the lack of clique separators ensures that we always make a progress in the algorithm, by "eating" at least one vertex of . The bound on the size of minimum cut between two nonadjacent vertices ensures that the bag created in this way is of size , and so are the sets  in the recursive calls. Consequently, our algorithm computes in an isomorphic-invariant way a tree decomposition with adhesions of size  and bags of size .

Well, there is one small catch - neither of the aforementioned steps is good to start the algorithm, to provide the initial call of the recursion; the Robertson-Seymour approximation starts with just  and . Here, we can start with  and  for every pair  of nonadjacent vertices, as in this call the step with the small set  will work fine. But this produces not a single isomorphic-invariant tree decomposition, but a family of  decompositions - which is still fine for the isomorphism test.

A cautious reader would also notice that the described algorithm results in a double-exponential dependency on the parameter  in the running time, contrary to the claim of the  term. Getting down to this dependency requires a bit more technical work; we refer to the full version of our paper for details.