scispace - formally typeset
Search or ask a question
Journal ArticleDOI

The graph isomorphism disease

01 Dec 1977-Journal of Graph Theory (John Wiley & Sons, Ltd)-Vol. 1, Iss: 4, pp 339-363
TL;DR: The present state of the art of isomorphism testing is surveyed, its relationship to NP-completeness is discussed, and some of the difficulties inherent in this particularly elusive and challenging problem are indicated.
Abstract: The graph isomorphism problem—to devise a good algorithm for determining if two graphs are isomorphic—is of considerable practical importance, and is also of theoretical interest due to its relationship to the concept of NP-completeness. No efficient (i.e., polynomial-bound) algorithm for graph isomorphism is known, and it has been conjectured that no such algorithm can exist. Many papers on the subject have appeared, but progress has been slight; in fact, the intractable nature of the problem and the way that many graph theorists have been led to devote much time to it, recall those aspects of the four-color conjecture which prompted Harary to rechristen it the “four-color disease.” This paper surveys the present state of the art of isomorphism testing, discusses its relationship to NP-completeness, and indicates some of the difficulties inherent in this particularly elusive and challenging problem. A comprehensive bibliography of papers relating to the graph isomorphism problem is given.
Citations
More filters
Journal ArticleDOI
TL;DR: Traces as mentioned in this paper is a graph isomorphism algorithm based on the refinement-individualization paradigm, and it is implemented in several of the key implementations of the program nauty.

1,602 citations

Journal ArticleDOI
TL;DR: This work presents improved partition refinement algorithms for three problems: lexicographic sorting, relational coarsest partition, and double lexical ordering that uses a new, efficient method for unmerging two sorted sets.
Abstract: We present improved partition refinement algorithms for three problems: lexicographic sorting, relational coarsest partition, and double lexical ordering. Our double lexical ordering algorithm uses a new, efficient method for unmerging two sorted sets.

1,267 citations

Proceedings ArticleDOI
29 Nov 2001
TL;DR: The empirical results show that the algorithm scales linearly with the number of input transactions and it is able to discover frequent subgraphs from a set of graph transactions reasonably fast, even though it has to deal with computationally hard problems such as canonical labeling of graphs and subgraph isomorphism which are not necessary for traditional frequent itemset discovery.
Abstract: As data mining techniques are being increasingly applied to non-traditional domains, existing approaches for finding frequent itemsets cannot be used as they cannot model the requirement of these domains. An alternate way of modeling the objects in these data sets is to use graphs. Within that model, the problem of finding frequent patterns becomes that of discovering subgraphs that occur frequently over the entire set of graphs.The authors present a computationally efficient algorithm for finding all frequent subgraphs in large graph databases. We evaluated the performance of the algorithm by experiments with synthetic datasets as well as a chemical compound dataset. The empirical results show that our algorithm scales linearly with the number of input transactions and it is able to discover frequent subgraphs from a set of graph transactions reasonably fast, even though we have to deal with computationally hard problems such as canonical labeling of graphs and subgraph isomorphism which are not necessary for traditional frequent itemset discovery.

1,181 citations


Cites background or methods from "The graph isomorphism disease"

  • ...To narrow down the search space, we first partition the vertices by their degrees and labels, which is a well-known technique called vertex invariants [ 18 ]....

    [...]

  • ...A canonical label is a unique code of a given graph [ 18 , 10]....

    [...]

Journal ArticleDOI
TL;DR: It is formally shown that the new distance measure is a metric, based on the maximal common subgraph of two graphs, which is superior to edit distance based measures in that no particular edit operations together with their costs need to be defined.

782 citations

Journal ArticleDOI
TL;DR: Theoretical Approach to Chemical Structure, Approximate Approaches versus Ambitious Computations, and Use of Signed Matrices.
Abstract: G. Clar 6n Rule versus Hückel 4n + 2 Rule 3464 H. Hydrocarbons versus Heteroatomic Systems 3465 IV. Hidden Treasures of Kekulé Valence Structures 3466 A. Conjugated Circuits 3467 B. Innate Degree of Freedom 3470 C. Clar Structures 3472 V. Graph Theoretical Approach to Chemical Structure 3473 A. Metric 3473 B. Chemical Graphs 3473 C. Isospectral Graphs 3473 D. Embedded Graphs 3475 E. Partial Ordering 3476 VI. On Enumeration of Benzenoid Hydrocarbons 3477 VII. Kekulé Valence Structures Count 3479 A. Non-branched Cata-condensed Benzenoids 3481 B. Branched Cata-condensed Benzenoids 3482 C. Benzenoid Lattices 3482 D. Peri-condensed Benzenoids 3483 E. Miscellaneous Benzenoids 3484 F. The Approach of Platt 3485 G. Computer Programs for Calculating K 3485 H. Transfer-Matrix Method 3486 I. Use of Recursion Relations 3486 J. Use of Signed Matrices 3487 VIII. Enumeration of Conjugated Circuits 3488 IX. Approximate Approaches versus Ambitious Computations 3490

664 citations

References
More filters
Book
01 Jan 1974
TL;DR: This text introduces the basic data structures and programming techniques often used in efficient algorithms, and covers use of lists, push-down stacks, queues, trees, and graphs.
Abstract: From the Publisher: With this text, you gain an understanding of the fundamental concepts of algorithms, the very heart of computer science. It introduces the basic data structures and programming techniques often used in efficient algorithms. Covers use of lists, push-down stacks, queues, trees, and graphs. Later chapters go into sorting, searching and graphing algorithms, the string-matching algorithms, and the Schonhage-Strassen integer-multiplication algorithm. Provides numerous graded exercises at the end of each chapter. 0201000296B04062001

9,262 citations

01 Jan 1972
TL;DR: Throughout the 1960s I worked on combinatorial optimization problems including logic circuit design with Paul Roth and assembly line balancing and the traveling salesman problem with Mike Held, which made me aware of the importance of distinction between polynomial-time and superpolynomial-time solvability.
Abstract: Throughout the 1960s I worked on combinatorial optimization problems including logic circuit design with Paul Roth and assembly line balancing and the traveling salesman problem with Mike Held. These experiences made me aware that seemingly simple discrete optimization problems could hold the seeds of combinatorial explosions. The work of Dantzig, Fulkerson, Hoffman, Edmonds, Lawler and other pioneers on network flows, matching and matroids acquainted me with the elegant and efficient algorithms that were sometimes possible. Jack Edmonds’ papers and a few key discussions with him drew my attention to the crucial distinction between polynomial-time and superpolynomial-time solvability. I was also influenced by Jack’s emphasis on min-max theorems as a tool for fast verification of optimal solutions, which foreshadowed Steve Cook’s definition of the complexity class NP. Another influence was George Dantzig’s suggestion that integer programming could serve as a universal format for combinatorial optimization problems.

7,714 citations

Proceedings ArticleDOI
03 May 1971
TL;DR: It is shown that any recognition problem solved by a polynomial time-bounded nondeterministic Turing machine can be “reduced” to the problem of determining whether a given propositional formula is a tautology.
Abstract: It is shown that any recognition problem solved by a polynomial time-bounded nondeterministic Turing machine can be “reduced” to the problem of determining whether a given propositional formula is a tautology. Here “reduced” means, roughly speaking, that the first problem can be solved deterministically in polynomial time provided an oracle is available for solving the second. From this notion of reducible, polynomial degrees of difficulty are defined, and it is shown that the problem of determining tautologyhood has the same polynomial degree as the problem of determining whether the first of two given graphs is isomorphic to a subgraph of the second. Other examples are discussed. A method of measuring the complexity of proof procedures for the predicate calculus is introduced and discussed.

6,675 citations

Journal ArticleDOI
TL;DR: In this article, the maximum number of cliques possible in a graph with n nodes is determined and bounds are obtained for the number of different sizes of clique possible in such a graph.
Abstract: A clique is a maximal complete subgraph of a graph. The maximum number of cliques possible in a graph withn nodes is determined. Also, bounds are obtained for the number of different sizes of cliques possible in such a graph.

907 citations