- Kruskal's algorithm
**Kruskal's algorithm**is analgorithm ingraph theory that finds aminimum spanning tree for a connected weighted graph. This means it finds a subset of the edges that forms a tree that includes every vertex, where the total weight of all the edges in the tree is minimized. If the graph is not connected, then it finds a "minimum spanning forest" (a minimum spanning tree for each connected component). Kruskal's algorithm is an example of agreedy algorithm .It works as follows:

* create a forest "F" (a set of trees), where each vertex in the graph is a separate tree

* create a set "S" containing all the edges in the graph

* while "S" isnonempty

** remove an edge with minimum weight from "S"

** if that edge connects two different trees, then add it to the forest, combining two trees into a single tree

** otherwise discard that edgeAt the termination of thealgorithm , the forest has only one component and forms a minimum spanning tree of the graph.This algorithm first appeared in "

Proceedings of the American Mathematical Society ", pp. 48–50 in1956 , and was written byJoseph Kruskal .Other algorithms for this problem include

Prim's algorithm ,Reverse-Delete algorithm , andBorůvka's algorithm .**Performance**Where "E" is the number of edges in the graph and "V" is the number of vertices, Kruskal's algorithm can be shown to run in "O"("E" log "E") time, or equivalently, "O"("E" log "V") time, all with simple data structures. These running times are equivalent because:

* "E" is at most "V"^{2}and $log\; V^2\; =\; 2\; log\; V$ is "O"(log "V").

* If we ignore isolated vertices, which will each be their own component of the minimum spanning tree anyway, "V" ≤ "E"+1, so log "V" is "O"(log "E").We can achieve this bound as follows: first sort the edges by weight using a

comparison sort in "O"("E" log "E") time; this allows the step "remove an edge with minimum weight from "S" to operate in constant time. Next, we use adisjoint-set data structure to keep track of which vertices are in which components. We need to perform O("E") operations, two 'find' operations and possibly one union for each edge. Even a simple disjoint-set data structure such as disjoint-set forests with union by rank can perform O("E") operations in "O"("E" log "V") time. Thus the total time is "O"("E" log "E") = "O"("E" log "V").Provided that the edges are either already sorted or can be sorted in linear time (for example with

counting sort orradix sort ), the algorithm can use more sophisticated disjoint-set data structures to run in "O"("E" α("V")) time, where α is the extremely slowly-growing inverse of the single-valuedAckermann function .**Example****Proof of correctness**The proof consists of two parts. First, it is proved that the algorithm produces a spanning tree. Second, it is proved that the constructed spanning tree is of minimal weight.

**panning tree**Let $P$ be a connected, weighted graph and let $Y$ be the subgraph of $P$ produced by the algorithm. $Y$ cannot have a cycle, since the last edge added to that cycle would have been within one subtree and not between two different trees. $Y$ cannot be disconnected, since the first encountered edge that joins two components of $Y$ would have been added by the algorithm. Thus, $Y$ is a spanning tree of $P$.

**Minimality**The proof is by "

reductio ad absurdum ". Assume that $Y$ is not a minimal spanning tree and among all minimum weight spanning trees pick $Y\_1$ which has the smallest number of edges which are not in $Y$. Consider the edge $e$ which was first to be added by the algorithm to $Y$ of those which are not in $Y\_1$.$Y\_1\; cup\; \{e\}$ has a cycle. Being a tree, $Y$ cannot contain all edges of this cycle. Therefore this cycle contains an edge $f$ which is not in $Y$. The graph $Y\_2=Y\_1\; cup\; \{e\}\; setminus\; \{f\}$ is also a spanning tree and therefore its weight cannot be less than the weight of $Y\_1$ and hence the weight of "e" cannot be less than the weight of "f". $Y\_2$By another application of the "reductio ad absurdum" we shall prove that the weight of "f" cannot be less than that of "e". Assume the contrary and remember that the edges are considered for addition to $Y$ in the order of non-decreasing weight. Therefore "f" would have been considered in the main loop before "e", i.e., it would be tested for the addition to a subforest $Y\_3\; subset\; Ycap\; Y\_1$ (recall that "e" is the first edge of $Y$ which is not in $Y\_1$). But "f" does not create a cycle in $Y\_1$, therefore it cannot create a cycle in $Y\_3$, and it would have been added to the growing tree.

The above implies that the weights of "e" and "f" are equal, and hence $Y\_2$ is also a minimal spanning tree. But $Y\_2$ has one more edge in common with $Y$ than $Y\_1$, which contradicts to the choice of $Y\_1$, "

Q.E.D ".**Pseudocode**1

**function**Kruskal("G") 2**for each**vertex "v" in "G" do 3 Define an elementary cluster "C"("v") ← {"v"}. 4 Initialize a priority queue "Q" to contain all edges in "G", using the weights as keys. 5 Define a tree "T" ← Ø //"T" will ultimately contain the edges of the MST 6 // n is total number of vertices 7**while**"T" has fewer than "n"-1 edges**do**8 // edge u,v is the minimum weighted route from/to v 9 ("u","v") ← "Q".removeMin() 10 // prevent cycles in T. add u,v only if T does not already contain an edge consisting of u and v. 11 // Note that the cluster contains more than one vertex only if an edge containing a pair of 12 // the vertices has been added to the tree. 13 Let "C"("v") be the cluster containing "v", and let "C"("u") be the cluster containing "u". 14**if**"C"("v") ≠ "C"("u")**then**15 Add edge ("v","u") to "T". 16 Merge "C"("v") and "C"("u") into one cluster, that is, union "C"("v") and "C"("u"). 17**return**tree "T"**References*** Joseph. B. Kruskal: " [

*http://links.jstor.org/sici?sici=0002-9939(195602)7%3A1%3C48%3AOTSSSO%3E2.0.CO%3B2-M On the Shortest Spanning Subtree of a Graph and the Traveling Salesman Problem*] ". In: "Proceedings of the American Mathematical Society", Vol 7, No. 1 (Feb, 1956), pp. 48–50

*Thomas H. Cormen ,Charles E. Leiserson ,Ronald L. Rivest , andClifford Stein . "Introduction to Algorithms ", Second Edition. MIT Press and McGraw-Hill, 2001. ISBN 0-262-03293-7. Section 23.2: The algorithms of Kruskal and Prim, pp.567–574.

* Michael T. Goodrich and Roberto Tamassia. "Data Structures and Algorithms in Java", Fourth Edition. John Wiley & Sons, Inc., 2006. ISBN 0-471-73884-0. Section 13.7.1: Kruskal's Algorithm, pp.632.**ee also***

Reverse-Delete algorithm

*Borůvka's algorithm

*Dijkstra's algorithm **External links*** [

*http://students.ceid.upatras.gr/~papagel/project/kruskal.htm Animation of Kruskal's algorithm (Requires Java plugin)*]

* [*http://www.cut-the-knot.org/Curriculum/Games/Mazes.shtml Create and Solve Mazes by Kruskal's and Prim's algorithms*] atcut-the-knot

* [*http://www-b2.is.tokushima-u.ac.jp/~ikeda/suuri/kruskal/Kruskal.shtml Minimum Spanning Tree Problem: Kruskal's Algorithm*]

*Wikimedia Foundation.
2010.*