From e4ee1f8a1e5707a4d8478a678a03ce664ae8f894 Mon Sep 17 00:00:00 2001 From: Camil Staps Date: Mon, 2 Nov 2015 15:49:23 +0100 Subject: Working on report --- Practical1/report/algorithm.tex | 64 ++++++++------- Practical1/report/analysis.tex | 147 ++++++++++++++++++++++++++++++++--- Practical1/report/implementation.tex | 9 ++- Practical1/report/report.tex | 5 +- 4 files changed, 182 insertions(+), 43 deletions(-) (limited to 'Practical1/report') diff --git a/Practical1/report/algorithm.tex b/Practical1/report/algorithm.tex index 7ce0fdd..f6272e4 100644 --- a/Practical1/report/algorithm.tex +++ b/Practical1/report/algorithm.tex @@ -45,7 +45,7 @@ But this latter recurrence is clearly more efficient for \emph{large} $d(v)$. Th If there is an m.i.s. $ms$ with $w\in ms$, $ms\ex w\with v$ is an m.i.s. as well. \end{thmproof} -If we then take $w$ instead of $v$ and use \autoref{eq:basic-structure}, it may be the case that $v>w$. In this case, we may bypass \autoref{eq:basic-structure} by applying the recurrence $\ms(G)=\ms(G\ex w)$ of \autoref{lem:dominance}. After this, it may be the case that we can apply \autoref{eq:with-two-neighbours} instead, since $d(v)$ will have been decreased by one. +If we then take $w$ instead of $v$ and use \autoref{eq:basic-structure}, it may be the case that $v>w$. In this case, we may bypass \autoref{eq:basic-structure} by applying the recurrence $\ms(G)=\ms(G\ex w)$ of \autoref{lem:dominance}. After this, since the maximum degree of any vertex in our graph is $4$, we will be able to apply \autoref{eq:with-two-neighbours} instead, since $d(v)$ will have been decreased by one to at most $3$. This gives us \autoref{alg:with-best-vertex}, which combines \autoref{eq:basic-structure} and \ref{eq:with-two-neighbours}. @@ -111,27 +111,27 @@ For the case that $d(v)=2$ we write $w,w'$ for the neighbours of $v$. \REQUIRE $G=(V,E)$ \IF{$G$ has multiple strongly connected components} \STATE $C\gets$ some strongly connected component - \RETURN $\ms(C) + \ms(G\ex C)$ \COMMENT{\autoref{lem:components}} - \ELSIF{$|G|=1$} - \RETURN $|G|$ \COMMENT{\autoref{lem:base-case}} + \RETURN $\ms(C) + \ms(G\ex C)$ \COMMENT{\autoref{lem:components}}\label{alg:ms:case-component} + \ELSIF{$|G|\le1$} + \RETURN $|G|$ \COMMENT{\autoref{lem:base-case}}\label{alg:ms:case-g1} \ELSE \STATE $v\gets v\in V$ with minimal $d(v)$ \STATE $w\gets w\in N(v)$ with maximal $d(w)$ \IF{$d(v)=1$} - \RETURN $1+\ms(G\ex\iN(v))$ \COMMENT{\autoref{lem:robson-ms-d1}} + \RETURN $1+\ms(G\ex\iN(v))$ \COMMENT{\autoref{lem:robson-ms-d1}}\label{alg:ms:case-d1} \ELSIF{$d(v)=2$} \STATE $w'\gets$ the other neighbour of $v$ \IF{$e(w,w')$} - \RETURN $1+\ms(G\ex\iN(v))$ \COMMENT{\autoref{lem:robson-ms-d2-edge}} + \RETURN $1+\ms(G\ex\iN(v))$ \COMMENT{\autoref{lem:robson-ms-d2-edge}}\label{alg:ms:case-d2-edge} \ELSE - \RETURN $\max(2+\ms(G\ex\iN(w)\ex\iN(w')), 1+\ms^2(G\ex\iN(v),N^2(v)))$ \COMMENT{\autoref{lem:robson-ms-d2-noedge}} + \RETURN $\max(2+\ms(G\ex\iN(w)\ex\iN(w')), 1+\ms^2(G\ex\iN(v),N^2(v)))$ \COMMENT{\autoref{lem:robson-ms-d2-noedge}}\label{alg:ms:case-d2-noedge} \ENDIF \ELSIF{$d(v)=3$} - \RETURN $\max(\ms^2(G\ex v,N(v)), 1+\ms(G\ex\iN(v)))$ \COMMENT{\autoref{eq:with-two-neighbours}} + \RETURN $\max(\ms^2(G\ex v,N(v)), 1+\ms(G\ex\iN(v)))$ \COMMENT{\autoref{eq:with-two-neighbours}}\label{alg:ms:case-d3} \ELSIF{$v>w$} - \RETURN $\ms(G\ex w)$ \COMMENT{\autoref{lem:dominance}} + \RETURN $\ms(G\ex w)$ \COMMENT{\autoref{lem:dominance}}\label{alg:ms:case-dominance} \ELSE - \RETURN $\max(\ms(G\ex w), 1+\ms(G\ex\iN(w)))$ \COMMENT{\autoref{eq:basic-structure}} + \RETURN $\max(\ms(G\ex w), 1+\ms(G\ex\iN(w)))$ \COMMENT{\autoref{eq:basic-structure}}\label{alg:ms:case-otherwise} \ENDIF \ENDIF \end{algorithmic} @@ -145,6 +145,10 @@ For the case that $d(v)=2$ we write $w,w'$ for the neighbours of $v$. The algorithm follows the same basic structure. In any case that \autoref{alg:ms} is different, it considers less cases (yet it is still correct by \autoref{thm:ms-correct}). Therefore, it can only be more efficient. \end{thmproof} +\begin{thmproof}[ms-terminates]{theorem}{\autoref{alg:ms} terminates with every input, if $\ms^2$ terminates with a smaller input.} + There is a base case on line \ref{alg:ms:case-g1} which is not recursive and handles the graphs with the lowest $k$ possible orders ($k=2$). In every other case, the same algorithm is run on a smaller input (or $\ms^2$ on a smaller input), so inductively the algorithm terminates with any input. +\end{thmproof} + \subsection{The helper function $\ms^n$} \label{sec:algorithm:helper-function} So far we have been using $\ms^n(G,S)$ as the size of the maximum independent set of $G$ given that it contains at least $n$ vertices in $S$. This section is devoted to arguing an efficient implementation of this helper function. In the above, we have only used $\ms^2$. Therefore, we do not need to go into details about the case $n\neq2$. @@ -180,39 +184,39 @@ We will write $s_1,s_2,\dots$ for the elements of $S$. \label{alg:ms2} \begin{algorithmic}[1] \REQUIRE $G=(V,E), S\subseteq V$ - \IF{$|S|=1$} - \RETURN $0$ \COMMENT{\autoref{lem:helper-1}} + \IF{$|S|\le1$} + \RETURN $0$ \COMMENT{\autoref{lem:helper-1}}\label{alg:ms2:case-s1} \ELSIF{$|S|=2$} \IF{$e(s_1,s_2)$} - \RETURN $0$ \COMMENT{\autoref{lem:helper-2-edge}} + \RETURN $0$ \COMMENT{\autoref{lem:helper-2-edge}}\label{alg:ms2:case-s2-edge} \ELSE - \RETURN $2+\ms(G\ex\iN(s_1)\ex\iN(s_2))$ \COMMENT{\autoref{lem:helper-2-noedge}} + \RETURN $2+\ms(G\ex\iN(s_1)\ex\iN(s_2))$ \COMMENT{\autoref{lem:helper-2-noedge}}\label{alg:ms2:case-s2-noedge} \ENDIF \ELSIF{$|S|=3$} \IF{$\exists_i[d(s_i)=0]$} - \RETURN $1+\ms^1(G\ex s_i,S\ex s_i)$ \COMMENT{Special case of \autoref{lem:components}} + \RETURN $1+\ms^1(G\ex s_i,S\ex s_i)$ \COMMENT{Special case of \autoref{lem:components}}\label{alg:ms2:case-s3-d0} \ELSIF{$s_1,s_2,s_3$ form a three-cycle} - \RETURN $0$ \COMMENT{Impossible to pick two vertices from a three-cycle} + \RETURN $0$ \COMMENT{Impossible to pick two vertices from a three-cycle}\label{alg:ms2:case-s3-cycle} \ELSIF{$\exists_{i,j,k}[i\neq k \land e(s_i,s_j) \land e(s_i,s_k)]$} - \RETURN $2 + \ms(G\ex\iN(s_j)\ex\iN(s_k))$ \COMMENT{Impossible to choose $s_i$ and another vertex, so we have to choose $s_j$ and $s_k$} + \RETURN $2 + \ms(G\ex\iN(s_j)\ex\iN(s_k))$ \COMMENT{Impossible to choose $s_i$ and another vertex, so we have to choose $s_j$ and $s_k$}\label{alg:ms2:case-s3-path} \ELSIF{$\exists_{i,j}[e(s_i,s_j)]$} \STATE $s_k\gets S\ex s_1\ex s_j$ - \RETURN $1 + \ms^1(G\ex\iN(s_k), S\ex s_k)$ \COMMENT{Impossible to choose $s_i$ and $s_j$, so we choose $s_k$ and one of $s_i,s_j$} + \RETURN $1 + \ms^1(G\ex\iN(s_k), S\ex s_k)$ \COMMENT{Impossible to choose $s_i$ and $s_j$, so we choose $s_k$ and one of $s_i,s_j$}\label{alg:ms2:case-s3-edge} \ELSIF{$\exists_{i,j}[N(s_i)\cap N(s_j)\neq\emptyset]$} - \RETURN $\ms^2(G\ex(N(s_i)\cap N(s_j)), S)$ \COMMENT{m.i.s. contains either $s_i$ or $s_j$, so not their common neighbours} + \RETURN $\ms^2(G\ex(N(s_i)\cap N(s_j)), S)$ \COMMENT{m.i.s. contains either $s_i$ or $s_j$, so not their common neighbours}\label{alg:ms2:case-s3-intersection} \ELSIF{$\exists_i[d(s_i)=1]$} - \RETURN $1+\ms^1(G\ex\iN(s_i),S\ex s_i)$ \COMMENT{Special case of \autoref{lem:one-neighbour}} + \RETURN $1+\ms^1(G\ex\iN(s_i),S\ex s_i)$ \COMMENT{Special case of \autoref{lem:one-neighbour}}\label{alg:ms2:case-s3-d1} \ELSE - \RETURN $\max(1+\ms^1(G\ex\iN(s_1),S\ex s_1), 2 + \ms^2(G\ex\iN(s_2)\ex\iN(s_3)\ex s_1, N(s_1)))$ \COMMENT{Either the m.i.s. does contain $s_1$, and one of $s_2,s_3$, or it doesn't. In the latter case, it must contain $s_2,s_3$, and at least two on $s_1$'s neighbours, using an argument similar to \autoref{lem:two-neighbours}. Robson misses the `$2+$' for this second case, but from the writing it is clear that it should be included.} + \RETURN $\max(1+\ms^1(G\ex\iN(s_1),S\ex s_1), 2 + \ms^2(G\ex\iN(s_2)\ex\iN(s_3)\ex s_1, N(s_1)))$ \COMMENT{Either the m.i.s. does contain $s_1$, and one of $s_2,s_3$, or it doesn't. In the latter case, it must contain $s_2,s_3$, and at least two on $s_1$'s neighbours, using an argument similar to \autoref{lem:two-neighbours}. Robson misses the `$2+$' for this second case, but from the writing it is clear that it should be included.}\label{alg:ms2:case-s3-otherwise} \ENDIF \ELSIF{$|S|=4$} \IF{$\lnot\exists_i[d(s_i)\leq3]$} - \RETURN $\max(1+\ms(G\ex\iN(s_1)),\ms^2(G\ex s_1,S\ex s_1))$ \COMMENT{m.i.s. either does or doesn't contain $s_1$} + \RETURN $\max(1+\ms(G\ex\iN(s_1)),\ms^2(G\ex s_1,S\ex s_1))$ \COMMENT{m.i.s. either does or doesn't contain $s_1$}\label{alg:ms2:case-s4-no-dle3} \ELSE - \RETURN $\ms(G)$ \COMMENT{More efficient than the $\max$ in this case; \autoref{lem:helper-general}} + \RETURN $\ms(G)$ \COMMENT{More efficient than the $\max$ in this case; \autoref{lem:helper-general}}\label{alg:ms2:case-s4-dle3} \ENDIF \ELSE - \RETURN $\ms(G)$ \COMMENT{No clever optimisation found; \autoref{lem:helper-general}} + \RETURN $\ms(G)$ \COMMENT{No clever optimisation found; \autoref{lem:helper-general}}\label{alg:ms2:case-otherwise} \ENDIF \end{algorithmic} \end{algorithm} @@ -221,6 +225,10 @@ We will write $s_1,s_2,\dots$ for the elements of $S$. This follows from \autoref{lem:helper-general} through \ref{lem:helper-2-noedge} and \autoref{lem:one-neighbour} and \ref{lem:components}. We have only made small efficiency enhancements that are clearly correct and are argued in the definition of the algorithm. \end{thmproof} +\begin{thmproof}[ms2-terminates]{theorem}{\autoref{alg:ms2} terminates for any input, if $\ms$ and $\ms^1$ terminate with a smaller input.} + There are base cases for the smallest cases, $|S|$-wise, that terminate directly. Every other case terminates directly or reduces to termination of $\ms$ on a smaller input $|G|$-wise, $\ms^1$ on a smaller input $|G|$-wise, or $\ms^2$ on a smaller input $|G|$ or $|S|$-wise. By induction, the algorithm terminates with any input. +\end{thmproof} + \bigskip As the observant reader may have noticed, we use $\ms^1$ in the definition of $\ms^2$. We therefore also need to write an efficient algorithm to apply this function. Of course, \autoref{lem:helper-general} and \ref{lem:helper-1} apply to $\ms^1$ as well. Note, that we always call $\ms^1(\cdots,S)$ with $|S|=2$. We will therefore not give a complete algorithm $\ms^1$, but one that works in the necessary cases. @@ -248,10 +256,10 @@ If the conditions of \autoref{lem:helper-ms1-2-2} are met, write $w_1,w_2$ for t This leads us to a slightly adapted version of $\ms^2$ that we can use for $\ms^1$. Pseudocode is given in \autoref{alg:ms1}. \begin{algorithm}[h!] - \caption{Finding the size of the m.i.s. of a graph, given that it should contain one node from a certain subgraph} + \caption{Finding the size of the m.i.s. of a graph, given that it should contain one node from a certain subgraph with two nodes} \label{alg:ms1} \begin{algorithmic}[1] - \REQUIRE $G=(V,E), S\subseteq V, d(s_1)\leq d(s_2)$ + \REQUIRE $G=(V,E), S\subseteq V, S=\{s_1,s_2\}, d(s_1)\leq d(s_2)$ \IF{$d(s_1)\leq1$} \RETURN $\ms(G)$ \COMMENT{\autoref{lem:helper-general}} \label{alg:ms1:case-d1} \ELSIF{$e(s_1,s_2)$} @@ -281,5 +289,9 @@ This leads us to a slightly adapted version of $\ms^2$ that we can use for $\ms^ This follows from \autoref{lem:helper-general} and \autoref{lem:helper-ms1-2-2} through \ref{lem:helper-ms1-2-2-dominance}. We have only made enhancements that are clearly correct and furthermore argued in the definition of the algorithm. \end{thmproof} +\begin{thmproof}[ms1-terminates]{theorem}{\autoref{alg:ms1} terminates with any input if the second argument has size $2$.} + Inductively by \autoref{thm:ms-terminates} and \ref{thm:ms2-terminates}. +\end{thmproof} + If we assume to have implementations of the more obvious graph algorithms such as the neighbours of a vertex, we now have by \autoref{thm:ms-correct} a correct implementation of the maximum independent set problem. diff --git a/Practical1/report/analysis.tex b/Practical1/report/analysis.tex index d796f96..aae5697 100644 --- a/Practical1/report/analysis.tex +++ b/Practical1/report/analysis.tex @@ -1,10 +1,12 @@ \section{Complexity analysis} \label{sec:analysis} -Robson provides us with a very detailed analysis proving that the $\ms$ algorithm runs in $\pazocal O(\mathrm{polynomial}(n)\cdot2^{cn})$, where $c<0.296$ and with $n$ the number of vertices of the input graph \cite{robson}. However, the argument is over my head, so I will provide my own, simpler, less exact analysis here. +Robson provides us with a very detailed analysis proving that the $\ms$ algorithm runs in $\pazocal O(\mathrm{polynomial}(n)\cdot2^{cn})$, where $c<0.296$ and with $n$ the number of vertices of the input graph \cite{robson}. However, the argument goes over my head, so I will provide my own, simpler, less exact analysis here. We write $T_{\ms}(n)$ for the running time of $\ms$ on a graph $G$ with $|G|=n$. For $\ms^n$ we write $T_{\ms^n}(p,q)$ for the appliance of $\ms^n$ on a graph $G$ with $|G|=p$ and a set $S$ to pick $n$ vertices from with $|S|=q$. In the case of $n=1$, we write $T_{\ms^1}(p)$ instead of $T_{\ms^1}(p,2)$, because we only use the case of $q=2$. +For the sake of brevity, I will use but not discuss here that basic graph operations such as dominance checking, edge checking, degree computation and others run in $\pazocal O(\mathrm{polynomial}(n))$. Informally, this can be checked by looking at the functions in the Java code accompanying this report. + \subsection{$\ms^1$: \autoref{alg:ms1}} \label{sec:analysis:ms1} @@ -13,17 +15,19 @@ This function is a simple case distinction. Let's therefore consider the complex \begin{table}[h] \centering \renewcommand{\arraystretch}{1.3} - \begin{tabular}{l | p{29mm} | p{19mm} | p{50mm}} + \begin{tabular}{r | p{29mm} | p{19mm} | p{50mm}} Ln. & Case & Complexity & Argumentation \\\hline\hline \ref{alg:ms1:case-d1} & $d(s_1)\leq 1$ - & $T_{\ms}(n-1)$ - & Simple recurrence \\\hline + & $T_{\ms}(n)$;\newline + $T_{\ms}(n-1)+\pazocal O(n)$ + & Simple recurrence; see below\\\hline \ref{alg:ms1:case-edge-d4} & $e(s_1,s_2), d(s_1)>3$ & $2T_{\ms}(n-5)$ & Both vertices have degree four, so we run $\ms$ twice on a graph with five vertices less\\\hline \ref{alg:ms1:case-edge-d3} & $e(s_1,s_2), d(s_1)\leq3$ - & $T_{\ms}(n)$ - & Simple recurrence \\\hline + & $T_{\ms}(n)$;\newline + $3T_{\ms}(n-2)$ + & Simple recurrence; see below\\\hline \ref{alg:ms1:intersection} & $N(s_1)\cap N(s_2)\neq\emptyset$ & $T_{\ms^1}(n-1)$ & At least one vertex removed \\\hline @@ -40,10 +44,18 @@ This function is a simple case distinction. Let's therefore consider the complex & $T_{\ms}(n-3) + T_{\ms^2}(n-4,4)$ & In the first case at least three vertices removed, in the second case four, and at most four to choose two from \end{tabular} - \caption{Worst-case time complexity analysis of the different cases in \autoref{alg:ms1}} + \caption{Worst case time complexity analysis of the different cases in \autoref{alg:ms1}} \label{tab:analysis-ms1} \end{table} +The case in line \ref{alg:ms1:case-d1} is problematic in the sense that the size of the graph is not reduced. However, we know that $d(s_1)\le1$, which means that there is either a strongly connected component of size $1$, namely $\{s_1\}$, or that there is a $v$, namely $s_1$, with $d(v)=1$. In the first case, the time complexity will reduce to $T_{\ms}(1)+T_{\ms}(n-1)=T_{\ms}(n-1) + \pazocal O(n)$, as we will see below in \autoref{tab:analysis-ms}. In the second case, the time complexity reduces to $T_{\ms}(n-2) < T_{\ms}(n-1) + \pazocal O(n)$. + +We need to do something similar for the case in line \ref{alg:ms1:case-edge-d3}. In this case, we know there is a vertex $v$ with $d(v)\le3$. This is broken down into a number of cases in \autoref{alg:ms}, of which the time complexity may be found in \autoref{tab:analysis-ms} below. It is easy to see that all of the relevant cases have a running time less than $T_{\ms}(n-2)+T_{\ms^2}(n-3,6)+T_{\ms^2}(n-1,3)$. Since $T_{\ms^2}(n-3,\dots)4$ is in line \ref{alg:ms:case-d2-noedge}. In \autoref{tab:analysis-ms} below we see that this particular case $3$ vertices were already removed from the graph. We will use this below in the analysis of \autoref{alg:ms}, line \ref{alg:ms:case-d2-noedge} in \autoref{tab:analysis-ms}. + +\begin{thmproof}[ms2-le-ms]{lemma}{There exists a $c$ for which for all graphs $G$ and subgraphs $S$ with $|S|\le4$, we have $T_{\ms^2}(|G|,|S|)\le T_{\ms}(n-2)$.} + Every case in \autoref{tab:analysis-ms2} suffices. Cases with complexity $T_{\ms^1}(n-x)$ suffice by \autoref{lem:ms1-le-ms}. +\end{thmproof} \subsection{$\ms$: \autoref{alg:ms}} \label{sec:analysis:ms} -%todo +Also this is a simple case distinction. Let's again put the different cases in a table, along with their line numbers. See \autoref{tab:analysis-ms}. + +\begin{table}[h] + \centering + \renewcommand{\arraystretch}{1.3} + \begin{tabular}{r | p{28mm} | p{20mm} | p{50mm}} + Ln. & Case & Complexity & Argumentation \\\hline\hline + \ref{alg:ms:case-component} & Multiple strongly connected components + & $T_{\ms}(|C|) + T_{\ms}(n-|C|)$ + & $|C|$ is the size of some strongly connected component\\\hline + \ref{alg:ms:case-g1} & $|G|\le1$ + & $\pazocal O(n)$ + & Counting the number of vertices\\\hline + \ref{alg:ms:case-d1} & $d(v)=1$ + & $T_{\ms}(n-2)$ + & $v$ and its neighbour removed\\\hline + \ref{alg:ms:case-d2-edge} & $d(v)=2$, $e(w,w')$ + & $T_{\ms}(n-3)$ + & $v$ and its two neighbours removed\\\hline + \ref{alg:ms:case-d2-noedge} & $d(v)=2$, otherwise + & $T_{\ms}(n-6) + T_{\ms^2}(n-3,6)$; \newline + $T_{\ms}(n-6) + T_{\ms}(n-3)$ + & $d(w)$ and $d(w')\ge2$; $N^2(v)$ is at most $6$ for any degree-2 vertex in a graph with maximum degree $4$; as we saw in \autoref{sec:analysis:ms2}, this reduces to the second sum\\\hline + \ref{alg:ms:case-d3} & $d(v)=3$ + & $T_{\ms}(n-4) + T_{\ms^2}(n-1,3)$; \newline + $T_{\ms}(n-4) + T_{\ms}(n-3)$ + & Simple recurrences; second form by \autoref{lem:ms2-le-ms}\\\hline + \ref{alg:ms:case-dominance} & $v>w$ + & $T_{\ms}(n-1)$; \newline + $T_{\ms}(n-3) + T_{\ms}(n-5)$ + & Simple recurrence; see below\\\hline + \ref{alg:ms:case-otherwise} & Otherwise + & $T_{\ms}(n-1) + T_{\ms}(n-5)$ + & $d(w)\ge d(v)\le4$ so $|\iN(w)|\le5$ + \end{tabular} + \caption{Worst case time complexity analysis of the different cases in \autoref{alg:ms}} + \label{tab:analysis-ms} +\end{table} + +We saw already in \autoref{sec:algorithm:good-v:dominance} that after removing $w$ in the case of dominance of $v$ over $w$ on line \ref{alg:ms:case-dominance}, we will have a vertex with degree $3$. Therefore, the recurrence $T_{\ms}(n-1)$ in this case reduces to one of the above cases with $n:=n-1$. This is at most $T_{\ms}(n-3)+T_{\ms}(n-5)$. + +We can do something similar in the `Otherwise' case. We can only be in this case if there is no vertex with degree less than $4$. Since all vertices have a maximum degree of $4$, this means that every vertex has degree $4$. In both subcases that are considered in line \ref{alg:ms:case-otherwise} of \autoref{alg:ms}, vertices are removed from the graph, leaving a graph with a vertex of maximum degree $3$. If we use that knowledge, we see from the other cases in \ref{tab:analysis-ms} that the time complexity of the `Otherwise' case can be rewritten as $T_{\ms}(n-3)+T_{\ms}(n-5)+T_{\ms}(n-7)+T_{\ms}(n-9)$. + +\begin{thmproof}[otherwise-half-of-cases]{lemma}{Of any $k$ consequent applications of $\ms$, at most $\frac k2$ will be using the inefficient `Otherwise' case.} + After applying the `Otherwise' case, there will be at least one vertex with degree lower than $4$. In the iteration directly after applying the `Otherwise' case, we don't need to apply it. +\end{thmproof} + +\begin{thmproof}[ms-complexity]{theorem}{\autoref{alg:ms} runs in $\pazocal O\left(2^{n/2}\right)$.} + We see that the time complexity of $\ms(n)$ depends on $\ms(n-c)$ with $c\ge2$. Furthermore, if $c=2$, we have $\ms(n)=\ms(n-2)$, so if $\ms(n)$ depends on $\ms(n-2)$, it depends on that \emph{alone}, giving a branching factor of $1$. Branching only occurs when $c\ge3$. Only in the `Otherwise' case, the branching factor is $4$, in the other cases it is $2$. + + We then consider the search tree $T$ where we compress nodes with branching factor $1$. By \autoref{lem:otherwise-half-of-cases}, only on one of two consecutive levels can we have a branching factor of $4$ (and in the worst case this will happen as much as possible). If we then combine the levels with branching factor $4$ with the levels directly below them (having branching factor $2$), we find a search tree $T'$ with in the worst case a constant branching factor of $8$. + + The original search tree $T$ has at most $\frac n3$ levels, since branching only occurs when $c\ge3$. Therefore, the compressed tree $T'$ has, in the worst case, $\frac n6$ levels (note, that this case is worse than the case of a tree with constant branching factor $2$ and $\frac n3$ levels, and therefore also worse than any case in between). + + In every vertex of the search tree $T'$ we need polynomial time, since we assumed that basic graph operations run in polynomial time. Therefore, \autoref{alg:ms} runs in $$\pazocal O\left(\left(8\cdot\mathrm{polynomial}(n)\right)^{n/6}\right) = \pazocal O\left(\mathrm{polynomial}(n)^{n/2}\right) = \pazocal O\left(2^{n/2}\right).$$ +\end{thmproof} + +This result is, naturally, coherent with Robson's result that the algorithm runs in $\pazocal O\left(\mathrm{polynomial}(n)\cdot2^{cn}\right)$ with $c<0.296$ \cite{robson}. We have proven that the bound holds with $c\le\frac12$. + +\subsection{Space complexity} +\label{sec:analysis:space} +A graph is stored as a list of nodes, and for every edge both nodes hold one reference. Therefore, the space complexity of a single graph $G=(V,E)$ is $\pazocal O(|V|+|E|)$. + +As discussed in \autoref{sec:implementation}, we're using a remove-and-restore strategy. For every recurrence, no matter what the branching factor, we do not need to create new objects: the whole algorithm runs on one object. This happens at the cost of using a stack of restore information, that are lists of vertices that are removed from the graph. Since at any moment a node will either be in the graph \emph{or} on that stack (and then only once so), the space complexity of the graph object is still $\pazocal O(|V|+|E|)$. + +There is one small exception to this: the \texttt{findStronglyConnectedComponent} method does create a new object. Space-wise, the worst case is then the case of a graph consisting of $|V|$ strongly connected components. In this case, we need $|V|$ graph objects, giving a quadratic worst case space complexity of $\pazocal O\left(|V|\cdot(|V|+|E|)\right)$. This is not taking into regard the garbage collector. A closer look at the implementation of the \texttt{maximumIndependentSetSize} method will show that in fact at most two graph objects are needed at any point in the program. In any case, thanks to the efficiency gained by applying \autoref{lem:components}, the algorithm runs very fast on graphs with many strongly connected components, so we do not need to worry about the theoretically quadratic space complexity. diff --git a/Practical1/report/implementation.tex b/Practical1/report/implementation.tex index 5c6bbff..6df8932 100644 --- a/Practical1/report/implementation.tex +++ b/Practical1/report/implementation.tex @@ -3,7 +3,7 @@ A Java application accompanies this report. I have chosen for a simple implementation of a vertex: -\begin{minted}[fontsize=\footnotesize,linenos,breaklines,tabsize=0]{java} +\begin{minted}[fontsize=\footnotesize,linenos,tabsize=0]{java} class Node implements Comparable> { private final T content; private final List> neighbours = new ArrayList<>(); @@ -39,7 +39,7 @@ The graph therefore has a stack of lists of nodes as a private member. Every ele That gives us the signature of the \texttt{Graph} class: -\begin{minted}[fontsize=\footnotesize,linenos,breaklines,tabsize=0]{java} +\begin{minted}[fontsize=\footnotesize,linenos,tabsize=0]{java} class Graph { private final List> nodes; private final Stack>> restorePoints = new Stack<>(); @@ -74,10 +74,11 @@ The \texttt{addNode} and \texttt{getNode} methods are simply wrappers for the \t Using these two classes in the problem at hand turns out to be as simple as: -\begin{minted}[fontsize=\footnotesize,linenos,breaklines,tabsize=0]{java} +\begin{minted}[fontsize=\footnotesize,linenos,tabsize=0]{java} Graph graph = new Graph(); int n_bins = readGraph(new Scanner(System.in), graph); - System.out.println(graph.maximumIndependentSetSize() >= n_bins ? "possible" : "impossible"); + int miss = graph.maximumIndependentSetSize(); + System.out.println(miss >= n_bins ? "possible" : "impossible"); \end{minted} A suitable implementation of \texttt{readGraph} is outside the scope of this report. diff --git a/Practical1/report/report.tex b/Practical1/report/report.tex index d13a648..2172329 100644 --- a/Practical1/report/report.tex +++ b/Practical1/report/report.tex @@ -48,6 +48,9 @@ \usepackage{algorithmic} \providecommand*{\algorithmautorefname}{Algorithm} +\renewcommand*{\subsectionautorefname}{subsection} +\renewcommand*{\subsubsectionautorefname}{subsection} + \usepackage{minted} \title{Miss Algorithm} @@ -76,7 +79,7 @@ \section{Acknowledgements} \label{sec:acknowledgements} -I'm grateful to Tom van der Zanden for pointing me to the term `maximum independent set', which eventually led me to Robson's algorithm \cite{robson}. +I'm grateful to Tom van der Zanden for pointing me to the term `maximum independent set' when discussing what I until then called a `colouring maximising problem'. The term `maximum independent set' eventually led me to Robson's algorithm \cite{robson}. \begin{thebibliography}{9} \bibitem{robson}{Robson, J.M. (1985). \emph{Algorithms for Maximum Independent Sets}.} -- cgit v1.2.3