Research and Advances
Architecture and Hardware Research highlights

Dissection: A New Paradigm For Solving Bicomposite Search Problems

Posted
  1. Abstract
  2. 1. Introduction
  3. 2. Representing Rubik's Cube as a Bicomposite Search Problem
  4. 3. The Basic Dissection Techniques
  5. 4. Improved Dissection Techniques
  6. 5. Conclusion
  7. Acknowledgments
  8. References
  9. Authors
  10. Footnotes
  11. Figures
Read the related Technical Perspective
Rubik's cube

Combinatorial search problems are usually described by a collection of possible states, a list of possible actions which map each current state into some next state, and a pair of initial and final states. The algorithmic problem is to find a sequence of actions which maps the given initial state into the desired final state. In this paper, we introduce the new notion of bicomposite search problems, and show that they can be solved with improved combinations of time and space complexities by using a new algorithmic paradigm called dissection. To demonstrate the broad applicability of our new paradigm, we show how to use it in order to untangle Rubik’s cube and to solve a typical NP-complete partition problem with algorithms which are better than any previously described algorithm for these problems.

Back to Top

1. Introduction

A central problem in the design of efficient algorithms is how to solve search problems, in which we are given a pair of states and a collection of possible actions, and we are asked to find how to get from the first state to the second state by performing some sequence of actions. In some cases, we only want to decide whether such a sequence exists at all, while in other cases it is clear that such sequences exist but we are asked to find the shortest possible sequence. Many search problems of this type have associated decision problems which are NP-complete, and thus we do not expect to find any polynomial time algorithms which can solve all their instances. However, what we hope to find are new exponential time algorithms whose exponents are smaller than in the best previously known algorithms. For example, the problem of breaking a cryptographic scheme whose key has n = 100 unknown bits cannot be solved in a practical amount of time via an exhaustive key search algorithm, since its time complexity of 2n would be beyond reach even for the largest currently available data center. However, if we manage to find a better cryptanalytic attack whose running time is 2n/2, we can break the scheme with a modest effort in spite of the exponential nature of this complexity function. One trick which is often helpful in such situations is to find a tradeoff between the time and space complexities of the attack: Exhaustive search requires a lot of time but a negligible amount of memory, and thus a tradeoff which uses more memory (in the form of large tables of precomputed values) in order to reduce the time (by skipping many computational steps) will be very beneficial. For reasons which are explained in the extended version of this paper (available in Dinur et al.2), we usually consider the product of the amount of time and the amount of space required by the algorithm as the appropriate complexity measure that we try to minimize. In the example above, breaking the cryptosystem with T = 2n time and an S = 1 space is infeasible, breaking it with T = 22n/3 time and S = 2n/3 space (whose product TS = 2n is the same as before) is better but still barely feasible, and breaking it in T = 2n/2 time and S = 2n/4 space (whose product TS = 23n/4 has a smaller exponent) is completely feasible.

A typical search problem is defined by a condition F (e.g., in the form of a CNF Boolean formula which is a conjunction of clauses) and a candidate solution X (e.g., in the form of a 0/1 assignment to all the variables in F), and the goal is to find among all the possible X at least one that satisfies the condition that F(X) is true. Such a representation has no internal structure in the sense that it uses the full description of F and the full value of X in order to decide whether F(X) is satisfied. However, we can usually replace the all-or-nothing choice of X by a sequence of smaller decisions. For example, we can start with the assignment of 0 to all the variables, and at each stage we can decide to flip the current value of one of the Boolean variables. At any intermediate point in this process F is in a state in which some of its clauses are satisfied and some are not, and our goal is to reach a state in which all the clauses are simultaneously satisfied. More generally, we say that a search problem is composite if solving it can be described by a sequence of atomic actions, which change the system from some initial state through a series of intermediate states until it reaches some final desired state. Most search problems have such a composite structure, which can be represented by the execution matrix described in Figure 1. The rows of this matrix represent states S0, S1, …, and the solution X is represented by the sequence of actions a1, a2, … on the left side of the figure in which action ai changes state Si−1 to state Si. In many cases, the atomic actions are invertible operations over the states, which makes it possible to map Si−1 to Si by using the action ai, and to map Si back to Si−1 by using the action. For example, in the case of Boolean formulas we are allowed to flip the current value of any one of the variables in X, but we can cancel its effect by flipping the same variable a second time, so in this case the action and its inverse happen to be the same. In this paper we consider only such invertible cases, which allow to split the search problem by applying some forward actions to the initial state, applying some inverse actions to the final state, and searching for a meet-in-the-middle (MITM) state which combines these parts. Our new dissection technique can be applied even when some of the actions are not invertible, but this makes its description more complicated and its complexity slightly higher, as discussed in Dinur et al.2

So far we have partitioned the execution matrix into multiple rows by dividing the solution process into a series of atomic actions. In order to apply our new dissection technique, we also have to partition the execution matrix into multiple columns by dividing each state Si into smaller chunks Si, j which we call substates, as described in Figure 2. However, only partitions in which the substates can be manipulated independently of each other will be useful to us. We say that a search problem has a bicomposite structure if it can be described by an execution matrix in which the knowledge of the action ai makes it possible to uniquely determine substate Si, j from Si−1, j and Si−1, j from Si, j, even when we know nothing about the other substates in the matrix. This immediately implies that if we choose any rectangle of any dimensions within the execution matrix, knowledge of the substates Si−1,j, Si−1,j+1, …, Si−1,k, along its top edge and knowledge of the actions ai, ai+1, …, a cacm5710_e.gif to its left suffices in order to compute the substates S cacm5710_e.gif ,j, S cacm5710_e.gif ,j+1, …, S cacm5710_e.gif ,k, along its bottom edge, and vice versa.

Not every search problem has such a bicomposite representation, but as we demonstrate in this paper there are many well-known problems which can be represented in such a way. When a problem is bicomposite, we can solve it with improved efficiency by using our generic new technique. Our main observation is that in such cases, we can improve the standard MITM algorithms (which try to match the forward and backward directions at a full intermediate state) by considering algorithms which only partially match the two directions at a partially specified intermediate state. In addition, we can reverse the logic of MITM algorithms, and instead of trying to converge from both ends of the execution toward some intermediate state which happens to be the same, we can start by partially guessing this intermediate state in all possible ways, and for each guessed value we can break the problem into two independent subproblems by proceeding from this intermediate state toward the two ends (exploiting the fact that in bicomposite problems we can determine the effect of long sequences of actions on partially specified states). We can then solve each one of these subproblems recursively by partially guessing additional substates in the execution matrix. We call this approach a dissection algorithm since it resembles the process in which a surgeon makes a sequence of short cuts at various strategically chosen locations in the patient’s body in order to carry out the operation. For example, in Section 4 we show how to solve the well-known combinatorial partition problem by first guessing only two-sevenths of the bits of the state which occur after performing three-sevenths of the actions, and then solving one of the resultant subproblems by guessing in addition one-seventh of the bits of the state which occurs after performing five-sevenths of the actions.

Our main purpose in this paper is to introduce the new ideas by applying them in the simplest possible way to several well-known search problems. We intentionally overlook several nuisance issues whose proper handling is not easy to explain, and ignore several possible optimizations which can further reduce the complexity of our algorithms. When we analyze the running time of our algorithms, we often assume for the sake of clarity that the instance we try to solve is randomly chosen, and that the intermediate states we try to guess are uniformly distributed.

The paper is organized as follows. In Section 2 we describe the problem of solving Rubik’s cube with the smallest possible number of steps as a bicomposite search problem, and in Section 3 we show how to solve it with time complexity which is approximately the square root of the size of the search space, and a space complexity which is approximately the fourth root of the size of the search space by using the simplest version of the new dissection algorithm. In Section 4 we describe several improvements of the basic dissection technique, and show how to use them in order to solve a different search problem called the “combinatorial partition problem” with combinations of time and space complexities which could not be achieved by any previously published algorithm, and which are more suitable for large scale FPGA-based hardware. We conclude the paper with some remarks in Section 5.

Back to Top

2. Representing Rubik’s Cube as a Bicomposite Search Problem

In this section we show how to construct a bicomposite representation of the well-known problem of solving a standard 3 × 3 × 3 Rubik’s cube.6 We can assume that we always hold the cube in a fixed orientation, in which the white center color is at the top, the yellow center color is on the left, etc. One of the 27 subcubes is at the center of the cube, and we can ignore it since it is completely invisible. The six subcubes at the center of each face are not moved when we rotate that (or any other) face, and thus we can ignore them as well in our state representation. The actions we can take are to rotate each one of the six faces of the cube by 90, 180, or 270 degrees (we are not allowed to rotate a center slice since this will change the standard orientation of the cube defined above). Consequently, we have a repertoire of 18 atomic actions we can apply to each state of the cube. Note that all these actions are invertible mappings on the state of Rubik’s cube in the sense that the inverse of a 90 degree rotation is a 270 degree rotation applied to the same face, and both of them are available as atomic actions.

Among the 27 − 6 − 1 = 20 subcubes which we can move, 12 have two visible colors and are called edge subcubes, and 8 have three visible colors and are called corner subcubes (Figure 3). Each such subcube can be uniquely described by the combination of colors on it, such as a blue-white (BW) edge subcube or a green-orange-red (GOR) corner subcube. In addition, each location on the cube can be described by its relevant sides (i.e., a combination of top/bottom, left/right, front/back). We can thus describe any state of the cube by a vector of length 20, whose ith entry describes the current location of the ith subcube (e.g., the first entry in the vector will always refer to the blue-green edge subcube, and specify that it is currently located at the top-front position). To complete the specification, we also have to choose some standard orientation of the colors, and note that edge subcubes can be either in the standard (e.g., BW) or in an inverted (e.g., WB) state, and each corner subcube can be in one of three possible orientations (e.g., GOR, ORG, or RGO). Note that any possible action can only move edge subcubes to edge subcubes and corner subcubes to corner subcubes. If we use the first 12 positions in the state vector to describe the current locations of the 12 edge subcubes (in some fixed order), then each entry in these positions can be described by a number between 1 and 24 (specifying in which one of the 12 possible positions it is currently located and in which one of its 2 possible orientations). Similarly, when we use the last 8 positions in the state vector to describe the current locations of the 8 corner subcubes (in some fixed order), then each one of these entries can again contain a number between 1 and 24, this time specifying in which one of the 8 possible positions it is located and in which of its 3 possible orientations. We can thus describe any state of Rubik’s cube by a vector of length 20 whose entries are numbers between 1 and 24. Any one of the 18 atomic actions will change 8 of the entries in this vector, by moving 4 edge subcubes and 4 corner subcubes to new positions and orientations, leaving the remaining 12 entries unchanged.

The problem of solving Rubik’s cube can now be formalized as the following search problem: We are given a first vector of length 20 (representing the initial scrambled state of the cube) and a second vector of length 20 (representing the standard unscrambled state of the cube). We would like to find a sequence of atomic actions (in the form of face rotations) that will change the first vector into the second vector. Finding some sequence is easy, and there are many algorithms which are described in the recreational mathematics literature to achieve this (some of which are described in Slocum6). However, these algorithms are typically quite wasteful, using between 50 and 100 actions in order to move subcubes to their correct positions one at a time in some fixed order, ignoring the effect of these actions on other subcubes. Some algorithms use fewer actions, but then they are much harder to describe since they require a very detailed case analysis of the full state before choosing the first action. Recently, Davidson et al.1 proved that 20 actions are necessary and sufficient if we want to solve any solvable state of Rubik’s cube, but this was an existential result and the authors did not present an efficient way to actually find such a sequence. Note that the number of possible sequences of 20 atomic actions is 1820 = 12748236216396078174437376 ≈ 283, but we can slightly reduce the size of the search space to 18 × 1519 = 399030807609558105468750 ≈ 278 by noticing that there is no point in rotating the same face twice in a row. However, even this reduced size cannot be exhaustively searched in a feasible amount of time.

To show that our representation of the search problem is bicomposite, assume that we know the current location and orientation of a particular subcube (namely, we know the value of Si−1, j in the execution matrix as a number between 1 and 24), and we apply to it some known face rotation action. We can then uniquely determine the new location and orientation of that particular subcube (namely, the value of Si, j) even when we know nothing about the current location and orientation of any other subcube (namely, all the other Sk, cacm5710_e.gif values in the execution matrix). Notice that many other natural representations of the states of Rubik’s cube do not have such a bicomposite structure. For example, if we associate the first entry in the state vector with a particular cube position (such as top-front) and use it to denote which edge subcube (such as BW) is currently located in it and in which orientation, then knowledge of just this entry in the first state does not tell us anything about which edge subcube (such as GR) replaces it at the top-front position if we rotate the top face by 90 degrees. Such a representation requires knowledge of other columns in the execution matrix, depending on which action was applied to the state, and thus we cannot use it in our new dissection technique.

As shown in Section 3, we can use the bicomposite representation in order to find for any given initial state of Rubik’s cube a sequence of up to 20 face rotations using a completely feasible combination of a time complexity which is the square root of the size of the search space (namely, in about 239 steps, or a few minutes on a standard PC) and a space complexity which is about the fourth root of this number (namely, about 219.5 memory locations, or a few megabytes). The resultant algorithm is completely generic, makes no use of the details of the problem besides its bicompositeness, and matches the complexities of the best previous algorithm for solving Rubik’s cube (designed about 25 years ago, see Fiat et al.3) which was highly specific and depended on the group-theoretic properties of the set of permutations defined by Rubik’s cube.

Back to Top

3. The Basic Dissection Techniques

We now assume that we are given an initial state vector S0 and a final state vector S cacm5710_e.gif of Rubik’s cube, and our goal is to find a series of atomic actions a1, a2, …, a cacm5710_e.gif that transform the initial state into the final state. As described in the previous section, we know that cacm5710_e.gif = 20 suffices to find a solution, and hence our goal is to find a1, a2, …, a20.

Our dissection algorithms are extensions of the classical MITM algorithm, which was first presented in 1974 by Horowitz and Sahni5 in order to solve the Knapsack problem. We can apply a MITM algorithm to almost any composite problem with invertible actions. When the size of the search space is 2n, the MITM algorithm requires about 2n/2 time and 2n/2 space. For the sake of completeness, we describe below how to apply this algorithm in the context of Rubik’s cube, whose search space has about 278 states. In this case, a time complexity of 278/2 = 239 is feasible, but a space complexity of 278/2 = 239 random access memory locations is too expensive.

The complete details of the algorithm are given in Figure 4. The first step of the algorithm is to iterate over all possible 20/2 = 10 action sequences a1, …, a10. We have 18 × 159 ≈ 239 such sequences, and for each one, we apply its actions to S0 and obtain a possible value for S10. We store the 239 values of a1, …, a10 in a list next to the corresponding value of S10, and sorta the list according to S10. Next, we iterate over all possible action vectors a11, …, a20. Again, there are about 239 such action vectors, and for each one, we apply its inverted actions to S20 and obtain a possible value for S10. We now search the sorted list for this value of S10, and for each match, we obtain the corresponding value of a1, …, a10 from the list and output a1, a2, …, a20 as a solution.

The MITM algorithm requires about 239 memory cells in order to store the sorted list, and its time complexity is about 239, which is the time required in order to iterate over each one of the action vectors a1, …, a10 and a11, …, a20.

*  3.1. Improving the MITM algorithm using dissection

In this section, we show how to improve the classical MITM algorithm on Rubik’s cube by using a basic version of our new dissection technique. The main idea here is to “dissect” the execution matrix in the middle by iterating over all the possible values of some part of the middle state S10. The size of the partial S10 that we iterate on is chosen such that it contains about 2n/4 = 219.5 partial states. Since S10 is represented as a 20-entry vector, where each entry can attain 24 values, we choose to iterate on its first 4 entries, which can assume about 244 ≈ 218.5 values. For each such partial value of S10, we use the bicomposite structure of the problem in order to independently work on the two partial execution matrices shown in Figure 5 as red and blue rectangles, and finally join the partial solutions in order to obtain the full action vector.

The complete details of the algorithm are given in Figure 6. We have an outer loop which iterates over all the possible values of S10,1, S10,2, S10,3, S10,4. Assuming that this value is correctly guessed, we first concentrate on the upper part of the execution matrix and find all the partial action vectors a1, …, a10 which transform S0,1, S0,2, S0,3, S0,4 into S10,1, S10,2, S10,3, S10,4. This is done using a simple MITM algorithm on this smaller execution matrix. For each solution a1, …, a10 that we obtain using the MITM algorithm, we apply its actions to the full state S0, obtain candidate values of the full S10 state, and store it next to a1, …, a10 in a list. After the MITM algorithm finishes populating the list, we sort it (e.g., in some lexicographic order) according to the value of S10.

We now focus on the bottom execution matrix and find all the partial action vectors a11, …, a20 which transform S10,1, S10,2, S10,3, S10,4 into S20,1, S20,2, S20,3, S20,4. We use the same idea that we used for the upper part, that is, we execute a MITM algorithm on the bottom execution matrix. For each solution a11, …, a20 that we obtain, we apply its inverse actions to S20 and obtain a value for S10. Then, we check for matches for S10 in the sorted list, and for each match, we output a full solution a1, a2, …, a20.

In order to analyze the algorithm, we fix a value of S10,1, S10,2, S10,3, S10,4 and estimate the average number of solutions that we expect for the upper (smaller) execution matrix. Namely, we calculate the expected number of action vectors a1, …, a10 which transform S0,1, S0,2, S0,3, S0,4 into S10,1, S10,2, S10,3, S10,4. First, we notice that the number of possible action vectors a1, …, a10 is about 239. Each such action vector transforms S0,1, S0,2, S0,3, S0,4 into an arbitrary partial state which matches S10,1, S10,2, S10,3, S10,4 with probability of about 1/(244) ≈ 2−18.5 (which is inverse-proportional to the number of possible values of S10,1, S10,2, S10,3, S10,4). Thus, the expected number of solution (that we store in our sorted list) is 239 × 2−18.5 = 220.5.

In general, the time complexity of the MITM algorithm is about square root of the search space, and thus its time complexity on the upper execution matrix is about 239/2 = 219.5. However, since in this case we could not split the problem into two parts of exactly equal sizes, we expect 220.5 solutions (which we enumerate and store), and thus its time complexity is slightly increased to 220.5. This is also the expected time complexity of the MITM algorithm on the bottom part (although here we do not store the solutions, but immediately check each one of them). Since we have an outer loop which we execute 244 ≈ 218.5 times, the expected time complexity of the full algorithm is about 218.5+20.5 = 239. The expected memory complexity is 220.5, required in order to store the solutions for the MITM on the upper part (note that we reuse this memory for each guess of S0,1, S0,2, S0,3, S0,4).

Back to Top

4. Improved Dissection Techniques

A closer look at the algorithm presented in Section 3 reveals that the algorithm treats the top and bottom parts of the execution matrix differently. Indeed, while the suggestions from the top part are stored in a table (L10 in the example of Figure 6), the suggestions from the bottom part are checked on-the-fly against the table values. As a result, while the number of suggestions in the top part is bounded from above by the size of the memory available for the algorithm, the number of suggestions from the bottom part can be arbitrarily large and generated on-the-fly in an arbitrary order.

This suggests that an asymmetric division of the execution matrix, in which the bottom part is significantly bigger than the top part, may lead to better performance of the algorithm.

In this section we show that this is indeed the case. As the algorithm for untangling the Rubik’s cube presented in Section 3 is already practical on a PC so that there is no significant value in further improving it, we choose another classical search problem, known as the combinatorial partition problem to be our running example in this section.

The problem is defined as follows. We are given a set of n integers, U = {x1, x2, …, xn}. Our goal is to partition U into two complementary subsets U1, U2 whose elements sum up to the same number, that is,

eq01.gif

The combinatorial partition problem is known to be NP-complete,4 and hence, one cannot expect a sub-exponential solution in general. Nevertheless, there are various techniques which allow to find a solution efficiently in various cases, especially when there exist many partitions (U1, U2) which satisfy Equation (1). We thus consider the “hardest” instance of the problem, in which each of the xis is of n digits in binary representation (i.e., xi ≈ 2n). In this case, at most a few solutions (U1, U2) are expected to exist, and no sub-exponential algorithms for the problem are known. For the sake of simplicity, we focus on the modular variant of the problem, in which Equation (1) is slightly altered to

eq02.gif

As a specific numeric example, consider the case n = 112. In this case, checking all 2112 possible partitions is, of course, completely infeasible. The standard MITM algorithm allows to reduce the time complexity to 256, but it increases the space complexity to 256 which is currently infeasible. As we show below, the problem can be represented as a bicomposite problem, and hence, the technique of Section 3 can be applied to obtain the better tradeoff of T = 256 and S = 228. While these numbers are almost practical, the relatively large amount of required memory disallows the use of FPGAs in the computation, which makes it barely feasible. We show below that by an asymmetric variant of the dissection algorithm, we are able to obtain the complexities T = 264 (which is only 28 times larger) and S = 216 (which is 212 times smaller), which allow for a significantly faster computation using memory-constrained FPGAs. Note that the asymmetric dissection algorithm outperforms the symmetric one by a factor of 16 according to the complexity measure S × T (the complexities are 280 vs. 284). Such a factor, while not extremely big, can make a difference in practical scenarios.

*  4.1. Representing combinatorial partition as a bicomposite search problem

In order to apply dissection algorithms to the combinatorial partition problem, we have to find a way to represent it as a bicomposite search problem.

First, we represent it as a composite problem. We treat the problem of choosing the partition (U1, U2) as a sequence of n atomic decisions, where the ith decision is whether to assign xiU1 or xiU2. We introduce a counter C which is initially set to zero, and then at the ith step, if the choice is xiU1 then C is replaced by C + xi (mod 2n), and if the choice is xiU2, C is replaced by Cxi (mod 2n). Note that the value of C after the nth step is cacm5710_b.gif (mod 2n), and hence, the sequence of choices leads to the desired solution if and only if the final value of C is zero.

In this representation, the partition problem has all the elements of a composite problem: an initial state (Cinitial = 0), a final state (Cfinal = 0), and a sequence of n steps, such that in each step, we have to choose one of two possible atomic actions. Our goal is to find a sequence of choices which leads from the initial state to the final state. In terms of the execution matrix, we define Si to be the value of C after the ith step (which is an n-bit binary number) and ai to be the action transforming Si−1 to Si, whose possible values are either CC + xi (mod 2n) or CCxi (mod 2n).

The second step is to represent the problem as a bicomposite problem. The main observation we use here is the fact that for any two integers a, b, the mth least significant bit (LSB) of a + b (mod 2n) depends only on the m LSBs of a and b (and not on their other digits). Hence, if we know the m LSBs of Si−1 and the action ai, we can compute the m LSBs of Si.

Using this observation, we define Si, j to be the jth LSB of Si. This leads to an n-by-n execution matrix Si, j for i, j ∈ 1, 2, …, n with the property that if we choose any rectangle within the execution matrix which includes the rightmost column of the matrix, knowledge of the substates cacm5710_c.gif along its top edge and knowledge of the actions ai, ai+1, …, a cacm5710_e.gif to its right suffices in order to compute the substates cacm5710_d.gif along its bottom edge.

Note that the condition satisfied by our execution matrix is weaker than the condition given in the definition of a bicomposite problem, since in our case, the “rectangle” property holds only for rectangles of a certain kind and not for all rectangles. However, as we show in the next subsection, even this weaker property is sufficient for applying all dissection algorithms we present.b

*  4.2. Dissection algorithm for the combinatorial partition problem

The basic idea in the algorithm is to divide the state matrix into seven (!) parts of n/7 steps each, where three parts belong to the top part St and four parts belong to the bottom part Sb. The partition is obtained by enumerating the 2n/7 LSBs of the state S3n/7.

For each value v of these bits, we perform a simple MITM algorithm in the top part, which yields about 23n/7 × 2−2n/7 = 22n/7 possible combinations of actions a1, a2, …, a3n/7 which lead to a state S3n/7 whose 2n/7 LSBs equal to the vector v. For each of these combinations, we compute the full value of the state S3n/7. The resulting values of S3n/7 are stored in a table, along with the corresponding combinations of a1, a2, …, a3n/7.

Then we consider the bottom part, and apply to it the dissection algorithm described in Section 3 (thus, dividing it into four chunks of n/7 steps each). This results in 24n/7 × 2−2n/7 = 22n/7 possible combinations of actions a(3n/7)+1, a(3n/7)+2, …, an which lead (in the inverse direction) to a state S3n/7 whose 2n/7 LSBs equal to the vector v. For each of these combinations, we compute the full value S3n/7 and compare it to the values in the table. If a match is found, this means that the corresponding sequences {a1, a2, …, a3n/7} and a(3n/7)+1, a(3n/7)+2, …, an match to yield a solution of the problem. Note that if a solution exists, then our method must find it, since it actually goes all over all possible combinations of actions (though, in a sophisticated way). The pseudocode of the algorithm is given in Figure 7.

The memory complexity of the algorithm is O(2n/7), as both the standard MITM algorithm for the top part and the dissection algorithm for the bottom part have this complexity. (For the bottom algorithm, the complexity is (24n/7)1/4 = 2n/7.)

The time complexity is 24n/7. Indeed, the enumeration in the state S3n/7 is performed over 22n/7 values, both the standard MITM algorithm for the top part and the dissection algorithm for the bottom part require 22n/7 time, and the remaining 22n/7 possible combinations of a(3n/7)+1, a(3n/7)+2, …, an are checked instantly. This leads to time complexity of 22n/7 × 22n/7 = 24n/7.

In the special case of n = 112, each of the seven chunks consists of 16 steps, the enumeration is performed on the 28 LSBs of the state S42, the memory complexity is 2n/7 = 216, and the time complexity is 24n/7 = 264.

*  4.3. Advanced dissection algorithms

The algorithms presented in Section 3 and in this section are the two simplest dissection algorithms, which demonstrate the general idea behind the technique. In the extended version of the paper,2 we present more advanced dissection algorithms, which include division of the matrix to “exotic” numbers of parts, such as 11 and 29, and show the optimality of such choices within our general framework.

So far we only considered search algorithms which are not allowed to fail (i.e., if there are any solutions to the problem then our algorithm will always find all of them, but its running time may be longer than expected if the instances are not randomly chosen or if the number of solutions is too large). In Dinur et al.,2 we also consider algorithms which may fail to find a solution with a small probability, and show how to improve the efficiency of our algorithms in this case by combining them with a classical technique called parallel collision search, devised by Wiener and van Oorschot in 1996.7

Back to Top

5. Conclusion

In this paper we introduced the notion of bicomposite search problems, and developed new types of algorithmic techniques called dissection algorithms in order to solve them with improved time and space complexities. We demonstrated how to use these techniques by applying them to two standard types of problems (Rubik’s cube and combinatorial partitions). However, some of the most exciting applications of these techniques are in cryptanalysis, which is beyond the scope of this paper. For example, many banks are still using a legacy cryptographic technique called triple-DES, which encrypts sensitive financial data by encrypting it three times with three independent keys. A natural question is whether using quadruple-DES (which encrypts the data four times with four independent keys) would offer significantly more security due to its longer key and more complicated encryption process. By using our new dissection techniques, we can show the surprising result that finding the full key of quadruple-DES could be achieved with essentially the same time and space complexities as finding the full key of the simpler triple-DES encryption scheme, and thus there is no significant security advantage in upgrading triple-DES to quadruple-DES.

Back to Top

Acknowledgments

The second author (O.D.) was supported in part by the Israel Science Foundation through grant No. 827/12 and in part by the German-Israeli Foundation for Scientific Research and Development through grant No. 2282-2222.6/2011. The third author (N.K.) was supported by the Alon Fellowship.

Back to Top

Back to Top

Back to Top

Back to Top

Figures

F1 Figure 1. An execution matrix of a composite search problem.

F2 Figure 2. An execution matrix of a bicomposite search problem.

F3 Figure 3. Rubik’s cube.

F4 Figure 4. Meet-in-the-middle procedure to solve the Rubik’s cube.

F5 Figure 5. Dissection of the Rubik’s cube execution matrix.

F6 Figure 6. Solving the Rubik’s cube using dissection into 4.

F7 Figure 7. Solving the partitioning problem using dissection into 7.

Back to top

    1. Davidson, M., Dethridge, J., Kociemba, H., Rokicki, T. God's number is 20, 2010. http://cube20.org.

    2. Dinur, I., Dunkelman, O., Keller, N., Shamir, A. Efficient dissection of composite problems, with applications to cryptanalysis, knapsacks, and combinatorial search problems. In CRYPTO, R. Safavi-Naini and R. Canetti, eds. Volume 7417 of Lecture Notes in Computer Science (2012). Springer, 719–740.

    3. Fiat, A., Moses, S., Shamir, A., Shimshoni, I., Tardos, G. Planning and learning in permutation groups. In FOCS. IEEE Computer Society, 1989, 274–279.

    4. Garey, M.R., Johnson, D.S. Computers and Intractability: A Guide to the Theory of NP-Completeness. W. H. Freeman, 1979.

    5. Horowitz, E., Sahni, S. Computing partitions with applications to the knapsack problem. J. ACM 21, 2 (1974), 277–292.

    6. Slocum, J. The Cube: The Ultimate Guide to the World's Bestselling Puzzle—Secrets, Stories, Solutions. Black Dog & Leventhal Publishers, 2011.

    7. van Oorschot, P.C., Wiener, M.J. Improving implementable meet-in-the-middle attacks by orders of magnitude. In CRYPTO, N. Koblitz, ed. Volume 1109 of Lecture Notes in Computer Science (1996). Springer, 229–236.

    a. For the sake of simplicity, we ignore logarithmic factors in our complexity analysis, and thus we assume that sorting is a linear time operation.

    b. For the sake of simplicity, we disregard the issue of carries. We note that in order to know all the carries required to execute our dissection algorithms, we guess the values of Si, j from the least significant bit to the most significant bit. For more details, refer to the extended version of this paper.2

    The previous version of this paper was published in CRYPTO, vol. 7417 (2012) Lecture Notes in Computer Science. Springer, 719–740.

Join the Discussion (0)

Become a Member or Sign In to Post a Comment

The Latest from CACM

Shape the Future of Computing

ACM encourages its members to take a direct hand in shaping the future of the association. There are more ways than ever to get involved.

Get Involved

Communications of the ACM (CACM) is now a fully Open Access publication.

By opening CACM to the world, we hope to increase engagement among the broader computer science community and encourage non-members to discover the rich resources ACM has to offer.

Learn More