584
Views
5
CrossRef citations to date
0
Altmetric
ORIGINAL ARTICLES

Exact and approximate algorithms for part-machine clustering based on a relationship between interval graphs and Robinson matrices

&
Pages 925-935 | Received 01 Sep 2006, Accepted 01 Feb 2007, Published online: 26 Sep 2007

Abstract

An important manufacturing cell formation problem requires permutations of the rows (parts) and columns (machines) of a part-machine incidence matrix such that the reordered matrix exhibits a block-diagonal form. Numerous objective criteria and algorithms have been proposed for this problem. In this paper, a new perspective is offered that is based on the relationship between the consecutive ones property associated with interval graphs and Robinson structure within symmetric matrices. This perspective enables the cell formation problem to be decomposed into two permutation subproblems (one for rows and one for columns) that can be solved optimally using dynamic programming or a branch-and-bound algorithm for matrices of nontrivial size. A simulated annealing heuristic is offered for larger problem instances. Results pertaining to the application of the proposed methods for a number of problems from the literature are presented.

1. Introduction

Group technology provides a framework for accommodating a diverse collection of products and activities associated with a manufacturing process. An especially noteworthy application of group technology is cellular manufacturing, which requires the identification of families of parts that are assigned to production cells consisting of groups of machines. The cell formation problem is a well-studied problem in the operations research and industrial engineering literature and encompasses a variety of issues that have been outlined carefully in a comprehensive review by CitationSelim et al. (1998). As observed by CitationShargal et al. (1995), an important initial step in the design of a cellular manufacturing system is the clustering of the part-machine matrix that contains important processing information within the system.

In this paper, we focus on the problem of inducing a block-diagonal form for binary part-machine matrices. A value of “one” (“zero”) in a part-machine matrix indicates that the part associated with the row requires (does not require) processing on the machine corresponding to the column. A block-diagonal form is established by permuting the rows and columns of the part-machine matrix such that the “ones” in the matrix group together along the diagonal. CitationShargal et al. (1995) indicate that methods for permuting part-machine matrices are typically based on one of three models: (i) minimum spanning trees (CitationMcCauley, 1972); (ii) the traveling salesman problem (CitationAskin et al., 1991); and (iii) the linear placement problem (CitationSuryanarayanan et al., 1991). Each of these methods attempts to produce a block-diagonal form via the independent establishment of permutations for parts and machines and the subsequent blending of these permutations.

Our goal in this paper is to present an alternative method for obtaining a block-diagonal form that is based on the consecutive ones property associated with interval graphs and the corresponding relationship to Robinson matrices. Accordingly, our approach is based on the pioneering work of CitationFulkerson and Gross (1965) and CitationKendall (1969) which, to the best of our knowledge, has not previously been addressed in the cell formation literature. A graph G = (V, E), with vertex set V and edge set E, is an interval graph if an interval (Δ v) on the real number line can be generated for each vertex, vV, such that the condition {u, v}∉ E ↔ Δ u ∩ Δ v = ø holds. CitationFulkerson and Gross (1965) proved that consecutive ones in the columns of the incidence matrix corresponding to G = (V, E) is a necessary and sufficient condition for a graph to be an interval graph. CitationKendall (1969) subsequently showed that if a matrix possesses consecutive ones in all columns, then the product of that matrix with its transpose must exhibit a perfect Robinson structure, whereby elements of the matrix are “never increasing” when moving away from the main diagonal in any direction (CitationRobinson, 1951).

Why are the consecutive ones property and Robinson matrices relevant to cell formation? The importance of the consecutive ones property stems from the well-established practice of permuting the rows and columns of the part-machine matrix to obtain a block-diagonal form (CitationShargal et al., 1995; CitationCrama and Oosten, 1996; CitationSarker and Khan, 2001). A block-diagonal form is characterized by cells that consist primarily of ones along the main diagonal and primarily of zeros away from the main diagonal. The ones that appear outside the blocks along the main diagonal are termed “exceptional elements”, and such elements correspond to bottleneck machines and exceptional parts (CitationMalakooti and Yang, 2002). Likewise, the occurrence of zeros within the blocks along the main diagonal produces an adverse effect on machine utilization within the corresponding cell (CitationChandrasekharan and Rajagopalan, 1986). The consecutive ones property is in accordance with the goal of block-diagonal form, minimization of exceptional parts, and maximization of machine utilization. If each block along the main diagonal consists of consecutive ones in all rows and columns of the block, and if there are no ones outside the block (i.e., consecutive zeros outside the main diagonal blocks), then the objectives of minimizing exceptional elements and maximizing utilization are satisfied perfectly. Further support of the importance of the consecutive ones property to cell formation is offered by the fact that all of the clustering efficiency measures for block-diagonal form used by CitationShargal et al. (1995) during their search for algorithms for part-machine clustering are based principally on counts of adjacent ones (and sometimes adjacent zeros) within rows and columns.

The relevance of the consecutive ones property to block-diagonal form and part-machine clustering establishes one aspect of the motivation for our paper. The fact that a permuted part-machine matrix with consecutive ones in all rows and columns must yield part and machine similarity matrices with a perfect Robinson structure provides the remainder of the motivation. Our premise is that permuting the part and machine similarity matrices to have at least an approximate Robinson structure will have a propensity to yield a reordered part-machine matrix with large bands of consecutive ones and block-diagonal form. We develop some effective criteria for producing a Robinson structure and demonstrate that exact procedures based on dynamic programming and a branch-and-bound algorithm are feasible for problems of nontrivial size. For larger part-machine matrices, we develop a simulated annealing heuristic. We apply these methods to a number of part-machine matrices in the published literature.

In Section 2 of this paper, we more formally describe the relationship between the consecutive ones property and Robinson matrices. An illustrative example is also presented in Section 2. Section 3 describes the dynamic programming, branch and bound, and simulated annealing methods used to obtain permutations for parts and machines based on a Robinson structure. Computational results for problems in the cell formation literature are reported in Section 4, and the paper concludes with a brief summary in Section 5.

2. The consecutive ones property and Robinson matrices

2.1. Basic definitions

We define A = [a ij ] as an n × m incidence matrix for n parts and m machines, where a ij = 1 if part i requires processing on machine j (for 1 ≤ slantislantn and 1 ≤ slantjslantm). The n × n matrix AA T is a symmetric similarity matrix for parts, where an element in row i and column i′ of the matrix represents the number of machines that parts i and i′ have in common. Similarly, the m × m matrix A T A is a symmetric similarity matrix for machines, where an element in row j and column j′ of the matrix represents the number of parts that machines j and j′ have in common.

We denote φ = (φ (1), φ (2),…, φ (n)) as a permutation of the n parts such that φ (k) is the part in position k of the permutation. An n × n permutation matrix for parts, Λ = [λ ij ], is constructed from φ by defining, for 1 ≤ slantislantn, λ iφ (i) = 1 and λ ij = 0 for 1 ≤ slantj ≠ φ (i) ≤ slantn. The set of all n¡ permutations of parts is represented as Φ (φ ∈ Φ). We define γ = (γ (1), γ (2), …, γ (m)) as a permutation of the m machines such that γ (k) is the machine in position k of the permutation. An m × m permutation matrix for machines, Ω = [ω ij ], is constructed from γ by defining, for 1 ≤ slantislantm, ω iγ (i) = 1 and ω ij = 0 for 1 ≤ slantj ≠ γ (i) ≤ slantm. The set of all m¡ permutations of machines is represented as Γ (γ ∈ Γ). With these definitions in place, a permutation of the original part-machine matrix, A, based on φ and γ can be concisely represented as ΛAΩ T . The correspondingly permuted part and machine similarity matrices are Λ (AA T ) Λ T and Ω (A T A) Ω T , respectively.

2.2. A numerical example

Consider the 12 × 10 part-machine matrix displayed in , and the permutations φ = (5, 1, 6, 10, 7, 8, 11, 2, 4, 9, 3, 12) and γ = (4, 1, 7, 8, 3, 6, 5, 2, 9, 10). For the moment, we will ignore how these permutations were obtained and focus solely on their results. The reordered part-machine matrix, ΛAΩ T , based on these permutations is shown in . The permuted matrix exhibits a block-diagonal form with ones grouping along the main diagonal. Moreover, within each row and column of the permuted matrix, the ones are consecutive. This consecutivity property ensures that interval assignments can be produced for both parts and machines. To illustrate, displays an interval assignment for machines. Notice that the interval for each machine overlaps only with the intervals for machines that produce some of the same parts. For example, the interval for machine 5 only overlaps with the intervals of machines for which it has at least one part in common (machines 8, 3, 6, 2, and 9). represents the permuted part similarity matrix, Λ (AA T ) Λ T . We observe that this matrix has a perfect Robinson structure because the elements never increase when moving away from the main diagonal within any row or column. This same property holds for the permuted machine similarity matrix, Ω (A T A) Ω T , in .

Fig. 1 A numerical example to illustrate the relationship between the consecutive ones property of the part-machine matrix and the Robinson structure of the part and machine similarity matrices: (a) the raw matrix, A; (b) the permuted matrix, ΛAΩ T ; (c) the permuted part matrix, Λ (AA T ) Λ T ; and (d) the permuted machine matrix, Ω(A T A) Ω T .

Fig. 1 A numerical example to illustrate the relationship between the consecutive ones property of the part-machine matrix and the Robinson structure of the part and machine similarity matrices: (a) the raw matrix, A; (b) the permuted matrix, ΛAΩ T ; (c) the permuted part matrix, Λ (AA T ) Λ T ; and (d) the permuted machine matrix, Ω(A T A) Ω T .

Fig. 2 An interval assignment for machines associated with the optimally permuted part-machine matrix in .

Fig. 2 An interval assignment for machines associated with the optimally permuted part-machine matrix in Fig. 1(b).

The example in (, , , ) establishes the premise of our proposed method. If we can obtain permutations of the rows and columns of the part-machine matrix such that the reordered matrix exhibits the consecutive ones property in all rows and columns, then, based on the classic result reported in CitationKendall (1969), the correspondingly permuted part and machine similarity matrices will have a perfect Robinson structure. Although the converse of Kendall's theorem is not always true, permutation of the part and machine similarity matrices into a Robinson structure should provide a strong propensity for the correspondingly reordered part-machine matrix to exhibit the consecutive ones property and block-diagonal form.

In practice, achieving consecutive ones in all rows and columns and a perfect Robinson structure for the part and machine similarity matrices is likely unattainable. In other words, there is no permutation of parts and machines that will provide a perfect Robinson structure for the respective similarity matrices. Nevertheless, if we can permute the similarity matrices to realize an approximate Robinson structure, then this should tend to produce a block-diagonal form in the part-machine matrix. In the next section, we present several optimization criteria that might be useful for approximating a Robinson structure.

2.3. Optimization criteria for an approximate Robinson structure

There are a host of possible criteria that can be devised to approximate a Robinson structure within a reordered similarity matrix. One option is to use the linear placement problem described by CitationSuryanarayanan et al. (1991) and CitationShargal et al. (1995). We adapt two within-row and column gradient criteria used by Brusco and Stahl (2005, Ch. 9) within the context of an anti-Robinson structure (i.e., never decreasing elements when moving away from the main diagonal within a row or column) in applied data analysis. Without loss of applicability, we display these criteria within the context of permuting the similarity matrix for parts, P = AA T . The linear placement model, under the assumption of equal distances between positions in the sequence (d kl = d lk = lk for 1 ≤ slantk < lslantn), is as follows:

The second criterion is an unweighted within-row-and-column gradient index:

where

The third criterion is the weighted within-row-and-column gradient index:

or

The linear placement criterion, f 1(φ), attempts to place similar parts close to one another and the imposed penalties are a function of the distances between positions in the sequence. The indices f 2(φ) and f 3(φ) both use two terms above the main diagonal of the reordered matrix. (Because P is symmetric, consideration of the below diagonal terms is redundant.) The first term in f 2(φ) and f 3(φ) measures whether or not a Robinson structure is present in row h above the main diagonal, whereas the second term determines the appropriateness of the structure in column l above the diagonal The two indices differ in that f 3(φ) assigns more credit to concordance with a Robinson structure and a greater penalty to discordance with a Robinson structure.

Based on our preliminary experiments, we concluded that f 3(φ) generally produces a slightly better block-diagonal structure than f 2(φ). This finding is likely attributable to the fact that f 3(φ) penalizes egregious violations of a Robinson structure more harshly than f 2(φ). For example, assume that elements of “one” and “five” were observed within a row (above the main diagonal) in a reordered matrix, and that the five is to the right of the one. This violation of the Robinson structure would be penalized a value of “−1” using f 2(φ), which measures only consistency or inconsistency with the structure. In contrast, a penalty of “1 – 5 = −4” would be realized with f 3(φ), which captures the seriousness of the violation. We have observed that the optimal part-machine permutations for f 1(φ) and f 3(φ) were identical in most cases and, therefore, do not advocate one criterion over the other with respect to performance. However, with respect to computational efficiency of the exact algorithms (particularly, the branch-and-bound algorithm), optimal permutations were typically obtained in much less time for f 3(φ) relative to f 1(φ). For this reason, the remainder of our paper focuses on the f 3(φ) index and we refer to the problem posed by Equation (Equation4) as the Robinson Permutation Problem (RPP).

3. Solution procedures

3.1. A dynamic programming algorithm

The dynamic programming approach for the RPP is comparable to implementations for several problems pertaining to row layout (CitationPicard and Queyranne, 1981; CitationKouvelis et al., 1995; CitationBrusco, 2004). We provide a sufficient description of the dynamic programming method as related to our implementation for the RPP. We employ the following notation:

S =

the set of indices for parts {1, 2, …, n}.

θ=

the power set of S, i.e., the set of all subsets of S.

θ k =

the set of all subsets in θ that have exactly k part indices.

S k =

S k ∈ θ k is an arbitrary subset from θ k . In other words, S k represents one of the n¡/(k¡(nk)¡) subsets of size k contained in θ k . We delineate S k i (for all iS k ) as S k with the further condition that part i is placed last in any partial sequence of the parts in S k .

F(S k i )=

the contribution to the RPP index that is accumulated from placing part i in the last position of the sequence for subset S k .

F* (S k )=

the best possible RPP index that can be achieved for subset S k .

τ* (S k )=

the part that should be placed last in the partial sequence of the objects in S k to produce the best possible RPP index value, F* (S k ).

The stages of the dynamic programming algorithm are characterized by the subset sizes. In the first stage, we indicate that the contribution to the RPP objective function is zero for all n subsets of size 1:

The recursion equations for stages 2 ≤ slantkslantn are established as

The values of F* (S k ) and τ* (S k ) are stored in 2 n -dimensional vectors and accessed using a binary representation of the subset. The optimal permutation is acquired by backtracking from the final subset in stage n, which is S n = S, using the stored values τ* (S k ) for 1 ≤ slantkslantn. As noted by CitationBrusco (2004), the principal limitation of the dynamic programming approach is the computer memory required to store the F* (S k ) and τ* (S k ) values. For current microcomputer platforms, memory constraints preclude optimal solutions for n ≈ 27 or greater. A small numerical demonstration of the dynamic programming algorithm is provided in the Appendix.

3.2. A branch-and-bound algorithm

Branch-and-bound methods have also proven effective for permutation problems (CitationBrusco, 2004; CitationBrusco and Stahl, 2005, Chs. 7–11). We have designed a branch-and-bound procedure for the RPP that can facilitate optimal solutions for problems with n = 35 or more objects. We begin by obtaining a lower bound for f 3(φ) using 20 restarts of a pairwise interchange heuristic similar to the CRAFT heuristic (CitationArmour and Buffa, 1963) for the quadratic assignment problem. Each restart uses a different, randomly generated, initial permutation. The best solution found across the 20 restarts provides an initial incumbent solution, φB, and corresponding lower bound, f LB. The steps of the branch-and-bound algorithm are as follows:

  • Step 0. Set t = 1, φ (t) = 1, and φ (k) = 0 for 2 ≤ slantkslantn. Obtain φB and f LB using a pairwise interchange heuristic. Compute the following three-dimensional array, Y = [y ijl ]: y ijl = (p ij + p jl − 2p il ) for 1 ≤ slantijlslantn.

  • Step 1. Set t = t + 1.

  • Step 2. Set φ (t) = φ (t) + 1.

  • Step 3. If φ (t) = φ (k) for any k = 1, …, t−1, then go to Step 2.

  • Step 4. If φ (t) = 2 and φ (k) ≠ 1 for any k = 1, …, t −1, then go to Step 2.

  • Step 5. If t = 1 and φ (t) > n, then return best solution and Stop.

  • Step 6. If t > 1 and φ (t) > n, go to Step 11.

  • Step 7. If t = n−1, then go to Step 8. Otherwise go to Step 9.

  • Step 8. Compute f 3(φ) for the complete sequence. If f 3(φ) > f LB, then set f LB = f 3(φ) and φB = φ. Go to Step 2.

  • Step 9. Conduct tests for insertion of the part in position t to position s. Define η t as the subset of parts assigned to the first t positions and perform the following test for t − 1 ≥ slantsslant 1:

    If Equation (Equation10) is satisfied for t −1 ≥ slantsslant 1, then go on to Step 10, otherwise go to Step 2.

  • Step 10. Conduct the bound test to determine whether the current partial permutation can lead to a value of f 3(φ) > f LB upon its completion.

    If Equation (Equation11) is satisfied, go to Step 1, otherwise go to Step 2.

  • Step 11. Perform depth retraction by setting φ (t) = 0 and t = t −1. Go to Step 2.

The branch-and-bound procedure constructs partial sequences by assigning parts to positions indexed by the pointer, t. Partial sequences are pruned if a part has already been assigned to a previous position in the sequence (Step 3), or if part 2 is assigned prior to part 1 in the sequence (Step 4). The latter pruning rule prevents evaluation of sequences that are the reverse order of other sequences where part 1 precedes part 2. The most important pruning rules are those in Steps 9 and 10. Step 9 prunes the partial solution if the relocation of the part in position t to an earlier position in the permutation, s, (1 ≤ slantsslantt−1) will lead to a better RPP criterion value. The bound test in Step 10 enacts a pruning operation if it is determined that pursuit of the partial sequence cannot ultimately lead to a complete permutation with an RPP criterion value greater than the current bound. All permutations have been implicitly or explicitly evaluated if t = 1 and φ (t) > n at Step 5, which results in termination of the algorithm. The computation time associated with the branch-and-bound algorithm tends to become excessive for n > 35.

3.3. A simulated annealing heuristic

Although exact methods, such as dynamic programming and branch-and-bound algorithms, are capable of providing guaranteed optimal solutions to RPP instances of nontrivial size, they are not computationally feasible for some of the larger part-machine matrices encountered in the empirical literature. Therefore, we also developed a simulated annealing heuristic for the RPP that is capable of accommodating larger test problems. The structure of the simulated annealing heuristic is closely related to those designed for related permutation problems (CitationHeragu and Alfa, 1992; CitationBrusco and Stahl, 2000), and the selection of algorithm parameters (cooling factor, temperature length, etc) were based on previous research as well as our own experimentation. The pseudocode for the simulated annealing heuristic is displayed in .

Fig. 3 Pseudocode for the simulated annealing heuristic.

Fig. 3 Pseudocode for the simulated annealing heuristic.

3.4. Partitioning the reordered matrix

It is important to clarify that the dynamic programming, branch-and-bound and simulated annealing algorithms for RPP are designed for seriating the rows and columns of the part-machine matrix, not partitioning them. Accordingly, our methods are designed in the spirit of the Bond Energy Algorithm (BEA) (CitationMcCormick et al., 1972). The BEA is a classic seriation procedure that spans several literature bases, and has served as the impetus for subsequent refinements in the cell formation context (CitationAskin et al., 1991; CitationShargal et al., 1995). The goodness-of-fit associated with a seriation procedure can be assessed using clustering efficiency measures such as those offered by CitationShargal et al. (1995). However, as noted by CitationCrama and Oosten (1996), a seriation procedure does not establish a partition of parts and machines. CitationShargal et al. (1995), for example, visually established blocks from the reordered matrix.

Although visual selection of blocks based on the permuted matrix is perfectly viable, we have written a computer program for partitioning the reordered part-machine matrix produced by our RPP algorithms. The program allows the user to specify the desired number of blocks, b, and then uses a simple exchange algorithm to move parts and machines from one block to another while preserving the order of parts and machines obtained by our permutation methods. The objective criterion of the partitioning program that we selected is to maximize the modified grouping efficiency index originally proposed by CitationKumar and Chandrasekharan (1990), and also employed by CitationCrama and Oosten (1996). The index is computed as follows:

where e is the number of operations (ones appearing in A), e O is the number of exceptional elements (ones appearing outside the blocks along the main diagonal), and e V is the number of voids (zeros appearing inside the blocks along the main diagonal). The algorithm could be modified, however, for a myriad of other clustering efficiency measures (see CitationSarker and Mondal (1999) and CitationSarker and Khan (2001) for reviews).

4. Computational results

We applied the dynamic programming, branch-and-bound and simulated annealing algorithms to the example data set in , as well as a number of cell formation problems from the research literature. The test problems from the literature include: (i) the 10 × 15 matrix presented by CitationChan and Milner (1983); (ii) the 19 × 12 matrix presented by CitationAskin et al. (1991); (iii) the 19 × 12 matrix presented by CitationVakharia and Wemmerlöv's (1990); (iv) the 20 × 8 matrix presented by CitationKing and Nakornchai (1982); (v) the 11 × 22 matrix and the 10 × 15 matrix presented by CitationSeifoddini (1988); (vi) the 23 × 20 matrix presented by CitationGroover (1987); (vii) the 16 × 30 matrix (problem 1) presented by CitationBoctor (1991); (viii) the 35 × 20 matrix presented by CitationCarrie (1973); (ix) the 35 × 20 matrix presented by CitationBoe and Cheng (1991); (x) the 40 × 24 matrix (problem 2) presented by CitationChandrasekharan and Rajagopalan (1989); and (xi) the 41 × 30 matrix presented by CitationKumar and Vanelli (1987).

A summary of results for each test problem is provided in . The results include the total CPU time required to obtain the RPP permutations using dynamic programming, branch-and-bound and simulated annealing algorithms. In addition, we report the total Jaccard index (CitationAnderberg, 1973, Ch. 4) for the optimally permuted part-machine matrices. The Jaccard index, which was recognized as an especially effective measure of clustering efficiency by CitationShargal et al. (1995), is computed for rows as follows:

where
The total Jaccard index is the sum of the row and column Jaccard indices. We also report, for each problem, the number of blocks, b, and modified grouping efficiency index, α . These results were obtained from the partitioning algorithm described in Section 3.4.

Table 1 Performance of dynamic programming (DP), branch and bound (BB), and simulated annealing (SA) for selected cell formation problemsFootnote*

The CPU times for both the dynamic programming and branch-and-bound algorithm are quite modest for all test problems where both dimensions of the matrix are 20 or fewer. The dynamic programming approach was faster than the branch-and-bound algorithm for the two smallest matrices, but required more CPU time for all remaining test problems where comparisons were possible. Dynamic programming was computationally infeasible for the last five test problems in because of computer memory limitations.

Although space limitations prevent the display of the RPP optimally permuted part-machine matrices for each of the test problems, we believe that closer inspection of at least one matrix is warranted. (, , , ) presents the results for the Chan–Milner matrix in a structure similar to that of (, , , ). The raw matrix is shown in and the RPP optimally permuted matrix is shown in . The optimally permuted matrix exhibits a strong block-diagonal structure, and there are only a few violations of the consecutive ones property in the rows and columns of . Accordingly, and exhibit a strong, although not perfect, Robinson structure. For example, consider the row for machine 1 in . When moving to the right in this row, away from the main diagonal, there is an increase from two to three when proceeding from column 4 to column 6.

Fig. 4 (a) A part-machine matrix from CitationChan and Milner (1983), A; (b) an optimal Robinson permutation of that matrix, ΛAΩ T ; (c) an optimal permutation of the part similarity matrix, Λ (AA T ) Λ T ; and (d) an optimal permulation of the machine similarity matrix, Ω (A T A) Ω T .

Fig. 4 (a) A part-machine matrix from CitationChan and Milner (1983), A; (b) an optimal Robinson permutation of that matrix, ΛAΩ T ; (c) an optimal permutation of the part similarity matrix, Λ (AA T ) Λ T ; and (d) an optimal permulation of the machine similarity matrix, Ω (A T A) Ω T .

The block-diagonal form of provides a visual representation that can be used to select partitions of part and machine indices, which enables the creation of cells. We display our partitioning program results for the matrix in for the b = 2 and b = 3 blocks. For example, consider the two-block partition that we have shown in with solid lines. The upper-left cell is comprised of parts {3, 4, 8, 9, 1, 6, 7} and machines {1, 3, 14, 4, 6, 7, 11, 2, 12}. The bottom-right cell consists of parts {2, 5, 10} and machines {5, 9, 10, 13, 15, 8}. This solution, which yields α = 0.605, has the desired property of no exceptional elements. However, the proportion of voids might be too large to enable a desired level of machine utilization. To improve machine utilization, a decision-maker might consider the three-block partition obtained by splitting the upper-left cell into two cells, as we have shown in with dashed lines. This solution has only six voids and, therefore, enables greater machine utilization. The price paid for the increase in machine utilization is five exceptional elements (i.e., note the five ones outside the boundaries of the dashed lines). Nevertheless, the grouping efficiency of the three-block solution is α = 0.800, which is clearly superior to the corresponding index for b = 2.

One of the most interesting aspects of is a comparison of the results for Carrie's matrix relative to those for the Boe–Cheng matrix. These two matrices have the same dimensions and nearly the same density. In fact, the Boe–Cheng matrix is a perturbed version of Carrie's matrix. The branch-and-bound algorithm provided optimal RPP part-machine permutations for both matrices; however, the differences in required computation time were enormous. For Carrie's matrix, which has a strong block-diagonal form resulting from optimal RPP permutation, the optimal permuted matrix was obtained in 18.56 seconds. Although the optimal RPP part-machine permutation for the Boe–Cheng matrix also has a block-diagonal form, it is not as strong as the form for Carrie's matrix. The required computation time for the Boe–Cheng matrix was 6690.34 seconds, more than 350 times that required for Carrie's matrix.

The Jaccard indices for the Carrie and Boe–Cheng RPP permuted matrices also provide explanation for the marked difference in CPU times, with the total Jaccard index for Carrie's matrix being more than 36% greater than the corresponding index for the Boe–Cheng. Although the computation time for the Boe–Cheng matrix seems excessive, it should be remembered that the solution procedure is solving a permutation problem for a 35 × 35 matrix as part of this process, which permits a solution space of 35¡/2 = 5.166 × 1039 permutations. In this light, the solution time of just under 2 hours does not seem to be an especially unreasonable consumption of resources. In addition, it is reassuring to observe that the simulated annealing algorithm obtains the optimal solution to the RPP in under 8 seconds. Similar results were realized for the two largest part-machine matrices in . The branch-and-bound algorithm obtained optimal RPP permutations for the Chandrasekharan–Rajagopalan 40 × 24 matrix and the Kumar–Vanelli 41 × 30 matrix in roughly 11.5 and 22 hours, respectively. However, the simulated annealing heuristic matched these optimal RPP solutions in 12.06 and 14.73 seconds, respectively.

5. Conclusions

The intended primary contribution of our paper was to position the part-machine clustering problem in cell formation within the context of the consecutive ones property and its relationship to interval graphs and Robinson matrices, which to our knowledge is unprecedented in the literature. At the same time, we recognized that it was not sufficient to link cell formation to Robinson matrices without providing any computational tools for fitting Robinson structure so as to establish block-diagonal form in the part-machine matrix. For this reason, we offered a secondary contribution that focused on the use of dynamic programming, branch-and-bound, and simulated annealing methods for obtaining Robinson-based permutation of matrices. Dynamic programming and the branch-and-bound algorithm have the advantage of guaranteeing an optimal solution for the selected RPP criterion. Our computational results show that both of these exact methods are effective for establishing block-diagonal patterning, within a reasonable time frame, for many part-machine matrices from the literature. However, even the more scalable of the procedures, the branch-and-bound algorithm, can require excessive amounts of computation time for matrices where one of the dimensions exceeds 35.

Our simulated annealing heuristic does not verify optimality, but can handle larger matrices. Other metaheuristics such as tabu search, genetic algorithms, and variable neighborhood search may also prove useful for larger instances of the RPP. Fortunately, the cell formation literature already contains a number of contributions using these methods (CitationShargal et al., 1995; CitationOnwubolu and Mutingi, 2001; CitationLei and Wu, 2005) and adaptation for RPP should be relatively straightforward. A perhaps more challenging, but no less important extension would be the development and testing of multicriterion cell formation approaches (see, for example, CitationMalakooti and Yang (2002)) that incorporate Robinson indices as part of the objective function.

Appendix A numerical illustration of the dynamic programming algorithm

Consider the 3 × 4 part-machine matrix, A, and part-similarity matrix, P shown in .

  • Stage 1 (initialize for recursion, k = 1).

  • F*(S 1 = {1}) = 0 and τ *(S 1 = {1}) = 1,

  • F*(S 1 = {2}) = 0 and τ *(S 1 = {2}) = 2,

  • F*(S 1 = {3}) = 0 and τ *(S 1 = {3}) = 3.

  • Stage 2 ( k = 2):

  • F(S 2 1 = {1,2}) = F*(S 1 = {2}) + p 21+ p 13− 2p 23 = 0 + 0 + 1 − 2(1) = −1.

  • F(S 2 2 = {1,2}) = F*(S 1 = {1}) + p 12+ p 23− 2p 13 = 0 + 0 + 1 − 2(1) = −1.

  • F(S 2 1 = {1,3}) = F*(S 1 = {3}) + p 31+ p 12− 2p 32 = 0 + 1 + 0 − 2(1) = −1.

  • F(S 2 3 = {1,3}) = F*(S 1 = {1}) + p 13+ p 32− 2p 12 = 0 + 1 + 1 − 2(0) = 2.

  • F(S 2 2 = {2,3}) = F*(S 1 = {3}) + p 32+ p 21− 2p 31 = 0 + 1 + 0 − 2(1) = −1.

  • F(S 2 3 = {2,3}) = F*(S 1 = {2}) + p 23+ p 31− 2p 21 = 0 + 1 + 1 − 2(0) = 2.

  • F*(S 2 = {1,2}) = max (F(S 2 1 = {1,2}), F(S 2 2 = {1,2})) = −1 and τ*(S 2 = {1,2}) = 1 or 2,

  • F*(S 2 = {1,3}) = max (F(S 2 1 = {1,3}), F(S 2 3 = {1,3})) = 2 and τ*(S 2 = {1,3}) = 3, F*(S 2 = {2,3}) = max (F(S 2 2 = {2,3}), F(S 2 3 = {2,3})) = 2 and τ*(S 2 = {2,3}) = 3.

  • Stage 3 ( k = 3):

  • F(S 3 1 = {1,2,3}) = F*(S 2 = {2,3}) = 2,

  • F(S 3 2 = {1,2,3}) = F*(S 2 = {1,3}) = 2,

  • F(S 3 3 = {1,2,3}) = F*(S 2 = {1,2}) = −1,

  • F*(S 3 = {1,2,3}) = max (F(S 3 1 = {1,2,3}), F(S 3 2 = {1,2,3}), F(S 3 3 = {1,2,3})) = 2. and τ *(S 3 = {1,2}) = 1 or 2.

To find the optimum we can place either part 1 or part 2 last in the sequence and backtrack. Backtracking from part 1 at Stage 3, we proceed to F*(S 2 = {2,3}) in Stage 2 and remove part 3, then we move to F*(S 1 = {2}) at Stage 1 and remove part 2. This yields the sequence 2-3-1. If we backtrack from part 2 at Stage 3, we obtain the reverse of the 2-3-1 sequence (i.e., 1-3-2). Both sequences are optimal because of the symmetry of P. Both permutations yield a perfect Robinson structure in P.

Fig. A1 The part-machine matrix A and the part similarity matrix P.

Fig. A1 The part-machine matrix A and the part similarity matrix P.

Biographies

Michael J. Brusco is the Synovus Professor of Marketing and Operations Research at Florida State University. His research focuses on the development of optimal and heuristic solution procedures for combinatorial optimization problems related to scheduling, location, layout, and data analysis. He is a past member of the Board of Directors for the Classification Society of North America, and currently serves on the Editorial Board of the Journal of Classification.

Douglas Steinley is an Assistant Professor of Psychological Sciences at the University of Missouri-Columbia. His research focuses on general multivariate statistical procedures, with emphasis on combinatorial optimization as related to exploratory data analysis. He is a member of the Board of Directors for the Classification Society of North America and currently is the Book Review Editor for the Journal of Classification.

Notes

*CPU times in seconds are for a 3.4 GHz, Pentium PC with 1GB of RAM (dynamic programming was infeasible when one of the matrix dimensions exceeded 26). The density is the proportion of ones in the matrix. The JI column contains the total Jaccard index (sum of row and column Jaccard indices). The b and α columns contain the number of blocks and the modified grouping index, respectively, after applying the partitioning algorithm.

References

  • Anderberg , M. R. 1973 . Cluster Analysis for Applications , New York, NY : Academic Press .
  • Armour , G. C. and Buffa , E. S. 1963 . A heuristic algorithm and simulation approach to relative location of facilities . Management Science , 9 : 294 – 309 .
  • Askin , R. G. , Creswell , S. H. , Goldberg , J. B. and Vakharia , A. J. 1991 . A Hamiltonian path approach to reordering the part-machine matrix for cellular manufacturing . International Journal of Production Research , 29 : 1081 – 1100 .
  • Boctor , F. F. 1991 . A linear formulation of the machine-part cell formation problem . International Journal of Production Research , 29 : 343 – 356 .
  • Boe , W. J. and Cheng , C. H. 1991 . A close neighbor algorithm for designing cellular manufacturing systems . International Journal of Production Research , 29 : 2097 – 2116 .
  • Brusco , M. J. 2004 . Optimal and heuristic methods for the minimum-backtracking row layout problem . IIE Transactions , 36 : 181 – 189 .
  • Brusco , M. J. and Stahl , S. 2000 . Using quadratic assignment methods to generate initial permutations for least-squares unidimensional scaling of symmetric proximity matrices . Journal of Classification , 17 : 197 – 223 .
  • Brusco , M. J. and Stahl , S. 2005 . Branch-and-Bound Applications in Combinatorial Data Analysis , New York, NY : Springer .
  • Carrie , A. S. 1973 . Numerical taxonomy applied to group technology and plant layout . International Journal of Production Research , 11 : 399 – 416 .
  • Chan , H. M. and Milner , D. A. 1983 . Direct clustering algorithm for group formation in cellular manufacturing . Journal of Manufacturing Systems , 1 : 65 – 74 .
  • Chandrasekharan , M. P. and Rajagopalan , R. 1986 . MODROC: An extension of rank order clustering for group technology . International Journal of Production Research , 24 : 1221 – 1233 .
  • Chandrasekharan , M. P. and Rajagopalan , R. 1989 . Groupability: An analysis of the properties of binary data matrices for group technology . International Journal of Production Research , 27 : 1035 – 1052 .
  • Crama , Y. and Oosten , M. 1996 . Models for machine-part grouping in cellular manufacturing . International Journal of Production Research , 34 : 1693 – 1713 .
  • Fulkerson , D. R. and Gross , O. A. 1965 . Incidence matrices and interval graphs . Pacific Journal of Mathematics , 15 : 835 – 855 .
  • Groover , M. P. 1987 . Automation, Production Systems and Computer Integrated Manufacturing , Englewood Cliffs, NJ : Prentice-Hall .
  • Heragu , S. S. and Alfa , A. S. 1992 . Experimental analysis of simulated annealing based algorithms for the layout problem . European Journal of Operational Research , 57 : 190 – 202 .
  • Kendall , D. G. 1969 . Incidence matrices, interval graphs, and seriation in archaeology . Pacific Journal of Mathematics , 28 : 565 – 570 .
  • King , J. R. and Nakornchai , V. 1982 . Machine-component group formation technology: review and extension . International Journal of Production Research , 20 : 117 – 133 .
  • Kouvelis , P. , Chiang , W. C. and Yu , G. 1995 . Optimal algorithms for row layout problems in automated manufacturing systems . IIE Transactions , 27 : 99 – 104 .
  • Kumar , C. S. and Chandrasekharan , M. P. 1990 . Grouping efficacy: A quantitative criterion for goodness of block diagonal forms of binary matrices in group technology . International Journal of Production Research , 28 : 233 – 243 .
  • Kumar , K. R. and Vanelli , A. 1987 . Strategic subcontracting for efficient disaggregated manufacturing . International Journal of Production Research , 25 : 1715 – 1728 .
  • Lei , D. and Wu , Z. 2005 . Tabu search approach based on a similarity coefficient for cell formation in generalized group technology . International Journal of Production Research , 43 : 4035 – 4047 .
  • Malakooti , B. and Yang , Z. 2002 . Multiple criteria approach and generation of efficient alternatives for machine-part family formation in group technology . IIE Transactions , 34 : 837 – 846 .
  • McCauley , J. 1972 . Machine grouping for efficient production . Production Engineer , 51 : 53 – 57 .
  • McCormick , W. T. , Schweitzer , P. J. and White , T. W. 1972 . Problem decomposition and data reorganization by a clustering technique . Operations Research , 20 : 993 – 1009 .
  • Onwubolu , G. C. and Mutingi , M. 2001 . A genetic algorithm approach to cellular manufacturing systems . Computers and Industrial Engineering , 39 : 125 – 144 .
  • Picard , J. C. and Queyranne , M. 1981 . On the one-dimensional space allocation problem . Operations Research , 29 : 371 – 391 .
  • Robinson , W. S. 1951 . A method for chronologically ordering archeological deposits . American Antiquity , 16 : 293 – 301 .
  • Sarker , B. R. and Khan , M. 2001 . A comparison of existing grouping efficiency measures and a new weighted grouping efficiency measure . IIE Transactions , 33 : 11 – 27 .
  • Sarker , B. R. and Mondal , S. 1999 . Grouping efficiency measures in cellular manufacturing: a survey and critical review . International Journal of Production Research , 37 : 285 – 314 .
  • Seifoddini , H. 1988 . Comparison between single linkage and average linkage clustering techniques in forming machine cells . Computers and Industrial Engineering , 15 : 210 – 216 .
  • Selim , H. M. , Askin , R. G. and Vakharia , A. J. 1998 . Cell formation in group technology: review, evaluation and directions for future research . Computers and Industrial Engineering , 34 : 3 – 20 .
  • Shargal , M. , Shekhar , S. and Irani , S. A. 1995 . Evaluation of search algorithms and clustering efficiency measures for machine-part matrix clustering . IIE Transactions , 27 : 43 – 59 .
  • Suryanarayanan , J. K. , Golden , B. L. and Wang , Q. 1991 . A new heuristic for the linear placement problem . Computers and Operations Research , 18 : 255 – 262 .
  • Vakharia , A. J. and Wemmerlöv , U. 1990 . Designing a cellular manufacturing system: A materials flow approach based on operations sequences . IIE Transactions , 22 : 84 – 97 .

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.