1. Introduction
1.1. Blockmodeling and Structural Equivalence
One early goal of social psychologists and sociologists was the understanding of role relations in social groups. This required collecting and organizing social relational data to reveal the fundamental network structure of social groups. Blockmodeling was a tool originally designed to accomplish this task. Although its origin stems from the study of role relations, blockmodeling has also become a useful tool for discerning the fundamental structure of any social network.
The key idea behind blockmodeling is to think of (social) actors as being equivalent in terms of the distribution of their ties in their social group. Equivalent actors are grouped into clusters, called positions, and this imposes also a partition of the relational ties into blocks. A block is composed of the set of ties between the actors in a pair of positions. If there are n actors in a group, then the size of its network is (n×n), and if there are K positions, then the blockmodel is a (K×K) network (with K≪n). The simpler and much smaller (K×K) network is an image of the larger (n×n) network. This image captures the underlying (essential) structure of the network, one that is more easily understood than the original network. In brief, blockmodeling methods partition the nodes and ties of a network in an attempt to reveal the basic structural properties of that network.
A key to successful blockmodeling is defining equivalence among actors. Lorrain and White (Reference Lorrain and White1971) introduced “structural equivalence” where actors are equivalent if they are connected to exactly the same other actors in the network: They are structurally identical and can be put in the same cluster. The premise of structural equivalence informed attempts to create programs that establish blockmodel images of networks, while recognizing that most networks do not conform exactly to structural equivalence. Broadly, these programs can be classified into two categories: Indirect and direct approaches. Breiger, Boorman, and Arabie (Reference Breiger, Boorman and Arabie1975) and Burt (Reference Burt1976) offered widely used algorithms for blockmodeling that are indirect in the sense that they transform a network into a set of (dis)similarity measures among actors and then cluster them. The problem with such indirect approaches is that there are many (dis)similarity measures and many potential clustering algorithms. Arbitrary choices about these items affect the blockmodels identified. Batagelj, Ferligoj, and Doreian (Reference Batagelj, Ferligoj and Doreian1992) noted that if a network does conform to structural equivalence, then the only possible types of blocks are null blocks (containing only null ties) and complete blocks (having no null ties). This permits a direct approach to fitting blockmodels by minimizing an explicit criterion function operationalizing departures from exact structural equivalence. Doreian, Batagelj, and Ferligoj (Reference Doreian, Batagelj and Ferligoj1994) found that the direct algorithm of Batagelj et al. (Reference Batagelj, Ferligoj and Doreian1992) frequently outperformed the Breiger et al. (Reference Breiger, Boorman and Arabie1975) algorithm in terms of fit and never performed worse in terms of fitting the data.
Doreian, Batagelj, and Ferligoj (Reference Doreian, Batagelj and Ferligoj2005) present generalized blockmodeling as a general approach to fitting blockmodels to social network data where a much wider set of ideal block types, beyond those implied by structural equivalence, are mobilized. Examples include ranked clusters models for classroom networks and nobility marriage systems, column-regular and row-regular block types for primate grooming networks, row-dominant and column-dominant blocks for Little League baseball teams and student discussion groups, and signed networks (with positive and negative blocks) for study of organizational data. Depending on the substantive context, a wide range of block types and blockmodel forms can be fitted to social networks. Nevertheless, structural equivalence remains the most used equivalence type in blockmodeling. Žnidaršič, Ferligoj, and Doreian (Reference Žnidaršič, Ferligoj and Doreian2012) show that structural equivalence is quite robust, even in the presence of substantial measurement error, in returning the correct structure of social networks in blockmodel form. Given its wide use and robustness, we focus on structural equivalence here.
1.2. Blockmodeling Variants
In addition to defining equivalence types, other important issues relevant for blockmodeling merit attention. Brusco and Steinley (Reference Brusco and Steinley2011) noted that blockmodeling can be categorized along a number of dimensions: (1) deterministic vs. stochastic, (2) one-mode vs. two-mode networks, (3) signed vs. unsigned networks, and (4) exploratory vs. confirmatory blockmodeling. Perhaps the most critical dimension pertains to the absence or presence of the assumption of a probabilistic model. Deterministic blockmodeling methods (Batagelj et al. Reference Batagelj, Ferligoj and Doreian1992; Brusco, Doreian, Mrvar, & Steinley, Reference Brusco, Doreian, Mrvar and Steinley2011; Brusco & Steinley, Reference Brusco and Steinley2007, Reference Brusco and Steinley2009a, Reference Brusco and Steinley2011; Doreian et al. Reference Doreian, Batagelj and Ferligoj2005; Doreian & Mrvar Reference Doreian and Mrvar1996; Hartigan Reference Hartigan1972) do not assume a probabilistic model and typically rely on exact or approximate algorithms to find blockmodels that minimize some measure of inconsistency with an ideal block structure. Contrastingly, stochastic blockmodeling methods (Airoldi, Blei, Fienberg, & Xing, Reference Airoldi, Blei, Fienberg and Xing2008; Bickel & Chen Reference Bickel and Chen2009; Fienberg, Meyer, & Wasserman, Reference Fienberg, Meyer and Wasserman1985; Goldenberg, Zheng, Fienberg, & Airoldi, Reference Goldenberg, Zheng, Fienberg and Airoldi2009; Govaert & Nadif Reference Govaert and Nadif2003; Handcock, Raftery, & Tantrum, Reference Handcock, Raftery and Tantrum2007; Holland, Laskey, & Leinhardt, Reference Holland, Laskey and Leinhardt1983; Nowicki & Snijders Reference Nowicki and Snijders2001; Steinley, Brusco, & Wasserman, Reference Steinley, Brusco and Wasserman2011; Wasserman & Faust Reference Wasserman and Faust1994) posit an underlying statistical model and incorporate maximum likelihood or pseudo-maximum likelihood methods to obtain blockmodels.
One-mode blockmodeling methods produce a partition of a single set of objects that define the nodes of the network, whereas two-mode methods obtain two separate but simultaneous partitions of the two sets of nodes. Unsigned networks assume that edges of the network are either absent or positively weighted when present, but signed networks allow negative weights as well (e.g., enemy ties as well as friendly ties that can vary also in strength). Exploratory blockmodeling methods make no assumption about the placement of block types within a blockmodel, whereas confirmatory methods do.
1.3. Multiobjective Blockmodeling
In general, direct blockmodeling provides a general framework for fitting blockmodels to social relational data by using an explicit criterion function capturing the difference between an ideal blockmodel structure and an empirical network. However, despite this generality, there remain problems. Even for a single network, it is possible that multiple equally well-fitting blockmodels (having a minimized criterion function) exist. Although the blockmodel structure may be stable, this presents serious problems with regard to interpreting such blockmodels. Moreover, when partitioning networks, using one criterion function seems limited. A more general approach would enable simultaneous optimization of multiple criteria. This is explicitly tackled here as a general problem that can pervade many aspects of blockmodeling.
We consider blockmodeling as a multiobjective combinatorial optimization problem. There are a number of advantages and opportunities associated with adopting this framework. One is that it can mitigate the well-known problem of multiple equally well-fitting partitions that often arises in blockmodeling (Doreian et al. Reference Doreian, Batagelj and Ferligoj2005; Doreian & Mrvar Reference Doreian and Mrvar2009). However, as more objective criterion functions are considered, the potential for equally well-fitting partitions diminishes. In this context, equally well-fitting partitions for one criterion are mitigated by the fact that other criteria are, effectively, breaking the ties among those partitions.
A second advantage of multiobjective blockmodeling is that it can preclude the need to specify weights for violations from an ideal block structure. For example, Doreian et al. (Reference Doreian, Batagelj and Ferligoj2005), Brusco and Steinley (Reference Brusco and Steinley2009a, Reference Brusco and Steinley2010), and others have discussed the use of differential weights for ones in null blocks and zeros in complete blocks. This is, effectively, an implicit transformation or weighted-sum approach to a multiobjective problem that considers these two types of violations independently. However, if we can approximate the Pareto efficient set for such problems, then a judicious selection can be made from the frontier, obviating the need for an a priori specification of weights.
A third opportunity for multiobjective programming arises in situations where an analyst has multiple network matrices for the same set of objects. This can occur when measurements are taken at multiple points in time; examples include Sampson’s (1968) monastery data and Newcomb’s (Reference Newcomb1961) fraternity data. They also can arise from multiple indicators of the same relation (e.g., Lemann & Solomon Reference Lemann and Solomon1952) or multiple relations for a set of actors (e.g., Sampson 1968). Multiobjective programming can be used to establish a blockmodel that fits multiple matrices well. To illustrate, assume that the optimal blockmodel for network matrix one fits network matrix two rather poorly, and vice versa. In many instances, multiobjective programming can identify a blockmodel that fits both network matrices (one and two) reasonably well.
We use B to represent a blockmodel for network data and f g(B) as the measure of fit of the blockmodel for criterion g for each of G criteria (1≤g≤G). The relevant criteria differ depending on the application and may include the number of inconsistencies from an ideal block structure, the correlation of the block with an ideal pattern, or a more complex (e.g., likelihood-based) measure. The focal point of multiobjective blockmodeling is the identification of Pareto efficient (or nondominated) blockmodels. To clarify this concept, we define \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\pmb{\mathcal{B}}$\end{document}
as the set of all blockmodelsFootnote 1 and let B correspond to a blockmodel from \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\pmb{\mathcal{B}}$\end{document}
(i.e., \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$B \in \pmb{\mathcal{B}}$\end{document}
). We assume, without loss of generality, that each of the G objectives should be minimized. In addition, a blockmodel B′ dominates B if f g(B′)≤f g(B) for all 1≤g≤G and f g(B′)<f g(B) for one or more of the G objectives. The Pareto efficient (or nondominated) set, \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}} \subset \pmb{\mathcal{B}}$\end{document}
, consists of all blockmodels, B′, that are not dominated by any other blockmodel of the same type.
One possible approach for approximatingFootnote 2 the set of nondominated blockmodels in \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}}$\end{document}
is to transform the multiobjective problem into a single-objective function that consists of a weighted sum of the objectives (Brusco & Stahl Reference Brusco and Stahl2001; Ehrgott & Wiecek Reference Ehrgott, Wiecek, Figueira, Greco and Ehrgott2005; Ferligoj & Batagelj Reference Ferligoj and Batagelj1992). This approach to multiobjective programming is hereafter referred to as the weighted-sum approach. Among the limitations of this approach are: (1) the challenge of choosing an appropriate set of weights and (2) the failure to identify unsupported Pareto efficient solutions.Footnote 3 The a priori selection of an appropriate set of weights is difficult because the analyst typically does not know precisely how the multiple objective functions will behave for different weights. Moreover, even if a broad set of weights is evaluated, there are often large gaps in the estimated nondominated set where interesting unsupported Pareto efficient solutions have been missed.
Given these shortcomings of the weighted-sum approach, we adopt what is known as a direct procedure for estimating the nondominated set of blockmodels. More specifically, we obtain results using an adaptation of the multiobjective tabu search (MOTS) paradigm developed by Kulturel-Konak, Smith, and Norman (Reference Kulturel-Konak, Smith and Norman2006) and also discussed by Kulturel-Konak, Coit, and Baheranwala (Reference Kulturel-Konak, Coit and Baheranwala2008). The multiobjective tabu search algorithm attempts to generate the entire nondominated set of blockmodels (both supported and unsupported). Our adaptation of the MOTS paradigm contains some critical modifications that substantially improve the quality of the estimated Pareto efficient set. The most important of these modifications are: (1) creating an initial population of the estimated Pareto efficient set with (near-)optimal blockmodels for each of the single objectives independently, (2) using finer granularities achieved by random selection of alternative weighting schemes for the objective functions rather than the all-or-nothing random selection of one criterion and the exclusion of all others, and (3) having a more exhaustive search prior to the switching of weighting schemes. Finer granularity allows for the search of more neighborhoods of the solution space, whereas an exhaustive search enables thoroughness of the investigation of each neighborhood.
In the next section, we describe the general purpose multiobjective tabu search algorithm that can be used for multiobjective blockmodeling. This is followed by Section 3, which focuses on an application within the context of generalized structural balance blockmodeling of a signed network. Section 4 reports the results for an application related to fitting a blockmodel based on multiple relations for the same set of objects. Section 5 offers an application to a larger two-mode unsigned network. Results for two sets of simulated data are provided in Section 6, along with precise recommendations for multiobjective blockmodeling. The paper concludes in Section 7 with a brief summary, as well as discussion of limitations and extensions.
2. Multiobjective Tabu Search Algorithm
Section 2.1 provides a brief description of tabu search, with particular emphasis on multiobjective tabu search. Section 2.2 provides a glossary of terms used for the presentation of the algorithm. The steps of the algorithm are presented in Section 2.3, and selection of a solution from the nondominated set is discussed in Section 2.4.
2.1. Overview of Tabu Search
Originally developed by Glover (Reference Glover1989, Reference Glover1990), tabu search is a popular metaheuristic for discrete optimization problems (see Glover & Laguna Reference Glover, Laguna and Reeves1993 for a survey). The “tabu” in tabu search involves the prohibition of specific neighborhood search operations (e.g., a relocation of an object to a different cluster) for a user-specified number of iterations, which is often referred to as the tabu length. By disallowing some neighborhood search moves, the tabu search algorithm is able to diversify the search and avoid returning to the same local optimum. This diversification often enables the algorithm to locate better local optima. Nevertheless, tabu search is a heuristic method, and, although a global optimum might be obtained, the algorithm is not guaranteed to find one.
In their adaptation of tabu search for multiobjective combinatorial optimization, Kulturel-Konak et al. (Reference Kulturel-Konak, Smith and Norman2006) randomly select one of the G objective functions to optimize at any particular time. Standard local-search operations such as pairwise interchange or object relocation are used to find the best non-tabu solution. Intensification of the search occurs as improved solutions are obtained by these operations. Diversification of the search can occur in a couple of different ways. First, a new objective function to optimize is randomly selected after a user-specified number of iterations have elapsed. In addition, after a prescribed number of iterations with no change in the nondominated set, the search process is restarted after replacing the current incumbent solution with one randomly selected from the nondominated set. A precise description of our implementation for multiobjective blockmodeling is provided in the following subsections.
2.2. Glossary
- \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\pmb{\mathcal{B}} $\end{document}

-
the set of all blockmodels;
- \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}} $\end{document}

-
the set of nondominated blockmodels;
- \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\pmb{\mathcal{B}}^{\mathrm{tabu}} $\end{document}

-
the set of blockmodels on the tabu list;
- B I
-
the incumbent blockmodel in the algorithm;
- n
-
the number of objects in the network;
- w
-
a set of objective function weights [w 1,…,w G], where w g≥0 is the weight for criterion g for 1≤g≤G and \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$\sum_{g = 1}^{G} w_{g} = 1$\end{document}
; - W
-
a q×G matrix of alternative weighting schemes, where each row contains a set of objective function weights, w. The number of weighting schemes, q, is an input parameter selected by the user;
- F(w,B)
-
the weighted objective function value, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$F(\mathbf{w},B) = \sum_{g = 1}^{G} w_{g}f_{g}(B)$\end{document}
, for blockmodel B given weighting scheme w; - ψ max
-
the termination criterion, which is defined as the maximum number of iterations permitted with no modification of the nondominated solutions list, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}}$\end{document}
; when this limit is reached, the algorithm terminates; - ξ max
-
the diversification limit, which is defined as a limit on number of iterations with no modification of the nondominated solutions list since the last diversification; when the limit is reached, diversification occurs via selection of a new incumbent solution from \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}}$\end{document}
; - τ max
-
the maximum length of the tabu list, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$\pmb{\mathcal{B}}^{\mathrm{tabu}}$\end{document}
; - ψ
-
a counter for the number of iterations with no modification in \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}}$\end{document}
; - ξ
-
a counter for the number of iterations with no modification in \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}}$\end{document}
, since the last diversification; - τ
-
a counter for the number of iterations at the current tabu length.
2.3. The Algorithm
- Step 0.
-
Initialization. Randomly construct an incumbent blockmodel, B I, and place it in the nondominated set of blockmodels such that \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}} = \{B^{\mathrm{I}}\}$\end{document}
. Alternatively, \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}}$\end{document}
can be initially populated by applying exact or approximation procedures to the single-objective optimization problems associated with each of the G criteria individually. Initialize parameters of the tabu search process: ψ max=1000; ξ max=10; \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\tau_{\max} = \mathit{random\_integer}[n, 3n]$\end{document}
; ψ=0, and ξ=0. - Step 1.
-
Select weighting scheme. Randomly select a weighting scheme, w from W.
- Step 2.
-
Increment counters. Set τ=τ+1, ξ=ξ+1, and ψ=ψ+1.
- Step 3.
-
Dynamic adjustment of tabu list length. If τ=τ max, then reset \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$\tau_{\max} = \mathit{random\_integer}\allowbreak[n, 3n]$\end{document}
and τ=0. - Step 4.
-
Neighborhood search. Conduct a search of the neighborhood of the incumbent blockmodel, B I, consisting of all single-object relocations. For each trial blockmodel, B, that is generated, perform the following substeps:
- Step 4a.
-
Add B to \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}}$\end{document}
if B is nondominated. - Step 4b.
-
Eliminate all blockmodels from \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}}$\end{document}
that are dominated by B. - Step 4c.
-
If \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}}$\end{document}
is modified in Step 4a or 4b, then set ξ=0 and ψ=0.
- Step 5.
-
Update tabu list. Choose the blockmodel \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$B \notin \pmb{\mathcal{B}}^{\mathrm{tabu}}$\end{document}
in Step 4 that minimizes F(w,B). Append B to the bottom of the tabu list, \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\pmb{\mathcal{B}}^{\mathrm{tabu}}$\end{document}
. If \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$|\pmb{\mathcal{B}}^{\mathrm{tabu}}| = \tau_{\max}$\end{document}
, then remove the oldest blockmodel from \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\pmb{\mathcal{B}}^{\mathrm{tabu}}$\end{document}
. - Step 6.
-
Intensification. If F(w,B)<F(w,B I), then set B I=B and go to Step 4; otherwise, go to Step 7.
- Step 7.
-
Termination. If ψ=ψ max, then return \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}}$\end{document}
and STOP; otherwise, proceed to Step 8. - Step 8.
-
Diversification. If ξ<ξ max, then return to Step 1 with B I as the current incumbent; otherwise, randomly select a new B I from \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}}$\end{document}
, set ξ=0, and return to Step 1.
The multiobjective tabu search algorithm for blockmodeling begins in Step 0 with the random construction of an initial blockmodel, B I, which is placed in the nondominated set, \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}} = \{B^{\mathrm{I}}\}$\end{document}
. As an alternative, \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}}$\end{document}
can be initially populated by applying exact or approximate solution procedures to each of the G single-optimization problems independently. We evaluate the merits of this latter approach in our computational studies in Sections 6.1 and 6.2.
The parameters of the algorithm are also initialized in Step 0. We use the same ψ max=1000 and \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\tau_{\max} = \mathit{random\_integer}[n, 3n]$\end{document}
parameter settings used by Kulturel-Konak et al. (Reference Kulturel-Konak, Smith and Norman2006) in their study. We also initially adopted the recommendation of Kulturel-Konak et al. (Reference Kulturel-Konak, Smith and Norman2006) by choosing ξ max=250. However, because the relocation heuristic embedded in Step 4 of our algorithm runs until convergence to a local minimum, we have found that it is more effective to change the weighting scheme much more frequently and have opted for ξ max=10. Our experience is that these settings work well across a broad range of problem sizes. A weighting scheme is selected in Step 1. This step affords an important modification from the original MOTS paradigm (Kulturel-Konak et al. Reference Kulturel-Konak, Smith and Norman2006), which selects an objective criterion in Step 1 rather than a weighting scheme. For example, in the case of G=2 criteria, the original MOTS paradigm only considers switching the selected criterion from f 1(B) to f 2(B) and vice versa. In our framework of weighting schemes, this corresponds to two choices at Step 1: (w 1=1 and w 2=0) or (w 1=0 and w 2=1). One problem with this approach is that the weight of zero implies that the algorithm is indifferent to values of the secondary objective function. We recommend defining a small positive constant, ε, that will allow at least some weight on the secondary criterion so as to break ties among alternative optima. The weights would then become (w 1=1−ε and w 2=ε) or (w 1=ε and w 2=1−ε). The value of ε should be chosen sufficiently small so that no sacrifice in the criterion weighted 1−ε would ever be permitted for an improvement in the criterion weighted ε. In our applications, ε=0.001 always suffices. Our implementation also enables the analyst to include other schemes with partial weights on both criteria, such as (w 1=0.6 and w 2=0.4) or (w 1=0.3 and w 2=0.7). The key advantage of this expansion is that it allows for a more rigorous search in more areas of the frontier of nondominated solutions.
Step 2 increments the key counters in the algorithm, namely the tabu counter (τ), the termination criterion counter (ψ), and the diversification counter (ξ). As recommended by Kulturel-Konak et al. (Reference Kulturel-Konak, Smith and Norman2006), dynamic adjustment of the length of the tabu list occurs in Step 3 every τ max iterations. The neighborhood search is conducted in Step 4 of the algorithm. Trial blockmodels are generated by attempting to relocate each object from its current cluster to one of the other clusters.Footnote 4 For each trial blockmodel, B, generated by this procedure, a check is made to see if the nondominated list, \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}}$\end{document}
, should be updated. If B is not dominated by any blockmodel in the current list, then it is appended to the list in Step 4a. If any blockmodels in \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}}$\end{document}
are dominated by the newly added blockmodel in Step 4a, then those dominated blockmodels are removed from the list in Step 4b. If \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}}$\end{document}
is modified in Step 4a (and, possibly, Step 4b as well), then the two counters for the number of iterations without modification of \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}}$\end{document}
are reset to ξ=0 and ψ=0.
The tabu list is updated in Step 5 by appending the non-tabu blockmodel from the neighborhood search that yields the best weighted objective function value. If the updated tabu list exceeds the maximum list size (i.e., if \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$|\pmb{\mathcal{B}}^{\mathrm{tabu}}| = \tau_{\max}$\end{document}
), then the first (oldest) blockmodel in the list is deleted. If the weighted objective function value for the best trial blockmodel is an improvement over the incumbent (i.e., F(w,B)<F(w,B I)), then the search process for the current weighting scheme is intensified at Step 6 by setting B I=B and returning to Step 4. Preserving the current weighting scheme as long as better F(w,B) values continue to be found represents another novel feature of our implementation of the MOTS paradigm If no better F(w,B) value is found in Step 6, then control is passed to Step 7 for a check of the termination condition.
The algorithm terminates in Step 7 if the number of iterations with no modification of the nondominated set reaches its maximum limit (i.e., ψ=ψ max), and \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}}$\end{document}
is returned as the output. After every ξ max iterations with no modification of \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}}$\end{document}
, diversification occurs in Step 8 by randomly selecting a new incumbent solution from the nondominated set. After the first ψ=ξ=10 iterations with no modification of \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}}$\end{document}
, the termination limit is not reached. Therefore, control passes from Step 7 to Step 8, where diversification occurs by randomly selecting a new incumbent solution from \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}}$\end{document}
and resetting ξ=0. If other 10 iterations with no modification occur, then ψ=20 and ξ=10. The termination limit is still not reached, but it is again time for diversification in Step 8 and resetting of ξ=0. In summary, the modification processes for ψ and ξ are as follows: (1) if a new nondominated solution is found, then both counters are reset as ξ=0 and ψ=0, (2) if the diversification limit is reached but the termination limit is not, then the diversification counter is reset to ξ=0, but the termination counter (ψ<1000) is not adjusted, or (3) if the termination limit is reached, then the algorithm terminates and \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}}$\end{document}
is returned as the output.
Multiobjective tabu search algorithms for blockmodeling have been prepared for the applications considered in this paper.Footnote 5 All programs are written in Fortran 90 and are available from the website: http://mailer.fsu.edu/~mbrusco.
2.4. Evaluation of the Approximated Nondominated Set
Ehrgott and Gandibleaux (Reference Erhgott and Gandibleaux2000) distinguish among a priori, a posteriori, and interactive multiobjective combinatorial optimization. In the a priori approach, the analyst (or decision-maker) prespecifies preferences or weights for the objective functions in advance, and these can be directly incorporated in the solution method (e.g., using the weighted-sum approach). This a priori specification is often impractical in exploratory blockmodeling because the analyst typically does not know the behavior of the individual objective functions beforehand. The a posteriori approach generates the approximated nondominated set and, subsequently, requires use of the analyst’s weights and preferences to choose one or more blockmodels from this set. The interactive approach is a multistep process that requires significant analyst participation to evaluate the quality of solutions during the multiobjective programming process. We adopt a perspective similar to one outlined by Ehrgott and Gandibleaux (Reference Erhgott and Gandibleaux2000, p. 448), which blends the a posteriori and interactive perspectives. The multiobjective blockmodeling algorithm is first used to establish a good approximation of the nondominated set. Subsequently, the preferences of the analyst would be used to focus the examination on specific regions of the approximated nondominated set by re-running the algorithm with an expanded set of weighting schemes. Although some criterion measures can be established to facilitate the selection of solutions from the nondominated set, the analyst’s ability to provide practical interpretation would remain an important component of real applications.
The output of the multiobjective tabu search algorithm is the approximated nondominated set, \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}}$\end{document}
. The size of this set can vary widely depending on the application; however, it generally tends to increase as a function of network size and the number of objective criteria (G). The evaluation of the nondominated set can range from the selection and interpretation of a single blockmodel from the set, to an evaluation of multiple blockmodels from various regions of the set. In the case of all minimization objectives, one simple criterion is to select the solution that is the minimum distance from the origin. Unfortunately, this approach does not take into account differences in measurement scales and deviation from the best possible values for the various objective functions. As an alternative, we adopt an approach proposed by Brusco and Stahl (Reference Brusco and Stahl2001) within the more general context of multiobjective combinatorial data analysis. This approach begins with the identification of the best-found (minimum) objective function value, \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$f_{g}^{*}$\end{document}
, for each criterion 1≤g≤G. These numbers are then used to establish an ideal point \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$(f_{1}^{*},\ldots,f_{G}^{*})$\end{document}
, from which the Euclidean distance from each blockmodel in \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}}$\end{document}
can be measured to that point. An analyst can select the blockmodel from \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$\pmb{\mathcal{B}}^{\mathrm{ND}}$\end{document}
based on its proximity to the ideal point or select the blockmodel with the smallest maximum distance across the G criteria.Footnote 6
3. Example 1. Sampson’s (1968) Monastery Data
3.1. The Network Matrix
For our first example of multiobjective blockmodeling, we use network data collected by Sampson (1968) corresponding to the relationships among trainee monks at a New England monastery. Blockmodel analyses of Sampson’s (1968) monastery data have been conducted several times in the network literature (Breiger et al. Reference Breiger, Boorman and Arabie1975; Brusco & Steinley Reference Brusco and Steinley2010; Doreian & Mrvar Reference Doreian and Mrvar1996; White, Boorman, & Breiger, Reference White, Boorman and Brieger1976). In our example, we consider the signed network matrix of affect ties among n=18 monks at time period T2. Each monk reported his three most-liked and three most-disliked monks and, accordingly, the elements of the signed network matrix, A=[a ij], range from −3 (most disliked) to +3 (most-liked). A value of a ij=0 indicates that monk i did not identify monk j as being either liked or disliked.
3.2. Structural Balance
We adopt the structural balance paradigm, based on Heider’s (Reference Heider1946) structural balance theory with important refinements by Cartwright and Harary (Reference Cartwright and Harary1956), Davis (Reference Davis1967), and Doreian and Mrvar (Reference Doreian and Mrvar1996), to analyze the monastery data. It posits that, ideally, the monks would be partitioned into distinct clusters of monks who have positive or neutral affect with other monks in their cluster, and either negative or neutral affect with monks outside their cluster. In other words, positive ties between monks should occur only within clusters and negative ties between monks occur only between clusters. As perfect structural balance seldom holds, Doreian and Mrvar offered an optimization approach to return partitions as close to those of exact structural balance as possible. Specifically, their approach identifies blockmodels by establishing a K-cluster partition of the monks that minimizes a weighted function of the G=2 types of inconsistencies: (a) negative ties within clusters and (b) positive ties between clusters. Denoting B={S 1,…,S K} as a blockmodel with S k representing the monks assigned to cluster k, these two objective criteria can be stated as follows:


By defining α (0≤α≤1) as the weight assigned to objective criterion f 1(B) and (1−α) as the weight assigned to criterion f 2(B), the single-objective, weighted-sum optimization problem studied by Doreian and Mrvar (Reference Doreian and Mrvar1996) is:
Doreian and Mrvar (Reference Doreian and Mrvar1996) used a relocation heuristic for obtaining solutions to the optimization problem posed by Equations (3) and (4). Brusco and Steinley (Reference Brusco and Steinley2010) recently presented an exact algorithm that can often obtain globally optimal solutions for signed networks with 40 or fewer objects. As for other weighted-sum approaches to multiobjective optimization, a key limitation of these algorithms is that they require specification of α. Although α=0.5 is a common assumption, this need not be the most appropriate alternative in all applications. For example, it might be much more disadvantageous from the standpoint of group cohesion to have members that dislike each other in the same cluster, than it is to have two members who like each other be in different clusters.Footnote 7 Thus, an important benefit of our proposed multiobjective method is that it avoids the prespecification of α. Moreover, it can uncover nondominated solutions that might not be recoverable using any value of α.
3.3. Results
Previous results for Sampson’s monastery data (see Doreian et al. Reference Doreian, Batagelj and Ferligoj2005 and Brusco & Steinley Reference Brusco and Steinley2010) have shown that, when minimizing the total number of inconsistencies, the optimal number of clusters for this network matrix is K=3. We used the branch-and-bound algorithm developed by Brusco and Steinley (Reference Brusco and Steinley2010) to initially populate the approximated nondominated set with optimal blockmodels for the single-objective optimization problems associated with Equations (1) and (2). The computation times to obtain these solutions were 0.03 and 0.02 seconds, respectively. Next, the multiobjective tabu search heuristic was applied to the network matrix for K=3 using two different weighting schemes, one coarse the other fine. The coarse weighting scheme was W=[(0.999,0.001),(0.001,0.999)], whereas the fine weighting scheme was W=[(0.999,0.001);(0.9,0.1);(0.8,0.2);(0.7,0.3),…,(0.3,0.7),(0.2,0.8),(0.1,0.9),(0.001,0.999)]. Both weighting schemes and parameters yielded the same nondominated set of 13 distinct blockmodels, shown in Table 1 with their objective function values and cluster assignments.Footnote 8 A visual display of the Pareto efficient frontier is provided in Figure 1.
The approximation of the Pareto efficient (nondominated) set for Sampson’s (1968) monastery affect data. The solid line connects the nine square markers that correspond to supported Pareto efficient blockmodels, whereas the four circular markers above the line correspond to the unsupported Pareto efficient blockmodels.

Results for Example 1—Sampson’s (1968) monastery data. The objective function values for the approximation of the Pareto efficient (nondominated) set of blockmodels. The results for the unsupported Pareto efficient blockmodels are highlighted in bold.

The solid lines on the Pareto efficient frontier in Figure 1 clearly distinguish the supported Pareto efficient blockmodels from those that are unsupported: the supported blockmodels are those connected by the line, whereas the unsupported blockmodels fall above the line. To illustrate the distinction, consider blockmodels #4, #5, and #6 from Table 1, which are displayed as the fourth square marker, first circular marker, and fifth square marker from left to right in Figure 1. Supported blockmodels #4 and #6 are connected by a solid line, but the unsupported blockmodel #5 lies just above that line. Assuming α=1/3, the weighted sum for blockmodel #4 is (1/3)(5)+(2/3)(30)=65/3, which is equal to the weighted sum for blockmodel #6 at (1/3)(9)+(2/3)(28)=65/3. The weighted sum for blockmodel #5 is slightly greater at (1/3)(8)+(2/3)(29)=66/3. Therefore, blockmodel #4 has the smallest weighted sum over the range 0≤α<1/3, blockmodels #4 and #6 have equal weighted sums at α=1/3, and blockmodel #6 has the smallest weighted sum over the range 1/3<α≤1. There is no value of α such that blockmodel #5 has the smallest weighted sum, and thus it is not supported by any weighting scheme.
Blockmodel #3 is precisely the same as the one reported by Doreian et al. (Reference Doreian, Batagelj and Ferligoj2005, p. 319), which minimizes the total number of inconsistencies at f 1(B)+f 2(B) (i.e., 2+33) =35. Blockmodel #4 (obtained by Breiger et al. Reference Breiger, Boorman and Arabie1975) also has 35 consistencies, and, accordingly, blockmodels #3 and #4 would be considered as equally well-fitting partitions at K=3 for a single-objective optimization problem focusing on minimizing the total sum of inconsistencies. These two blockmodels are displayed in Table 2. It is important to recognize, however, that these two blockmodels are only equally well-fitting under the assumption that the two types of inconsistencies are equally important (e.g., assuming α=0.5 in the weighted-sum context). For α<0.5, blockmodel #4 is preferred,Footnote 9 whereas blockmodel #3 would be superior for α>0.5.
Blockmodels #3 (top panel) and #4 (bottom panel) from Table 1 for Sampson’s (1968) monastery data. The K=3 boxes along the main diagonal of each matrix encapsulate the within-cluster elements. Inconsistencies (highlighted in bold) are the negative elements within the clusters and the positive elements between the clusters.

The Euclidean distance of each blockmodel from the ideal point also can be used by the analyst as part of the comparative process. The ideal point for the network matrix is (\documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$f_{1}^{*} = 0$\end{document}
, \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$f_{2}^{*} = 14$\end{document}
). Based on the Euclidean distances of the objective functions for each blockmodel to this point, blockmodel #6 is preferred with a distance of 16.64. Blockmodel #4 is a close second with a distance of 16.76 and might be favored over blockmodel #6 if there was a desire for a smaller value of f 1(B).
The analysis of the monastery network data demonstrates the principal advantage of the multiobjective tabu search algorithm, a full picture of the efficient frontier. This full picture enables the analyst to make an informed selection of one or more blockmodels.
4. Example 2. House Data
4.1. The Network Matrix
For our second example, we demonstrate multiobjective blockmodeling for finding a single blockmodel based on multiple network matrices for the same set of objects. For this example, we use data obtained by Lemann and Solomon (Reference Lemann and Solomon1952) that correspond to the measurement of four different types of network relations among female students at three different off-campus dormitories (houses) at an eastern college. The four relations pertained to (1) double-dating behavior, (2) maintaining friendship after college, (3) establishing a roommate relationship, and (4) setting up a weekend visit with family. To obtain the measurements for each relation, each woman was asked to identify the three women in the house with whom they would like to engage in the activity, as well as the three women with whom they would not like to engage in the activity. This data collection process results in three sets of matrices (one for each house), with each set containing four matrices (one for each relation). Following the convention of previous blockmodeling analyses of these data (Brusco et al. Reference Brusco, Doreian, Mrvar and Steinley2011; Doreian Reference Doreian2008), we label the sets as House A (n=21), House B (n=17), and House C (n=20). For each set, we use network matrices (A1,A2,A3,A4) measured for the G=4 different relations. A matrix element a ijg=1 if woman i listed woman j as one of the three women with whom they would like to do activity g, −1 if woman i listed woman j as one of the three women with whom they would not like to do activity g, and 0 otherwise.
4.2. Structural Balance Across Multiple Matrices
Once again, we adopt structural balance for this application; however, there are two salient differences from the example in Section 3. First, rather than measuring multiple fit measures for a single network matrix (as in Section 3), here we use the same fit measure for multiple network matrices (i.e., matrices measured for the double-dating, friendship, roommate, and weekend relations). Second, in this example, there are G=4 objective functions instead of two as in the example in Section 3. The four objective functions can be succinctly represented as follows:
Thus, the objective function for each network matrix is the number of inconsistencies associated with the blockmodel, which includes both negative elements within the clusters and positive elements between the clusters. The single-objective, weighted-sum optimization problem is:

4.3. Results
Previous results for Lemann and Solomon’s (Reference Lemann and Solomon1952) dormitory data have shown that a reasonable choice for the number of clusters is K=4. We began by using branch-and-bound to obtain, for each of the three houses, the single-objective optimal blockmodels for f 1(B), f 2(B), f 3(B), and f 4(B) independently. The four single-objective blockmodels for each house were used to initially populate the nondominated set for that house. The multiobjective tabu search heuristic was then applied to the set of network matrices for each house, assuming K=4 and using two different weighting schemes. Defining ε=0.001 and θ=ε/(G−1), the coarse weighting scheme was W=[(1−ε,θ,θ,θ),(θ,1−ε,θ,θ),(θ,θ,1−ε,θ),(θ,θ,θ,1−ε)], whereas the fine weighting scheme was W=[(1−ε,θ,θ,θ),(θ,1−ε,θ,θ,),(θ,θ,1−ε,θ),(θ,θ,θ,1−ε),(0.7 0.1 0.1 0.1),(0.4 0.2 0.2 0.2),(0.1 0.7 0.1 0.1),(0.2 0.4 0.2 0.2),(0.1 0.1 0.7 0.1),(0.2 0.2 0.4 0.2),(0.1 0.1 0.1 0.7),(0.2 0.2 0.2 0.4),(0.25 0.25 0.25 0.25)]. For each of the three houses, both weighting schemes yielded the same nondominated set of blockmodels. The computation times for the multiobjective tabu search algorithm when using the coarse weighting scheme were 0.18, 0.18, and 0.17 seconds for Houses A, B, and C, respectively. For the fine weighting scheme, the computation times were 0.14, 0.15, and 0.16 seconds for Houses A, B, and C, respectively.
In conjunction with the results for Sampson’s (1968) monastery data in the previous section, these results suggest that, for small network matrices, the multiobjective tabu search algorithm is not overly sensitive to the choice of weights. For each house, the objective function values and cluster assignments for each of the blockmodels in the nondominated set are shown in Table 3. The table also highlights (in bold) the blockmodels that would be obtained via aggregation of the four network relations into a single network matrix.
Results for Example 2—Lemann and Solomon’s (Reference Lemann and Solomon1952) dormitory data. The objective function values for the approximation of the Pareto efficient (nondominated) set of blockmodels. The results highlighted in bold are what would be obtained by fitting a blockmodel to an aggregation of the four network relations.

The number of blockmodels in the nondominated sets for House A, House B, and House C are 5, 16, and 6, respectively. The ideal point for House A is (\documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$f_{1}^{*} = 17$\end{document}
, \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$f_{2}^{*} = 15$\end{document}
, \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$f_{3}^{*} = 13$\end{document}
, \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$f_{4}^{*} = 15$\end{document}
). Based on the Euclidean distances of the objective functions for each blockmodel to this point, blockmodel A-3 is the best choice with a distance of 2.45. This blockmodel is also the one obtained by fitting a blockmodel to an aggregation of the four network relations into a single network matrix (Doreian Reference Doreian2008, p. 255). Similar results were obtained for House C, which has an ideal point of (\documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$f_{1}^{*} = 13$\end{document}
, \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$f_{2}^{*} = 14$\end{document}
, \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$f_{3}^{*} = 18$\end{document}
, \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$f_{4}^{*} = 15$\end{document}
). Blockmodel C-3, which is also what was obtained by aggregating the four network relations (Doreian Reference Doreian2008), has the minimum distance (3.74) from this ideal point. Thus, in the cases of Houses A and C, there are relatively few nondominated blockmodels, which is evident of a concordant structure among the four relations. Under such conditions, aggregation across multiple relations appears to offer a perfectly viable strategy.
By contrast, there was greater discordance among the network relations for House B, which resulted in a larger nondominated set. The principal source of discordance for House B is the roommate relation. Blockmodel B-14 yields the minimum objective function value of \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$f_{3}^{*} = 17$\end{document}
for the roommate relation but very poor objective function values for the other three relations. Other blockmodels, such as B-2, yield exceptional objective function values for the dating, friendship, and weekend relations but a poor objective function value for the roommate relation.
The ideal point for House B is (\documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$f_{1}^{*} = 18$\end{document}
, \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$f_{2}^{*} = 18$\end{document}
, \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$f_{3}^{*} = 17$\end{document}
, \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$f_{4}^{*} = 19$\end{document}
). Based on the Euclidean distances of the objective functions for each blockmodel to this point, blockmodel B-3 is the best choice with a distance of 6.63. This blockmodel is one of two equally well-fitting partitions obtained when applying a generalized blockmodeling algorithm to an aggregation of the four network relations into a single network matrix (Doreian Reference Doreian2008). The other equally well-fitting partition from the aggregation approach is blockmodel B-2, which has an appreciably greater distance of 9.06 from the ideal point. In fact, blockmodels B-4 and B-5 are closer to the ideal point than B-2. Thus, unlike Houses A and C where fitting a blockmodel to an aggregation of the relations resulted in arguably the best-fitting blockmodel in the nondominated set, the satisfactoriness of the aggregation approach is more ambiguous for House B. The results gleaned from the multiobjective tabu search heuristic indicate that one of the two equally well-fitting partitions (B-3) for the aggregated relations is an exceptionally good solution from the nondominated set, whereas the other (B-2) is not. Again, having the efficient frontier permits a more informed choice from a set of blockmodels. This suggests that the usual measures regarding the consistency of multiple relations are limited when one relation differs enough from the other indicators of the underlying signed relationship.
5. Example 3. Turning Point Project
5.1. The Network Matrices
For our third example, we demonstrate multiobjective blockmodeling using network data related to the Turning Point Project (TPP).Footnote 10 There were 108 organizations that collectively signed some radical advertisements regarding the environment in the New York Times (NYT) during 1999–2000 where different subsets of organizations signed different advertisements. It is reasonable to think that organizations signing advertisements together would show some consistency with those organizations sharing board members. When the board data were collected from their websites, it was clear that only 64 of them were involved in sharing board members. The following analysis is restricted to these 64 organizations. There are two network matrices associated with this application: (1) a shared-board network matrix and (2) an advertisement signing matrix. The first of these (A1) is a board interlock matrix of ties among n=64 organizations participating in the TPP and involved in sharing board members. A matrix element a ij1=1 if organization i shares board member relationships with organization j and a ij1=0 otherwise. The second matrix (A2) is a 64×25 two-mode matrix that reflects ties between organizations and full one-page advertisements in the NYT that they signed. More specifically, elements of this two-mode matrix assume values of a ij2=1 if organization i signed advertisement j and a ij2=0 otherwise. The 25 advertisements were in five categories defined by the organizers of the TPP. These categories were: Ecological catastrophe (EC), Economic globalization (EG), Genetic Engineering (EG), Industrial agriculture (IA), and Techno-mania (TM). We have taken this classification as given by the TPP for the following analysis. These categories provide insight into the natural or logical partitioning of the advertisements into underlying themes.
5.2. Short- vs. Long-Term Common Interests
This application of multiobjective blockmodeling is grounded by some assumptions regarding common interests among organizations in both the long and short terms. For example, if joint board memberships signal common interests, then these are long-term interests that are aligned given that board memberships are most often durable. The signing of ads is more likely to mobilize short-term interests. Nevertheless, as noted above, it is reasonable to anticipate some correspondence between the joint board memberships and the signing of ads, yet the extent is unclear. Multiobjective blockmodeling is a useful tool for examining the extent to which there is some degree of correspondence. We seek to find a partition of the 64 organizations that produces a well-fitting blockmodel for both the board interlock and ad signing network matrices. The objective function for the board interlock network matrix is
Thus, the objective function for the board interlock network is to minimize the sum of the absence of joint board membership ties within clusters, plus the sum of interlocked board ties that occur between clusters.
The objective function for the signed ad network is to minimize the total number of inconsistencies with an ideal block structure of null and complete blocks. More specifically, the objective function is the sum of 1s that occur in blocks that have mostly 0s, plus the sum of 0s that occur in blocks that have mostly 1s. This objective function is concordant with those used in recent applications of two-mode blockmodeling (Brusco & Steinley Reference Brusco and Steinley2011; Brusco, Doreian, Steinley, & Mrvar, in press; Doreian, Batagelj, & Ferligoj, Reference Doreian, Batagelj and Ferligoj2004). As noted previously, we assume that the ads are partitioned into L=5 five natural categories (C 1,C 2,…,C 5) defined by an underlying theme.Footnote 11 The objective function for the signed ad network matrix is
where
and
By defining w 1 and w 2 as objective function weights for the board interlock and signed ads network matrices, respectively, the single-objective, weighted-sum optimization problem is

5.3. Results
We began by applying 100 restarts of the bond-energy algorithm (McCormick, Schweitzer, & White, Reference McCormick, Schweitzer and White1972) to obtain a permuted version of A2 that would facilitate a reasonable choice of K. Visual inspection of the permuted matrix suggested the appropriateness of an 8-cluster model.Footnote 12 It was not possible to apply an exact solution method to obtain optimal solutions to the single-objective problems, and, therefore, we used a relocation heuristic algorithm to obtain the single-objective blockmodels used to initially populate the nondominated set. Next, the multiobjective tabu search heuristic was applied to the network matrix for K=8 using the same two weighting schemes described in Section 3.3. The coarse weighting scheme, W=[(0.999,0.001),(0.001,0.999)], yielded a nondominated set consisting of 73 blockmodels, whereas the fine weighting scheme, W=[(0.999,0.001),(0.9,0.1),(0.8,0.2),(0.7,0.3),…(0.3,0.7),(0.2,0.8),(0.1,0.9),(0.001,0.999)], produced 77 blockmodels. The total computation times for the multiobjective tabu search heuristic were 100.70 and 100.53 seconds for the coarse and fine weighting schemes, respectively.
We conducted a “cross-checking” of the results by comparing the approximated nondominated sets from the two weighting schemes to one another and discarding any blockmodel in one set that is dominated by one or more blockmodels in the other. When comparing the blockmodels obtained via the two-weighing schemes, it was found that 52 of the 73 blockmodels obtained for the coarse weighting scheme were actually nondominated. Contrastingly, all 77 of the blockmodels obtained with finer granularity in the weights were nondominated. The conclusion from this finding is that finer granularity can afford better estimation of the nondominated set for larger network matrices, particularly when there is discordance among the objective criteria. One approach that can be applied is to begin with a coarse weighing scheme and generate the nondominated set. If a satisfactory blockmodel is not identified, then the analyst can interactively use the method by restarting the algorithm with an expanded (finer) set of weights. For example, restarting the algorithm using the fine weighting scheme, but with the coarse weighting scheme blockmodels to initially populate the nondominated set, resulted in an increase in the number of blockmodels that were actually nondominated from 52 to 73.
The Pareto efficient frontier (using the finer granularity weighting scheme) for the TPP network is provided in Figure 2. Because of the disparity in the raw objective function values, Figure 2 displays the nondominated set in terms of percentage deviation above the ideal point.
The approximation of the Pareto efficient (nondominated) set for the Turning Point Project networks.

The ideal point for the network matrix is (\documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$f_{1}^{*} = 392$\end{document}
, \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$f_{2}^{*} = 165$\end{document}
). The minimization of one criterion to the exclusion of the other produces an extremely poor objective function value for the other criterion. For example, the partition of organizations yielding f 1=392 produces a value of f 2=321 (95 % greater than the ideal value). Similarly, the partition of organizations yielding f 2=165 produces a value of f 1=838 (114 % greater than the ideal value). The big differences in the single-objective solutions principally stem from the f 1 criterion exhibiting a propensity for similarly sized clusters, whereas the f 2 criterion is more amenable to differences in the cluster sizes. At face value, the results suggest considerable antagonism between the two objective criteria. This is important because it might point to a real difference between long-term interests (in terms of shared board memberships) and short-term interests (in the signing of specific ads). However, the sharply bending curve in Figure 2 reveals that very small sacrifices in f 1 will produce substantial improvements in f 2 and vice versa. This insight suggests that, even with antagonism between a pair of objective criteria, multiobjective blockmodeling is sufficiently flexible to handle this antagonism in a useful fashion.
Based on the criterion of Euclidean distance of the blockmodel from the ideal point (on a percentage basis), we selected a blockmodel with objective function values of f 1=486 and f 2=201, which are 24 % and 22 % greater than their ideal values, respectively. A visual representation of the blockmodel for the signed ads network is displayed in Figure 3. Figure 4 contains the number of inconsistencies that occur within and between clusters for the board interlock matrix. Along the main diagonal of Figure 4 is the sum of the total number of absences of board ties among organizations in that cluster. The entries off the main diagonal contain the number of ties that occur between organizations in two different clusters. Each cell also expresses (in parentheses) the number of inconsistencies as a percentage of the total possible ties that could have occurred. The percentages along the main diagonal seem large at first glance, averaging 67.2 %. However, these figures should be considered in light of the fact that the sparseness of the board interlock matrix results in a minimum possible percentage average of 55 %.
Blockmodel of 2-mode Turning Point Project network.

Inconsistencies for the board interlock matrix. Each cell contains the number of inconsistencies observed for each pair of clusters. Along the main diagonal, the entries are the total absences of board ties among organizations within the same cluster. The above-diagonal elements are the total board ties between organizations in different clusters (the lower triangle of the matrix is the mirror image of the upper triangle). The percentage of inconsistencies relative to the total number of elements is shown in parentheses.

The first cluster consists of six organizationsFootnote 13 (“eii”, “foe”, “gpus”, “icta”, “ran”, “fde”) that share a total of four joint board membership ties among them. Most notably, “fde” and “eii” share board memberships with one another and are each linked to two other organizations in the cluster. Figure 3 reveals that these six organizations are linked to the signing of four of the five groups of advertisements (EC, GE, EG, and TM). The second cluster consists of eight organizations (“ala”, “cbd”, “dw”, “esc”, “nfpa”, “nfn”, “mw”, “wp”) that tend to sign only ads in the EC group.Footnote 14 The organizations in this cluster share a total of six joint board memberships among them, and “mw” contributes three of these ties.
Cluster three is the only other cluster (besides cluster one) that heavily signs ads in more than one group. The third cluster consists of eight organizations (“cfs”, “crg”, “faw”, “ff”, “iatp”, “isec”, “oca”, “pan”) that are heavy signers of ads in both the GE and IA groups. The organizations in this cluster share a total of four joint board memberships among them, mainly stemming from “cfs”. Clusters four, five, and six are heavy signers of ads in groups IA, EG, and TM, respectively. Cluster five is especially noteworthy among these, as it is a cluster of eight organizations that share a total of 10 joint board memberships. Accordingly, the organizations in this cluster seem especially concordant in terms of sharing both long-term (joint board membership) and short-term (ad signing) ties. The seventh and eighth clusters each share a reasonable degree of joint board membership ties; however, neither of these two clusters consistently endorsed ads in a particular category.
The analysis of the TPP network data reveals that, again, the multiobjective tabu search heuristic provides a solution that minimizes the distance from the ideal point when the objectives are considered simultaneously and breaks ties between seemingly comparable blockmodels. Additionally, the analysis emphasizes the importance of granularity in approaching large network problems. Finally, multiobjective blockmodeling offers the analyst a way to reconcile and solve problems that include highly antagonistic networks.
6. Simulation Results and Recommendations for Practice
6.1. Simulation Study 1
The evaluation of multiobjective programming methods is nontrivial (Erhgott & Gandibleaux Reference Erhgott and Gandibleaux2000). The consistency and interpretability of the results obtained by the multiobjective tabu search heuristic for blockmodeling in the examples provided in Sections 3, 4, and 5 offer some positive evidence of its effectiveness. To further substantiate the merits of the multiobjective tabu search approach for blockmodeling, we conducted a simulation study to ensure that the method performs as intended for problems where at least a portion of the nondominated set is known with certainty.
The context of the simulation study was biobjective blockmodeling of a signed network matrix based on structural balance theory, the same problem considered in Section 3. We generated 20 signed network matrices of size 20×20. All test problems were generated under the assumption of four clusters, each containing five objects. For the first 10 problems, the probability of a positive tie between two objects in the same cluster was 60 %, the probability of a negative tie was 10 %, and the probability of no tie was 30 %. Similarly, the probability of a negative tie between two objects in different clusters was 60 %, the probability of a positive tie was 10 %, and the probability of no tie was 30 %. For Problems 11 through 20, the probabilities of the correct/incorrect signs within and between clusters were changed from 60 %/10 % to 50 %/20 %, and, therefore, these latter 10 problems were less well structured and would presumably have a larger nondominated set.
Multiobjective blockmodeling solutions for each of the 20 test problems were obtained under four different versions for the heuristic: (a) a coarse weighting scheme of W=[(0.999,0.001),(0.001,0.999)] and using a random blockmodel to initially populate the nondominated set, (b) a fine weighting scheme of W=[(0.999,0.001),(0.9,0.1),(0.8,0.2),…,(0.2,0.8),(0.1,0.9),(0.001,0.999)] and using a random blockmodel to initially populate the nondominated set, (c) the coarse weighting scheme using the single-objective optimal solutions to initially populate the nondominated set, and (d) the fine weighting scheme using the single-objective optimal solutions to initially populate the nondominated set. For each test problem and each of the four versions, data were collected with respect to both the approximated nondominated set and computation time. For each test problem, the nondominated sets obtained under the four versions were compared to one another, and any blockmodels that were dominated by a blockmodel obtained under one or more of the other versions were discarded. A summary of the results is provided in Table 4.
Results for 20 test problems in simulation Study 1. The “B&B” columns contains the computation time for the branch-and-bound algorithm to obtain the single-objective optimal blockmodels used to initially populate the nondominated set. The four versions of the multiobjective blockmodeling algorithm are: (a) weighting scheme of [1,0;0,1] and using a random blockmodel to initially populate the nondominated set, (b) weighting scheme of [1,0;0.9,0.1;0.8,0.2;…0,1] and using a random blockmodel to initially populate the nondominated set, (c) weighting scheme of [1,0;0,1] and using the single-objective optimal solutions to initially populate the nondominated set, and (d) weighting scheme of [1,0;0.9,0.1;0.8,0.2;…0,1] and using the single-objective optimal solutions to initially populate the nondominated set. Values in bold type in the last four columns indicate that all of the supported Pareto efficient blockmodels were found.

The results in Table 4 reveal that the computation times for multiobjective tabu search algorithm were very modest, regardless of the test problem characteristics, weighting scheme, and initial population of the nondominated set. It is particularly interesting to observe that, despite the fact that the sizes of the nondominated sets for Problems 11 through 20 were much larger than those for Problems 1 through 10, the computation times were comparable across these two sets of problems. Most of the computation times were less than one second. By contrast, much greater computation time variability is observed for the branch-and-bound algorithm used to obtain the single-objective optimal blockmodels for populating the initial nondominated set for versions (c) and (d). The branch-and-bound algorithm is sensitive to the data characteristics. The computation times for Problems 11 through 20 are typically greater because they are less well structured (i.e., have more inconsistencies) than Problems 1 through 10.
Because the test problems are limited to n=20 objects, we were able to use the branch-and-bound algorithm under different weighting schemes (i.e., [(0.999,0.001),(0.9,0.1),(0.8,0.2),…,(0.2,0.8),(0.1,0.9),(0.001,0.999)]) to obtain “supported” Pareto efficient solutions. This enabled us to obtain an objective measure of how well each of the four versions performed with respect to the recovery of blockmodels that are guaranteed to be supported Pareto efficient solutions in the dominated set. Heuristic versions (d) and (c) performed exceptionally well, finding all of the supported Pareto efficient solutions for all 20 test problems, respectively. Versions (a) and (b) failed to obtain all of the Pareto efficient solutions for a total of 5 and 7 problems, respectively. With one exception (version (b) on Problem 19), all four versions obtained all Pareto efficient solutions for Problems 11 through 20 and were always comparable in terms of the size of the approximated nondominated set for these problems. The different versions of the heuristic exhibited more variability in their nondominated sets for Problems 1 through 10. In light of the failure of versions (a) and (b) to identify the complete set of Pareto efficient solutions for several of these problems, we conclude the initial population of the nondominated set is a vital component of our methodology. There was no pronounced difference in performance between version (c) and (d), and, accordingly, the coarse weighting scheme was sufficient for these smaller problems.
6.2. Simulation Study 2
To determine whether the required granularity of the weighting scheme might be a function of problem size, we repeated simulation Study 1 after increasing the number of objects from n=20 to n=100. All test problems were generated under the assumption of four clusters, each containing 25 objects. The same probabilities used in Study 1 were used for data generation to produce 20 test problems. The nondominated set was initially populated using 1000 restarts of a relocation heuristic for the single-objective problems. The multiobjective tabu search heuristic was then applied using two different weighting schemes: (1) coarse, W=[(0.999,0.001),(0.001,0.999)], and (2) fine, W=[(0.999,0.001),(0.9,0.1),(0.8,0.2),…,(0.2,0.8),(0.1,0.9),(0.001,0.999)]. The two weighting schemes were compared in terms of computation time, as well as the size of the nondominated sets after cross-checking the results of the two weighting schemes and discarding dominated blockmodels. The results are reported in Table 5.
Results for 20 test problems in simulation Study 2. The “RH” column contains the computation time for 1000 restarts of the relocation heuristic used to obtain the single-objective optimal blockmodels used to initially populate the nondominated set. The two heuristic versions correspond to coarse weights: [(0.999,0.001),(0.001,0.999)] and fine weights [(0.999,0.001),(0.9,0.1),(0.8,0.2),…(0.2,0.8),(0.1,0.9),(0.001,0.999)]. The term “cross-checking” refers to comparison of the approximated nondominated sets from the two weighting schemes to one another and the discarding of any blockmodel in one set that is dominated by one or more blockmodels in the other.

There are three important aspects with respect to the computation time results in Table 5. First, the computation times for the multiobjective tabu search algorithm are drastically larger than those in Table 4, which indicates that it is sensitive to the size of the network. Second, the computation times for Problems 11 through 20 are roughly two to three times greater than those for Problems 1 through 10, which suggests that the form of structure in the network affects computational time and network size. Third, the computation times for the fine weighting scheme are, in most cases, appreciably less than those for the coarse weighting scheme. At first glance, this seems counterintuitive. However, it is attributable to the fact that the coarse weighting scheme typically fleshes out the nondominated set more slowly than the fine weighting scheme. Once several cycles at each set of weights in the fine weighting scheme have occurred, the frontier is generally well approximated. Contrastingly, with the coarse scheme, the algorithm must make many cycles, gradually adding nondominated blockmodels to the frontier.
The results in Table 5 also reveal that a finer weighting scheme can often afford significant improvements in the approximated nondominated set for larger problems. The coarse and fine weighting schemes initially establish comparably sized nondominated sets. However, after cross-checking between these two sets, the nondominated set for the fine weighting scheme is larger for 16 of the problems and very close in number to the coarse weighting scheme on three of the remaining four problems (i.e., Problems 9, 10, and 19). By contrast, cross-checking often substantially reduces the number of blockmodels in the nondominated set for the coarse weighting scheme, sometimes by more than 50 %. Problems 1, 2, 7, 11, and 18 are among those where the nondominated set of the coarse weighting scheme is severely pruned by the cross-checking process. Consider, for example, Problem 2, where the initial sizes of the nondominated set are comparable at 289 and 313 for the coarse and fine weighting schemes, respectively. After cross-checking, only 13 of the blockmodels in the fine set are dominated by blockmodels in the coarse set, thus reducing the number for the fine weighting scheme from 313 to 300. By comparison, 159 blockmodels in the coarse set are dominated by blockmodels from the fine set, thus reducing the number for the coarse weighting scheme from 289 to 130.
6.3. Recommendations for Practice
The results of the first two numerical illustrations and simulation Study 1 suggest that the multiobjective tabu search algorithm for blockmodeling performs comparably for both fine and coarse weighting schemes. The finer weighting scheme only offered a significant advantage in the cases of the Turning Point Project example and simulation Study 2, which are larger networks with greater discordance between the objective functions. The results of both simulation studies and the Turning Point Project example indicate that there are distinct improvements in the nondominated set that can be realized by initially populating the nondominated set with (near-)optimal blockmodels for each objective function independently. Based on these results, we recommend the following steps for an analyst using the applied multiobjective blockmodeling procedure.
Step 1. When possible, obtain an optimal blockmodel for each objective function independently and use these to initially populate the nondominated set. If this is not practical, then use a heuristic procedure to obtain near-optimal blockmodels for the initial population.
Step 2. Run the multiobjective tabu search algorithm using the coarse weighting scheme W=[(1−ε,θ,…,θ),(θ,1−ε,θ,…,θ),…,(θ,…,θ,1−ε)], where ε is a small positive constant (0<ε≤0.001), and θ=ε/(G−1). In other words, there are G sets of weights, each associated with a weight of (1−ε) for one criterion and an equal spread of weights of θ=ε/(G−1) among the other G−1 criteria. In our applications, using ε=0.001 was always sufficient to ensure that no improvement in the G−1 criteria weighted at θ would be worth a sacrifice of the criterion weighted (1−ε).
Step 3. The analyst’s preferences for tradeoffs among the competing criteria are applied to the approximate nondominated set obtained in Step 2 to select a blockmodel. To facilitate this process, the analyst could be presented with Euclidean distances from the ideal point, perhaps after some normalization of the objective functions (see Brusco & Stahl Reference Brusco and Stahl2001).
Step 4. If the analyst identifies a suitable blockmodel (or set of blockmodels) in Step 3, then the selection process is complete. Otherwise, the multiobjective tabu search algorithm can be reapplied using the approximated nondominated set (or, possibly, a select region of that set) from Step 2 as the initial population, as well as an expansion of the weighting scheme to a finer set of weights. The updated frontier is then reevaluated via a return to Step 3.
7. Conclusions
7.1. Summary
We presented a multiobjective programming model for blockmodeling of network matrices. A general multiobjective tabu search algorithm was offered as a plausible method for direct approximation of the Pareto efficient frontier. Among the benefits of our proposed model and method are: (1) the ability to obtain blockmodels that fit well with respect to multiple blockmodeling criteria, particularly when relationships among those criteria are antagonistic, (2) mitigation of the problem of multiple equally well-fitting partitions, (3) avoidance of the need to prespecify weights for different objective functions. The multiobjective programming model and method can be adapted for many different types of blockmodeling applications. In this paper, we have reported results for three applications. The first application (Monastery data) was to a well-known signed network corresponding to affect ties among 18 monks at a monastery (Sampson 1968). Using the sum of negative affect ties within groups and the sum of positive affect ties between groups as the objective criteria, the multiobjective tabu search algorithm was applied to the network matrix. For this small network, the algorithm was insensitive to parameter choices, rapidly generating the Pareto efficient set of blockmodels and enabling the analyst to select blockmodels from the set without the need to prespecify a weighting scheme.
The second application (House data) focused on finding a partition of actors that yielded a good blockmodel fit for G=4 distinct network relations. The relations pertain to double-dating, post-college friendship, potential roommate, and weekend outing ties among women in off-campus dormitories in an eastern college. Using multiobjective blockmodeling, it was possible to rapidly estimate the nondominated set of blockmodels and identify partitions that yield reasonable tradeoffs among the different relations.
The third application (TPP data) used multiobjective blockmodeling to find a partition of organizations that maximized joint board memberships within clusters, as well as shared signing of ads in the Turning Point Project. Thus, like the House data in the second application, the focus was on multiple network relations for the same objects. However, unlike the House data, the relationship between the two criteria in the TPP data was highly antagonistic. That is, the optimal partitions for the board interlock network yielded poor results for the signed ads network and vice versa. Nevertheless, multiobjective blockmodeling identified blockmodels that afford excellent tradeoffs between the two criteria.
7.2. Limitations
Among the inherent limitations of the multiobjective tabu search heuristic presented in this paper are: (1) the potential need to reset parameters, (2) the need to prespecify a weighting scheme, (3) the fact that the nondominated set is only approximated, and (4) the challenge of choosing solutions from the Pareto efficient set. The parameters selected seem fairly robust across different network sizes, objective functions, and objective function configurations. The weighting scheme selection is somewhat more important; however, we offer a structured approach in Section 6.3. Like most heuristics for combinatorial optimization problems, the proposed multiobjective tabu search algorithm is not guaranteed to obtain globally optimal solutions, and, accordingly, we are careful to point out that, in general, the algorithm only approximates the Pareto efficient frontier. For our applications to Sampson’s (1968) monastery data and Lemann and Solomon’s (1952) dormitory data, we were able to use a previously published branch-and-bound algorithm (Brusco & Steinley Reference Brusco and Steinley2010) to verify the supported Pareto efficient solutions on the frontier, but this is typically not possible for larger networks. Finally, the selection of a blockmodel from the Pareto efficient set will almost certainly require some information that reflects the preferences of the analyst. We offer some simple measures of distance from the ideal point to help facilitate this task.
7.3. The Larger Framework of Multiobjective Clustering
Multiobjective blockmodeling can be perceived as a special case of multiobjective clustering (see Ferligoj & Batagelj Reference Ferligoj and Batagelj1992, for an overview). Multiobjective clustering applications date back, at least, to the work of Delattre and Hansen (Reference Delattre and Hansen1980), who presented an exact algorithm for a biobjective partitioning problem with objectives of diameter and split. Brusco and Stahl (Reference Brusco and Stahl2005, Chapter 6) and Brusco and Steinley (Reference Brusco and Steinley2009b) discuss exact solution methods for biobjective clustering problems with within-cluster sum-of-squares objective functions for two distinct sets of clustering variables. These clustering applications are similar to those encountered in multiobjective market segmentation (Brusco, Cradit, & Stahl, Reference Brusco, Cradit and Stahl2002; Brusco, Cradit, & Tashchian, Reference Brusco, Cradit and Tashchian2003; DeSarbo & Grisaffe Reference DeSarbo and Grisaffe1998), where simulated annealing approaches have been especially popular. More recently, genetic/evolutionary algorithms have proved successful in a variety of clustering contexts (Handl, Kell, & Knowles, Reference Handl, Kell and Knowles2007; Handl & Knowles Reference Handl and Knowles2007; Liu, Ram, Lusch, & Brusco, Reference Liu, Ram, Lusch and Brusco2010; Ripon, Tsang, Kwong, & Ip, Reference Ripon, Tsang, Kwong and Ip2006; Saha & Bandyopadhyay Reference Saha and Bandyopadhyay2010) and could potentially be adapted as a viable alternative for multiobjective blockmodeling. Accordingly, one possible extension to our paper is the design of alternative heuristic procedures for direct estimation of the Pareto efficient set and the subsequent comparison of their estimated frontiers to those obtained by our algorithm.
It is also important to observe that, just as other multiobjective programming approaches for clustering can be adapted for the blockmodeling problems studied herein, the multiobjective tabu search heuristic presented herein can be adapted for other clustering problems. For example, it might be interesting to compare an adaptation of the MOTS paradigm for multiobjective market segmentation to some of the simulated annealing, genetic algorithm, and local-search heuristic procedures used in previous studies (Brusco et al., Reference Brusco, Cradit and Stahl2002, Reference Brusco, Cradit and Tashchian2003; DeSarbo & Grisaffe Reference DeSarbo and Grisaffe1998; Liu et al. Reference Liu, Ram, Lusch and Brusco2010).
7.4. Extensions to Stochastic Blockmodeling
We have framed our coverage of multiobjective blockmodeling within the context of deterministic blockmodeling. Our multiobjective programming model formulation is clearly a discrete optimization problem, whereas stochastic blockmodeling procedures typically rely on continuous (or smooth) optimization procedures such as the Newton–Raphson method, the E–M algorithm (Dempster, Laird, & Rubin, Reference Dempster, Laird and Rubin1977), Markov chain Monte Carlo estimation (Gilks, Richardson, & Spiegelhalter, Reference Gilks, Richardson and Spieglhalter1996), or iteratively reweighted least squares (Wasserman & Pattison Reference Wasserman and Pattison1996). Nevertheless, there is recent work in stochastic blockmodeling where it appears that researchers are now using combinatorial algorithms to optimize likelihood-based optimization criteria (Karrer & Newman Reference Karrer and Newman2011; Steinley et al. Reference Steinley, Brusco and Wasserman2011). In light of these developments, it seems possible that our proposed multiobjective programming based framework might hold some promise for stochastic blockmodeling as well. For example, suppose that two likelihood-based criteria were under consideration for a stochastic blockmodeling application and a discrete optimization solution procedure for fitting the blockmodel was deemed appropriate. Our multiobjective programming approach would be viable for such an application. The key here is that the feasibility of the approach is not necessarily contingent on the deterministic vs. stochastic distinction, but rather whether the model fitting process is implemented using discrete (combinatorial optimization) or continuous (smooth) optimization procedures.
Acknowledgements
We are grateful to the Editor, Associate Editor, and three anonymous reviewers for their thoughtful and constructive comments, which led to a substantially improved paper. Douglas Steinley’s work was supported in part by NIAAA Grant 5K25AA017456.








