Next Article in Journal
Coloring Invariant for Topological Circuits in Folded Linear Chains
Next Article in Special Issue
From the Quasi-Total Strong Differential to Quasi-Total Italian Domination in Graphs
Previous Article in Journal
Mathematical Algorithm for Identification of Eukaryotic Promoter Sequences
Previous Article in Special Issue
Study of Parameters in the Genetic Algorithm for the Attack on Block Ciphers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improving Multivariate Microaggregation through Hamiltonian Paths and Optimal Univariate Microaggregation

1
Department of Computer Engineering and Mathematics, Universitat Rovira i Virgili, Av. Països Catalans 26, 43007 Tarragona, Catalonia, Spain
2
Department of Informatics, University of Piraeus, Karaoli kai dimitriou 80, 18534 Pireas, Greece
*
Author to whom correspondence should be addressed.
Symmetry 2021, 13(6), 916; https://doi.org/10.3390/sym13060916
Submission received: 1 April 2021 / Revised: 4 May 2021 / Accepted: 14 May 2021 / Published: 21 May 2021
(This article belongs to the Special Issue Theoretical Computer Science and Discrete Mathematics)

Abstract

:
The collection of personal data is exponentially growing and, as a result, individual privacy is endangered accordingly. With the aim to lessen privacy risks whilst maintaining high degrees of data utility, a variety of techniques have been proposed, being microaggregation a very popular one. Microaggregation is a family of perturbation methods, in which its principle is to aggregate personal data records (i.e., microdata) in groups so as to preserve privacy through k-anonymity. The multivariate microaggregation problem is known to be NP-Hard; however, its univariate version could be optimally solved in polynomial time using the Hansen-Mukherjee (HM) algorithm. In this article, we propose a heuristic solution to the multivariate microaggregation problem inspired by the Traveling Salesman Problem (TSP) and the optimal univariate microaggregation solution. Given a multivariate dataset, first, we apply a TSP-tour construction heuristic to generate a Hamiltonian path through all dataset records. Next, we use the order provided by this Hamiltonian path (i.e., a given permutation of the records) as input to the Hansen-Mukherjee algorithm, virtually transforming it into a multivariate microaggregation solver we call Multivariate Hansen-Mukherjee (MHM). Our intuition is that good solutions to the TSP would yield Hamiltonian paths allowing the Hansen-Mukherjee algorithm to find good solutions to the multivariate microaggregation problem. We have tested our method with well-known benchmark datasets. Moreover, with the aim to show the usefulness of our approach to protecting location privacy, we have tested our solution with real-life trajectories datasets, too. We have compared the results of our algorithm with those of the best performing solutions, and we show that our proposal reduces the information loss resulting from the microaggregation. Overall, results suggest that transforming the multivariate microaggregation problem into its univariate counterpart by ordering microdata records with a proper Hamiltonian path and applying an optimal univariate solution leads to a reduction of the perturbation error whilst keeping the same privacy guarantees.

1. Introduction

Knowledge retrieval and data processing are catalysts for innovation. The continuous advances in information and communication technologies (ICT) and the efficient processing of data allow the extraction of new knowledge by discovering non-obvious patterns and correlations in the data. Nevertheless, such knowledge extraction procedures may threaten individuals’ privacy if the proper measures are not implemented to protect it [1,2,3]. For instance, an attacker may use publicly available datasets to obtain insights about individuals and extract knowledge by exploiting correlations that were not obvious from examining a single dataset [4]. Therefore, before disclosing any data, privacy protection procedures (e.g., anonymization, pseudonymization, aggregation, generalization) must be applied. A wide variety of privacy models and protection mechanisms have been proposed in the literature so as to guarantee anonymity (at different levels depending on the utilized model) when disclosing data [5]. Since most privacy protection methods are based on modifying/perturbing/deleting original data, their main drawback is that they negatively affect the utility of the data. Hence, there is a need for finding a proper trade-off between data utility and privacy.
One of the most well-known disciplines studying methods to protect individuals’ private information is Statistical Disclosure Control (SDC [6]), which seeks to anonymize microdata sets (i.e., datasets consisting of multiple records corresponding to individual respondents) in a way that it is not possible to re-identify the respondent corresponding to any particular record in the published microdata set. Microaggregation [7], which perturbs microdata sets by aggregating the attributes’ values of groups of k records so as to reduce re-identification risk by achieving k—anonymity, stands out among the most widely used families of SDC methods. It is usually applied by statistical agencies to limit the disclosure of sensitive microdata, and it has been used to protect data in a variety of fields, namely healthcare [8], smart cities [9], or collaborative filtering applications [10], to name a few.
Although the univariate microaggregation problem can be optimally solved in polynomial time, optimal multivariate microaggregation is an NP-hard problem [11]. Thus, finding a solution for the multivariate problem requires heuristic approaches that aim to minimize the amount of data distortion (often measured in terms of information loss), whilst guaranteeing a desired privacy level (typically determined by a parameter k that defines the cardinality of the aggregated groups).

1.1. Contribution and Research Questions

In this article, we propose a novel solution for the multivariate microaggregation problem, inspired by the heuristic solutions of the Traveling Salesman Problem (TSP) and the use of the optimal univariate microaggregation algorithm of Hansen and Mukherjee (HM) [12]. Given an ordered numerical vector, the HM algorithm creates the optimal k-partition (i.e., the optimal univariate microaggregation solution). Hence, our intuition is that, if we feed the HM algorithm with a good ordering of the records in a multivariate dataset, it would output a good k-partition of the multivariate dataset (although not necessarily optimal).
Ordering the records of a univariate dataset is trivial. However, ordering those records in a multivariate dataset, in which every record has p attributes, is not obvious since it is not apparent how to determine the precedence of an element over another. Thus, the primary question is:
Q1:How to create this ordering, when the records are in R p .
We suggest that a possible order for the records in R p is determined by the Hamiltonian path resulting from solving the Traveling Salesman Problem, in which the goal is to find the path that travels through all elements of a set only once, whilst minimizing the total length of the path. Optimally solving the TSP is known to be NP-Hard, but very good heuristic solutions are available. Hence, our intuition is that good heuristic solutions of the TSP (i.e., those with shorter path lengths) would provide a Hamiltonian path, that could be used as an ordered vector for the HM optimal univariate microaggregation algorithm, resulting in a good multivariate microaggregation solution.
The quality of a TSP solution is measured in terms of “path length”, the shorter the length the better the solution. However, the quality of the microaggregation is measured in terms of information loss. Given a cardinality parameter k (which sets the minimum size of the aggregation clusters), the lower the information loss, the better the microaggregation. Hence, the next questions that we aim to answer are:
Q2:Are the length of the Hamiltonian path and the information loss of the microaggregation related?, or Do shorter Hamiltonian paths lead to microaggregation solutions with lower information loss?
and
Q3:Is the length of the Hamiltonian path the only factor affecting information loss or does the particular construction of the path (regardless of the length) affect the information loss?
Overall, the key question is:
Q4:Does this approach provide better solutions (in terms of information loss) than the best performing microaggregation methods in the literature?
In order to answer these questions, we have tested seven TSP solvers, combined with the HM algorithm (virtually applied in a multivariate manner, or Multivariate HM (MHM)). Particularly, we have tested the “Concorde” heuristic, which, to the best of our knowledge, is the first time it is used for microaggregation. In addition, we have tested well-known classic microaggregation methods (i.e., MDAV and V-MDAV), and an advanced refinement of the former (i.e., MDAV-LK-MHM).
With the aim to test all the aforementioned approaches on a variety of datasets, we have used three classical benchmarks (i.e., Census, Tarragona, and EIA) and three novel datasets containing trajectory data retrieved from public sources, which lead to our last research question:
Q5:Do TSP-based microaggregation methods perform better than current solutions on trajectories datasets?

1.2. Plan of the Article

The rest of the article aims to answer the research questions above, and it is organized as follows: Section 2 provides the reader with some fundamental knowledge on Statistical Disclosure Control and microaggregation. In addition, it introduces the basics of the Traveling Salesman Problem and an overview of the existing heuristics to solve it. Next, Section 3 analyzes related work and highlights the novelty of our proposal compared with the state of the art. Section 4 describes our proposal, which is later thoroughly tested and compared with well-known classical and state-of-the-art microaggregation methods in Section 5. Section 6 discusses the research questions and the main benefits of our proposal. The article concludes in Section 7 with some final remarks and comments on future research lines.

2. Background

2.1. Statistical Disclosure Control and Microaggregation

Statistical disclosure control (SDC) has the goal of preserving the statistical properties of datasets, whilst minimizing the privacy risks related to the disclosure of confidential information from individual respondents. Microaggregation is a family of SDC methods for microdata, which use data perturbation as a protection strategy.
Given an original data file D and a privacy parameter k, microaggregation can be defined as follows: Let us assume a microdata set D with p continuous numerical attributes and n records. Clusters (also referred to as groups or subsets in this context) of D are formed with n i records in the i-th cluster ( n i k and n = i = 1 g n i ), where g is the number of resulting clusters, and k a cardinality constraint. Optimal microaggregation is defined as the one yielding a k-partition maximizing the within-clusters homogeneity. Optimal microaggregation requires heuristic approaches since it is an NP-hard problem [11] for multivariate data. Microaggregation heuristics can be classified into two main families:
  • Fixed-size microaggregation: These heuristics cluster the elements of D into k-partitions where all clusters have size k, except perhaps one group which has a size between k and 2 k 1 , when the total number of records is not divisible by k.
  • Variable-size microaggregation: These heuristics cluster the elements of D into k-partitions where all clusters have sizes in ( k , 2 k 1 ) . Note that it is easy to show that any cluster with size larger than ( 2 k 1 ) could be divided in several smaller clusters of size between k and 2 k 1 in which its overall within-cluster homogeneity is better than that of the single larger cluster.
Therefore, a microaggregation process consists in constructing a k-partition of the dataset, this is a set of disjoint clusters (in which the cardinality is between k and 2 k 1 ) and replacing each original data record by the centroid (i.e., the average vector) of the cluster to which it belongs, hence creating a k-anonymous dataset D . With the aim to reduce the information loss caused by the aggregation, the clusters are created so that the records in each cluster are similar.

2.2. Data Utility and Information Loss

The sum of square error (SSE) is commonly used for measuring the homogeneity in each group. In terms of sums of squares, maximizing within-groups homogeneity is equivalent to finding a k-partition minimizing the within-groups sum of square error (SSE) [13] defined as:
S S E = i = 1 g j = 1 n i ( x i , j x i ¯ ) ( x i , j x i ¯ ) ,
where x i , j is the j-th record in group i, and x i ¯ is the average record of group i. The total sum of squares (SST), an upper bound on the partitioning information loss, can be computed as follows:
S S T = i = 1 n ( x i x ¯ ) ( x i x ¯ ) ,
where x i is the i-th record in D, and x i ¯ is the average record of D. Note that all the above equations use vector notation, so x i R p .
The microaggregation problem consists in finding a k-partition with minimum SSE, this is, the set of disjoint subsets of D so that D = m = 1 g s m , where s m is the m-th subset, and g is the number of subsets, with minimum SSE. However, a normalized measure of information loss (expressed in percentage) is also used:
I l o s s = S S E S S T × 100 .
In terms of information loss, the worst case scenario for microaggregation would happen when all records in D are replaced in D by the average of the dataset (i.e., S S E = S S T I l o s s = 100 ), and the best case scenario implies that D = D (i.e., k = 1 , no aggregation), which leads to S S E = I l o s s = 0 . Obviously, the latter case is optimal in terms of information loss, but it offers no privacy protection, at all. Hence, values for the protection parameter k are greater than one, typically: k = 3 , 4, 5, or 6, and are chosen by privacy experts in statistical agencies so as to adapt to the needs of each particular dataset.

2.3. Basics on the Traveling Salesman Problem

The Traveling Salesman Problem (TSP) [14] consists of finding a particular Hamiltonian cycle. The problem can be stated as follows: a salesman leaves from one city and wants to visit (exactly once) each other city in a given group and, finally, return to the starting city. The salesman wonders in what order he should visit these cities so as to travel the shortest possible total distance.
In terms of graph theory, the TSP can be modeled by a graph G = ( V , E ) , where cities are the nodes in set V = { v 1 , v 2 , . . . , v n } and each edge e i j E has an associated weight w i j representing the distance between nodes i and j. The goal is to find a Hamiltonian cycle, i.e., a cycle which visits each node in the graph exactly once, with the least total weight. An alternative approach to the Hamiltonian cycle to solve the TSP is finding the Shortest Hamiltonian path through a graph (i.e., a path which visits each node in the graph exactly once). As an example, Figure 1 shows a short Hamiltonian path for the Eurodist dataset, which contains the distance (in km) between 21 cities in Europe.
Finding an optimal solution to the TSP is known to be NP-Hard. Hence, several heuristics to find good but sub-optimal solutions have been developed. TSP heuristics typically fall into two groups: those involving minimum spanning trees for tour construction and those with edge exchanges to improve existing tours. There are numerous heuristics to solve the TSP [15,16]. In this article, we have selected a representative sample of heuristics, including well-known approaches and top performers from the state-of-the-art:
  • Nearest Neighbor algorithm: The algorithm starts with a tour containing a randomly chosen node and appends the next nearest node iteratively.
  • Repetitive Nearest Neighbor: The algorithm is an extension of the Nearest Neighbor algorithm. In this case, the tour is computed n times, each one considering a different starting node and then selecting the best tour as the outcome.
  • Insertion Algorithms: All insertion algorithms start with a tour that originated from a random node. In each step, given two nodes already inserted in the tour, the heuristic selects a new node that minimizes the increase in the tour’s length when inserted between such two nodes. Depending on the way such the next node is selected, one can find different variants of the algorithm. For instance, Nearest Insertion, Farthest Insertion, Cheapest Insertion, and Arbitrary Insertion.
  • Concorde: This method is currently one of the best implementations for solving the symmetric TSP. It is based on the Branch-and-Cut method to search for optimal solutions.

3. Related Work on Microaggregation

There is a wide variety of heuristics to solve the multivariate microaggregation problem in the literature. One of the most well-known methods is the Maximum Distance to Average Vector (MDAV), proposed by Domingo-Ferrer et al. [17]. This method iteratively creates clusters of k members considering the furthest records from the dataset centroid. A variant of MDAV was proposed by Laszlo et al., namely the Centroid-Based Fixed Size method (CBFS) [18], which also has optimized versions based on kd-tree neighborhood search, such as KD-CBFS and KD-CBFSapp [19]. The Two Fixed Reference Points (TFRP) method was proposed by Chang et al. [20]. It uses the two most extreme points of the dataset at each iteration as references to create clusters. Differential Privacy-based microaggregation was explored by Yang et al. [21], which created a variant of the MDAV algorithm that uses the correlations between attributes to select the minimum required noise to achieve the desired privacy level. In addition, V-MDAV, a variable group-size heuristic based on the MDAV method was introduced by Solanas et al. in Reference [13] with the aim to relax the cardinality constraints of fixed-size microaggregation and allow clusters to better adapt to the data and reduce the SSE.
Laszlo and Mukherjee [18] approached the microaggregation problem through minimum spanning trees, aimed at creating graph structures that can be pruned according to each node’s associated weights to create the groups. Lin et al. proposed a Density-Based Algorithm (DBA) [22], which first forms groups of records in density descending order, and then fine-tunes these groups in reverse order. The successive Group Selection based on sequential Minimization of SSE (GSMS) method [23], proposed by Panagiotakis et al., optimizes the information loss by discarding the candidate cluster that minimizes the current SSE of the remaining records. Some methods are built upon the HM algorithm. For example, Mortazavi et al. proposed the IMHM method [24]. Domingo-Ferrer et al. [17] proposed a grouping heuristic that combines several methods, such as Nearest Point Next (NPN-MHM), MDAV-MHM, and CBFS-MHM.
Other approaches have focused on the efficiency of the microaggregation procedure, for example, the Fast Data-oriented Microaggregation (FDM) method proposed by Mortazavi et al. [25] efficiently anonymizes large multivariate numerical datasets for multiple successive values of k. The interested readers can find more detailed information about microaggregation in References [5,26].
The most similar work related to ours is the one presented in Reference [27] by Heaton and Mukherjee. The authors use TSP tour optimization heuristics (e.g., 2-opt, 3-opt) to refine a path created with the information of a multivariate microaggregation method (e.g., MDAV, MD, CBFS). Notice that, in our proposed method (described in the next section), we use tour construction TSP heuristics instead of optimization heuristics; thus, we eliminate the need for using a multivariate microaggregation method as a pre-processing step, and we decrease the computational time without hindering data utility.

4. Our Method

Our proposal is built upon two main building blocks: a TSP tour construction heuristic (H), and the optimal univariate microaggregation algorithm of Hansen and Mukherjee (HM). As we have already explained in Section 2, the HM algorithm is applied to univariate numerical data, because it requires the input elements to be in order. However, we virtually use it with multivariate data; thus, when we do that, we refer to it as Multivariate Hansen-Mukherjee (MHM), although, in practice, the algorithm is univariate. Since our proposal is based on a Heuristic (H) to obtain a Hamiltonian Path and the MHM algorithm, we have come to call it HMHM-microaggregation or ( H M ) 2 -Micro, for short.
Given a multivariate microdata set (D) with p columns and r rows, we model it as a complete graph G ( N , E ) , where we assume that each row is represented by a node n i N (or a city, if we think in terms of the TSP), and each edge e i j E represents the Euclidean distance between n i and n j (or the distance between cities in TSP terms). Hence, we have a set of nodes N = { n 1 , n 2 , n r } each representing rows of the microdata set in a multivariate space R p .
In a nutshell, we use H over G to create a Hamiltonian path ( H p a t h ) that travels across all nodes. H p a t h is a permutation ( Π N = { π 1 N , π 2 N , π r N } ) of the nodes in N, and de facto it determines an order for the nodes (i.e., it provides a sense of precedence between nodes). Hence, although D is multivariate, its rows represented as nodes in N can be sorted in a univariant permutation H p a t h that we use as input to the MHM algorithm. As a result, the MHM algorithm returns the optimal univariate k-partition of H p a t h , this is, the set of disjoint subsets S = { s 1 , s 2 , s t } defining the clusters of N. Hence, since each node n i represents a row in D, which is indeed multivariate, we have obtained a multivariate microaggregation of the rows in D and provided a solution for the multivariate microaggregation. Notice that, although MHM returns the optimal k-partition of H p a t h , it does not imply that the resulting microaggregation of D is optimal A schematic of our solution is depicted in Figure 2.
Although the foundation of our proposal described above is pretty straightforward, it has the beauty of putting together complex mathematical building blocks from the multivariate and univariate worlds in a simple yet practical manner. In addition, our solution is very flexible, since it allows the use of any heuristic H to create the Hamiltonian path H p a t h , and it allows for comprehensive studies, such as the one we report in the next section.
Note that most TSP heuristics output a Hamiltonian cycle. However, since we need a Hamiltonian path, we use the well-known solution of adding a dummy node in the graph (i.e., a theoretical node in which its distance to all other nodes is zero), and we cut the cycle by eliminating this node, so as to obtain a Hamiltonian path.
For the sake of completeness, we summarize our proposal step-by-step in Algorithm 1, and we next comment on it. Our solution can be seen as a meta-heuristic to solve the multivariate microaggregation problem, since it can accommodate any Heuristic (H) able to create a Hamiltonian cycle from a complete graph (G), and it could deal with any privacy parameter (k). Thus, our algorithm receives as input a numerical multivariate microdata set D with p columns (attributes) and r rows, that have to be microaggregated, a Heuristic H, and a privacy parameter k (see Algorithm 1: line 1). In order to avoid bias towards higher magnitude variables, the original dataset D (understood as a matrix) is standardized by subtracting to each element the average of its column and dividing it by the standard deviation of the column. The result is a standardized dataset D s t d in which each column has zero mean and unitary standard deviation (see Algorithm 1: line 2). Next, the distance matrix M d i s t is computed. Each element m i j M d i s t contains the Euclidean distance between row i and row j in D s t d ; hence, M d i s t is a square matrix ( r × r ) (see Algorithm 1: line 3). In order to be able to cut the Hamiltonian Cycle and obtain a Hamiltonian path, we add a dummy node to the dataset by adding a zero column and a zero row to M d i s t and generate M d i s t d u m , which is a square matrix ( r + 1 × r + 1 ) (see Algorithm 1: line 4). M d i s t d u m is, in fact, a weighted adjacency matrix that defines a graph G ( N , E ) with nodes N = { n 1 , , n r + 1 } and edges E = { e 11 , e i , j e r + 1 , r + 1 } = { M d i s t d u m 1 , 1 , M d i s t d u m r + 1 , r + 1 } . With this matrix as an input, we could compute a Hamiltonian Cycle H c y c l e on G by applying a TSP heuristic H (see Algorithm 1: line 5). Notice that this Heuristic H could be anyone that gets as input a weighted graph and returns a Hamiltonian cycle. Some examples are: Concorde, Nearest Neighbor, Repetitive Nearest Neighbor, and Insertion Algorithms. After obtaining H c y c l e , we cut it by removing the dummy node (see Algorithm 1: line 6), and we obtain a Hamiltonian path H p a t h that defines a permutation ( Π N = { π 1 N , π 2 N , π r N } ) of the nodes in N, as well as determines an order for the nodes that can be inputted to the MHM algorithm to obtain its optimal k-partition (S) (see Algorithm 1: line 7). S is a set of disjoint subsets S = { s 1 , s 2 , s t } defining the clusters of nodes in N. Hence, with S and D, we could create a microaggregated dataset D by replacing each row in D by the average vector of the k-partition subset to which it belongs (see Algorithm 1: line 8).
After applying the algorithm, we have transformed the original dataset D into a dataset D that has been microaggregated so as to guarantee the privacy criteria established by k.
Algorithm 1 ( H M ) 2 -Micro
1:
  function ( H M ) 2 -MICRO ( Microdata set D, TSP-Heuristic H, Privacy Parameter k)
2:
         D s t d = StandardizeDataset(D)
3:
         M d i s t = ComputeDistanceMatrix( D s t d )
4:
         M d i s t d u m = InsertDummyNode( M d i s t )
5:
         H c y c l e = CreateHamiltonianCycle( M d i s t d u m , H)
6:
         H p a t h = CutDummyNode( H c y c l e )
7:
        S = MHM( H p a t h , D s t d , k)
8:
         D = BuildMicroaggregatedDataSet(D, S);
9:
  return D
10:
end function

5. Experiments

With the aim to practically validate the usefulness of our multivariate microaggregation proposal, we have thoroughly tested it on six datasets (described in Section 5.1) that serve as benchmarks. In addition, we are interested in knowing (if and) to what extend our method outperforms the best performing microaggregation methods in the literature. Hence, we have compared our proposal with these methods (described in Section 5.2), and the results of all these tests are summarized in Section 5.3. Overall, considering four different values for the privacy parameter k { 3 , 4 , 5 , 6 } , ten microaggregation algorithms, 50 repetitions per case, and six datasets, we have run over 12.000 microaggregation tests, which allow us to provide a statistically solid set of results.

5.1. Datasets

We used six datasets as benchmarks for our experiments. We can classify those datasets into two main groups: The first group comprises three well-known SDC microdata sets that have been used for years as benchmarks in the literature, namely “Census”, “EIA”, and “Tarragona”. The second group comprises three mobility datasets containing real GPS traces from three Spanish cities, namely “Barcelona”, “Madrid”, and “Tarraco”. Notice that we use the term “Tarraco”, the old Roman name for the city of Tarragona, in order to avoid confusion with the classic benchmark dataset “Tarragona”. The features of each dataset are next summarized:
The Census dataset was obtained using the public Data Extraction System of the U.S. Census Bureau. It contains 1080 records with 13 numerical attributes. The Tarragona dataset was obtained from the Tarragona Chamber of Commerce. It contains information on 834 companies in the Tarragona area with 13 variables per record. The EIA dataset was obtained from the U.S. Energy Information Authority, and it consists of 4092 records with 15 attributes. More details on the aforementioned datasets can be obtained in Reference [28].
The Barcelona, Madrid, and Tarraco datasets consist of OpenStreetMap [29] GPS traces collected from those cities: Barcelona contains the GPS traces of the city of Barcelona within the area determined by the parallelogram formed by latitude (41.3726866, 41.4078446) and longitude (2.1268845, 2.1903992). The dataset has 969 records with 30 GPS locations each. Madrid contains the GPS traces of the city of Madrid within the area determined by the parallelogram formed by latitude (40.387613, 40.483515) and longitude (−3.7398145, −3.653985). The dataset has 959 records with 30 GPS locations each. Tarraco contains the GPS traces of the city of Tarragona within the area determined by the parallelogram formed by latitude (41.0967083, 41.141174) and longitude (1.226008, 1.2946691). The dataset has 932 records with 30 GPS locations each.
In all trajectories datasets, each record consists of 30 locations represented as (latitude and longitude). Hence, each record has 60 numerical values. These locations were extracted from each corresponding parallelogram according to the amount of recorded tracks and their length.
All datasets are available in our website: https://www.smarttechresearch.com/publications/symmetry2021-Maya-Casino-Solanas/ (accessed on 1 May 2021).

5.2. Compared Methods

We have selected a representative set of well-known and state-of-the-art methods to assess the value of our approach. We have selected two classic microaggregation methods (i.e., MDAV and V-MDAV), as baselines. In the case of V-MDAV, the method was run for several values of γ { 0 , 2 } , and the best result is reported. Although some other newer methods might have achieved better results, they are still landmarks that deserve to be included in any microaggregation comparison.
For newer and more sophisticated methods, we have considered the work of Heaton and Mukherjee [27], in which they study a variety of microaggregation heuristics, including methods, such as CBFS and MD. Thus, instead of comparing our proposal with all those methods, we have taken the method that Heaton and Mukherjee reported as the best performer, namely the MDAV-LK-MHM method. This method, which is based on MDAV, first creates a microaggregation using MDAV, next improves the result of MDAV by applying the LK heuristic, and it finally applies MHM to obtain the resulting microaggregation (cf. Reference [27] for further details on the algorithm).
Regarding our proposal (i.e., ( H M ) 2 -Micro), as we already discussed, it can be understood as a meta-heuristic able to embody any heuristic H that returns a Hamiltonian Cycle. Hence, with the aim to determine the best heuristic, we have analyzed seven alternatives, namely Nearest Neighbor, Repetitive Nearest Neighbor, Nearest Insertion, Farther Insertion, Cheapest Insertion, Arbitrary Insertion, and (our suggestion) Concorde. Table 1 summarizes some features of all selected methods, including the reference to the original article where the method was described. For our method, each reference points to the article describing the TSP heuristic.
The implementation of all these methods have used the R package sdcMicro [28], the TSP heuristics implemented in Reference [32], and the LK heuristics implemented in Reference [33]. LK has been configured so that the algorithm runs once at each iteration parameter RUN=1 until a local optimum is reached. This same criteria was followed for the other TSP heuristics. In this regard, the heuristics we used consider a random starting node at each run. Hence, each experiment has been repeated 50 times to guarantee statistically sound outcomes regardless of this random starting point.

5.3. Results Overview

By using the datasets and methods described above, we have analyzed the Information Loss (expressed in percentage), as a measure of data utility (cf. Section 2 for details). It is assumed that, given a privacy parameter k that guarantees that the microaggregated dataset is k-anonymous, the lower the Information Loss the better the result and performance of the microaggregation method. The results are reported in Table 2, Table 3, Table 4, Table 5, Table 6 and Table 7 with the best (lowest) information loss highlighted in green.
Overall, it can be observed that our method, ( H M ) 2 -Micro, with the Concorde heuristic is the best performer in 79% of the experiments, and it is the second best in the remaining 21% (for which the MDAV-LK-MHM outperforms it by a narrow margin of less than 2%). Interestingly enough, although ( H M ) 2 -Micro, with both Nearest Insertion and Farthest-Insertion, is not the best performer in any experiment, it outperforms MDAV-LK-HMH 50% of the times. The rest of the methods obtain less consistent results and highly depend on the dataset.
When we analyze the results more closely for each particular dataset, we observe that, in the case of the “Census” dataset (cf. Table 2), our method with Concorde outperforms all methods for all values of k. In addition, despite the random nature of TSP-heuristics, the values of σ are very stable, denoting the robustness of all methods, yet slightly higher on average in the case of the methods with higher Information Loss. It is worth emphasizing though, that, in all runs, our method with Concorde and the MDAV-LK-MHM method obtained better results than MDAV and V-MDAV (i.e., the max values obtained in all runs are lower than the outcomes obtained by MDAV and V-MDAV).
For the “EIA” dataset (cf. Table 3), MDAV-LK-MHM is the best performer for all values of k except k = 5 , for which our proposal with Concorde performs better. In this case, the results obtained by these two methods are very close. Similarly to the results in “Census”, the max values obtained by these two methods outperform MDAV and V-MDAV. In the case of “Tarragona” (cf. Table 4), our method with Concorde outperforms all other methods. Surprisingly, both MDAV and V-MDAV obtain better results than MDAV-LK-MHM, which performs poorly in this dataset.
So, it can be concluded that the overall winner for the classical benchmarks (i.e., Census, EIA, and Tarragona) is our method, ( H M ) 2 -Micro, with the Concorde heuristic, that is only marginally outperformed by MDAV-LK-MHM in the EIA dataset.
Regarding the other three datasets containing GPS traces (i.e., Barcelona, Madrid and Tarraco), our method, ( H M ) 2 -Micro, with the Concorde heuristic, is the best performer in 83% of the cases and comes second best in the remaining 17%. For the Barcelona dataset (cf. Table 5), MDAV-LK-MHM and ( H M ) 2 -Micro, with the Concorde heuristic, perform very well and similarly. The methods with the worst Information Loss are MDAV and V-MDAV. Our method, ( H M ) 2 -Micro, with the Insertion heuristics, have a remarkable performance, obtaining values similar to those of MDAV-LK-MHM and Concorde. Nevertheless, it is worth noting that the max (worst) values obtained by MDAV-LK-MHM and Concorde are still better than the averages obtained by the other methods. In the case of the Madrid dataset (cf. Table 6), our method, ( H M ) 2 -Micro, with the Concorde heuristic, achieves the minimum (best) value of Information Loss for all values of k. We can also observe that our method with Insertion heuristics offers higher performance than MDAV-LK-MHM. Finally, the results for the Tarraco dataset (cf. Table 7) show that the minimum (best) Information Loss value is obtained by our method with the Concorde heuristic in all cases. In this case, MDAV-LK-MHM performs poorly, and, for k = 3 and k = 4 , MDAV and V-MDAV are better.
We have already discussed that all studied methods (with the exception of MDAV and V-MDAV) have a non-deterministic component emerging from the random selection of the initial node. This random selection affects the performance of the final microaggregation obtained. With the aim to analyze the effect of this non-deterministic behavior, we have studied the standard deviation of all methods for all values of k and for all datasets. In addition, we have visually inspected the variability of the results by using box plot diagrams.
Since the results are quite similar and consistent across all datasets, for the sake of clarity, we only reproduce here the box plots for the “Census” dataset (see Figure 3), and we leave the others in Appendix A for the interested reader.
In Figure 3, we can observe that the Information Loss values increase with k, but all methods have the same behavior regardless of the value of k. In addition, it is clear that the most stable methods are ( H M ) 2 -Micro, with Concorde, and MDAV-LK-MHM.
Overall, we observe some expected differences depending on the datasets. However, the behavior of the best performing methods is stable. Particularly, the datasets with GPS traces (i.e., Barcelona, Madrid, and Tarraco) show more stable results. In summary, the best method was our ( H M ) 2 -Micro with Concorde, exhibiting the most stable results across all datasets.

6. Discussion

Over the previous sections, we have presented our microaggregation method, ( H M ) 2 -Micro, its rationale, and its performance against other classic and state-of-the-art methods on a variety of datasets. In the previous section, we have reported the main results, and we will discuss them next by progressively answering the research questions that we posed in the Introduction of the article.
Q1:How to create a suitable ordering for a univariate microaggregation algorithm, when the records are in R p .
A main takeaway of this article is that, by using a combination of TSP tour construction heuristics (e.g., Concorde) and an optimal univariate microaggregation algorithm, we are properly ordering multivariate datasets in a univariate fashion that leads to excellent multivariate microaggregation solutions. Other approaches to order R p points might consider projecting them over the principal component. However, the information loss associated with this approach makes it unsuitable. In addition, other more promising approaches, like the one used in MDAV-LK-MHM, first create a k-partition and set an order based on maximum distance criteria. Although this approach might work well in some cases, we have clearly seen that Hamiltonian paths created by TSP-heuristics, like Concorde, outperform this approach. Hence, based on the experiments of Section 5, we can conclude that TSP-heuristics, like Concorde, provide an order for elements in R p that is suitable for an optimal univariate microaggregation algorithm to output a consistent multivariate microaggregation solution with low Information Loss (i.e., high data utility). Moreover, from all analyzed heuristics, it is clear that the best performer is Concorde, followed by insertion heuristics.
Q2:Are the length of the Hamiltonian path and the information loss of the microaggregation related?, or Do shorter Hamiltonian paths lead to microaggregation solutions with lower information loss?
When we started this research, our intuition was that good heuristic solutions of the TSP (i.e., those with shorter path lengths) would provide a Hamiltonian path, that could be used as an ordered vector for the HM optimal univariate microaggregation algorithm, resulting in a good multivariate microaggregation solution. From this intuition, we assumed that shorter Hamiltonian paths would lead to lower Information Loss in microaggregated datasets.
In order to validate (or disproof) this intuition, we have analyzed the Pearson correlation between the Hamiltonian path length obtained by all studied heuristics (i.e., Nearest Neighbor, Repetitive Nearest Neighbor, Nearest Insertion, Farther Insertion, Cheapest Insertion, Arbitrary Insertion, and Concorde) and the SSE of the resulting microaggregation. We have done so for all studied datasets and k values. The results are summarized in Table 8, and all plots along with a trend line are available in Appendix B.
From the correlation analysis, it can be concluded that there is a positive correlation between the Hamiltonian path length and the SSE. This is, the shorter the path length the lower the SSE. This statement holds for all k and for all datasets (although Census exhibits a lower correlation). Hence, although this result is not a causality proof, it can be safely said that good solutions of the TSP problem lead to good solutions of the multivariate microaggregation problem. In fact, the best heuristic (i.e., Concorde) always results in the lowest (best) SSE.
Interested readers can find all plots in Appendix B. However, for the sake of clarity, let us illustrate this result by discussing the case of the Madrid dataset with k = 6 , depicted in Figure 4. In the figure, the positive correlation is apparent. In addition, it is clear that heuristics tend to form clusters. In a nutshell, the best heuristic is Concorde, followed by the insertion family of methods (i.e., Nearest Insertion, Furthest Insertion, Cheapest Insertion, and Arbitrary Insertion), followed by Repetitive Nearest Neighbor and Nearest Neighbor.
Although Figure 4 clearly illustrates the positive correlation between the path length and the SSE, it also shows that heuristics tend to cluster and might indicate that not only the path but the heuristic (per se) plays a role in the reduction of the SSE. This indication leads us to our next research question.
Q3:Is the length of the Hamiltonian path the only factor affecting information loss or does the particular construction of the path (regardless of the length) affect the information loss?
In the previous question, we have found clear positive correlation between the path length and the SSE. However, we have also observed apparent clusters suggesting that the very heuristics could be responsible for the minimization of the SSE. In other words, although the path length and SSE are positively correlated when all methods are analyzed together, would this correlation hold when heuristics are analyzed one at a time? In order to answer this question, we have analyzed the results of each heuristic individually, and we have observed that there is still positive correlation between path length and SSE, but it is very weak or almost non-existent (i.e., very close to 0), as Figure 5 illustrates.
The results shown in Figure 5 are only illustrative, and a deeper analysis that is out of the scope of this paper would be necessary. However, our initial results indicate that, although there is positive correlation between path length and SSE globally, this correlation weakens significantly when analyzed on each heuristic individually. This result suggests that it is not only the length of the path but the way in which this path is constructed what affects the SSE. This would explain why similar methods (e.g., those based on insertion) behave similarly in terms of SSE although their paths’ length varies.
Q4:Does ( H M ) 2 -Micro provide better solutions (in terms of information loss) than the best performing microaggregation methods in the literature?
This question has been already answered in Section 5.3. However, for the sake of completeness, we summarize it here: The results obtained after executing more than 12,000 tests suggest that our solution ( H M ) 2 -Micro obtains better results than classic microaggregation methods, such as MDAV and V-MDAV. Moreover, when ( H M ) 2 -Micro uses the Concorde heuristic to determine the Hamiltonian path, it outperforms the best state-of-the-art methods consistently. In our experiments, ( H M ) 2 -Micro with Concorde was the best performer 79% of the times and was the second best in the remaining 21%.
Q5:Do TSP-based microaggregation methods perform better than current solutions on trajectories datasets?
( H M ) 2 -Micro with Concorde is the best overall performer. Moreover, if we focus on those datasets with trajectory data (i.e., Barcelona, Madrid, and Tarraco), the results are even better. It is the best performer in 83% of the tests and the second best in the remaining 17%. This good behavior of the method could result from the very foundations of the TSP; however, there is still plenty of research to do in this line to reach more solid conclusions. Location privacy is a very complex topic that encompasses many nuances beyond k-anonymity models (such as the one followed in this article). However, this result is an invigorating first step towards the analysis of novel microaggregation methods applied to trajectory analysis and protection.

7. Conclusions

Microaggregation has been studied for decades now, and, although finding the optimal microaggregation is NP-Hard and a polynomial-time microaggregation algorithm has not been found, steady improvements over microaggregation heuristics have been made. Hence, after such a long research and polishing process, finding new solutions that improve the best methods is increasingly difficult. In this article, we have presented ( H M ) 2 -Micro, a meta-heuristic that leverages the advances in TSP solvers and combines them with the optimal univariate microaggregation to create a flexible and robust multivariate microaggregation solution.
We have studied our method and thoroughly compared it to classic and state-of-the-art microaggregation algorithms over a variety of classic benchmarks and trajectories datasets. Overall, we have executed more than 12,000 tests, and we have shown that our solution embodying the Concorde heuristic outperforms the others. Hence, we have shown that our TSP-inspired method could be used to guarantee k-anonymity of trajectories datasets whilst reducing the Information Loss, thus increasing data utility. Furthermore, our proposal is very stable, i.e., it does not change significantly its performance regardless of the random behavior associated with initial nodes selection.
In addition to proposing ( H M ) 2 -Micro, we have found clear correlations between the length of Hamiltonian Paths and the SSE introduced by microaggregation processes, and we have shown the importance of the Hamiltonian Cycle construction algorithms over the overall performance of microaggregation.
Despite these relevant results, there is still much to do in the study of microaggregation and data protection. Future work will focus on scaling up ( H M ) 2 -Micro to high-dimensional and very-large datasets. Considering the growing importance of Big Data and Cloud Computing, adapting our solution to distributed computation environments is paramount. Moreover, adjusting TSP heuristics to leverage lightweight microaggregation-based approaches is an interesting research path to follow. In addition, although the values of the privacy parameter k are typically low (i.e., 3, 4, 5, 6), we plan to study the effect of larger values of k on our solution. Last but not least, since microaggregation is essentially a data-oriented procedure, we will study how our solution adapts to data structures from specific domains, such as healthcare, transportation, energy, and the like.
All in all, with ( H M ) 2 -Micro, we have set the ground for the study of multivariate microaggregation meta-heuristics from a new perspective, that might continue in the years to come.

Author Contributions

Conceptualization, A.M.-L. and A.S.; methodology, A.S.; software, A.M.-L.; validation, A.M.-L., F.C. and A.S.; formal analysis, A.S.; investigation, A.M.-L., F.C. and A.S.; resources, A.S.; data curation, A.M.-L.; writing—original draft preparation, A.M.-L. and A.S.; writing—review and editing, F.C. and A.S.; visualization, F.C.; supervision, F.C. and A.S.; project administration, A.S.; funding acquisition, A.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the European Commission under the Horizon 2020 Programme (H2020), as part of the project LOCARD (https://locard.eu (accessed on 1 May 2021)) (Grant Agreement no. 832735), and by the Spanish Ministry of Science & Technology with project IoTrain RTI2018-095499-B-C32. The content of this article does not reflect the official opinion of the European Union. Responsibility for the information and views expressed therein lies entirely with the authors.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Dataset and high resolution data can be found in: https://www.smarttechresearch.com/publications/symmetry2021-Maya-Casino-Solanas/ (accessed on 1 May 2021).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Information Loss Variability Box Plots

Figure A1. Information Loss variability for each value of k over the Census dataset.
Figure A1. Information Loss variability for each value of k over the Census dataset.
Symmetry 13 00916 g0a1
Figure A2. Information Loss variability for each value of k over the EIA dataset.
Figure A2. Information Loss variability for each value of k over the EIA dataset.
Symmetry 13 00916 g0a2
Figure A3. Information Loss variability for each value of k over the Tarragona dataset.
Figure A3. Information Loss variability for each value of k over the Tarragona dataset.
Symmetry 13 00916 g0a3
Figure A4. Information Loss variability for each value of k over the Barcelona dataset.
Figure A4. Information Loss variability for each value of k over the Barcelona dataset.
Symmetry 13 00916 g0a4
Figure A5. Information Loss variability for each value of k over the Madrid dataset.
Figure A5. Information Loss variability for each value of k over the Madrid dataset.
Symmetry 13 00916 g0a5
Figure A6. Information Loss variability for each value of k over the Tarraco dataset.
Figure A6. Information Loss variability for each value of k over the Tarraco dataset.
Symmetry 13 00916 g0a6

Appendix B. Correlation Analysis between “Path Length” and “SSE”

Figure A7. Relation between SSE and Path Length for Census and k { 3 , 4 , 5 , 6 } .
Figure A7. Relation between SSE and Path Length for Census and k { 3 , 4 , 5 , 6 } .
Symmetry 13 00916 g0a7
Figure A8. Relation between SSE and Path Length for EIA and k { 3 , 4 , 5 , 6 } .
Figure A8. Relation between SSE and Path Length for EIA and k { 3 , 4 , 5 , 6 } .
Symmetry 13 00916 g0a8
Figure A9. Relation between SSE and Path Length for Tarragona and k { 3 , 4 , 5 , 6 } .
Figure A9. Relation between SSE and Path Length for Tarragona and k { 3 , 4 , 5 , 6 } .
Symmetry 13 00916 g0a9
Figure A10. Relation between SSE and Path Length for Barcelona and k { 3 , 4 , 5 , 6 } .
Figure A10. Relation between SSE and Path Length for Barcelona and k { 3 , 4 , 5 , 6 } .
Symmetry 13 00916 g0a10
Figure A11. Relation between SSE and Path Length for Madrid and k { 3 , 4 , 5 , 6 } .
Figure A11. Relation between SSE and Path Length for Madrid and k { 3 , 4 , 5 , 6 } .
Symmetry 13 00916 g0a11
Figure A12. Relation between SSE and Path Length for Tarraco and k { 3 , 4 , 5 , 6 } .
Figure A12. Relation between SSE and Path Length for Tarraco and k { 3 , 4 , 5 , 6 } .
Symmetry 13 00916 g0a12

References

  1. Ye, H.; Cheng, X.; Yuan, M.; Xu, L.; Gao, J.; Cheng, C. A survey of security and privacy in big data. In Proceedings of the 2016 16th International Symposium on Communications and Information Technologies (ISCIT), Qingdao, China, 26–28 September 2016; pp. 268–272. [Google Scholar]
  2. Vatsalan, D.; Sehili, Z.; Christen, P.; Rahm, E. Privacy-Preserving Record Linkage for Big Data: Current Approaches and Research Challenges. In Handbook of Big Data Technologies; Springer International Publishing: Cham, Switzerlands, 2017; pp. 851–895. [Google Scholar]
  3. Mehmood, A.; Natgunanathan, I.; Xiang, Y.; Hua, G.; Guo, S. Protection of big data privacy. IEEE Access 2016, 4, 1821–1834. [Google Scholar] [CrossRef] [Green Version]
  4. Barbaro, M.; Zeller, T.; Hansell, S. A face is exposed for AOL searcher no. 4417749. New York Times, 9 August 2006. [Google Scholar]
  5. Zigomitros, A.; Casino, F.; Solanas, A.; Patsakis, C. A Survey on Privacy Properties for Data Publishing of Relational Data. IEEE Access 2020, 8, 51071–51099. [Google Scholar] [CrossRef]
  6. Domingo-Ferrer, J.; Mateo-Sanz, J.M. Practical data-oriented microaggregation for statistical disclosure control. Knowl. Data Eng. IEEE Trans. 2002, 14, 189–201. [Google Scholar] [CrossRef] [Green Version]
  7. Torra, V. Microaggregation for Categorical Variables: A Median Based Approach. In Privacy in Statistical Databases; Springer: Berlin/Heidelberg, Germany, 2004; pp. 162–174. [Google Scholar]
  8. Solanas, A.; Martinez-Balleste, A.; Mateo-Sanz, J.M. Distributed architecture with double-phase microaggregation for the private sharing of biomedical data in mobile health. IEEE Trans. Inf. Forensics Secur. 2013, 8, 901–910. [Google Scholar] [CrossRef]
  9. Martinez-Balleste, A.; Perez-Martines, P.A.; Solanas, A. The pursuit of citizens’ privacy: A privacy-aware smart city is possible. IEEE Commun. Mag. 2013, 51, 136–141. [Google Scholar] [CrossRef]
  10. Casino, F.; Domingo-Ferrer, J.; Patsakis, C.; Puig, D.; Solanas, A. A k-anonymous approach to privacy preserving collaborative filtering. J. Comput. Syst. Sci. 2015, 81, 1000–1011. [Google Scholar] [CrossRef]
  11. Domingo-Ferrer, J.; Sebé, F.; Solanas, A. A polynomial-time approximation to optimal multivariate microaggregation. Comput. Math. Appl. 2008, 55, 714–732. [Google Scholar] [CrossRef] [Green Version]
  12. Hansen, S.L.; Mukherjee, S. A polynomial algorithm for optimal univariate microaggregation. IEEE Trans. Knowl. Data Eng. 2003, 15, 1043–1044. [Google Scholar] [CrossRef]
  13. Solanas, A.; Martinez, A. VMDAV: A Multivariate Microaggregation With Variable Group Size. In Proceedings of the 17th COMPSTAT Symposium of the IASC, Rome, Italy, 28 August–1 September 2006; pp. 917–925. [Google Scholar]
  14. Shmoys, D.B.; Lenstra, J.; Kan, A.R.; Lawler, E.L. The Traveling Salesman Problem; John Wiley & Sons, Incorporated: Hoboken, NJ, USA, 1985; Volume 12. [Google Scholar]
  15. Rosenkrantz, D.J.; Stearns, R.E.; Lewis, P.M., II. An analysis of several heuristics for the traveling salesman problem. SIAM J. Comput. 1977, 6, 563–581. [Google Scholar] [CrossRef]
  16. Applegate, D.L.; Bixby, R.E.; Chvatal, V.; Cook, W.J. The Traveling Salesman Problem: A Computational Study; Princeton University Press: Princeton, NJ, USA, 2006. [Google Scholar]
  17. Domingo-Ferrer, J.; Martinez-Ballesté, A.; Mateo-Sanz, J.; Sebé, F. Efficient multivariate data-oriented microaggregation. VLDB J. 2006, 15, 355–369. [Google Scholar] [CrossRef]
  18. Laszlo, M.; Mukherjee, S. Minimum spanning tree partitioning algorithm for microaggregation. IEEE Trans. Knowl. Data Eng. 2005, 17, 902–911. [Google Scholar] [CrossRef]
  19. Solé, M.; Muntés-Mulero, V.; Nin, J. Efficient microaggregation techniques for large numerical data volumes. Int. J. Inf. Secur. 2012, 11, 253–267. [Google Scholar] [CrossRef]
  20. Chang, C.; Li, Y.; Huang, W. TFRP: An efficient microaggregation algorithm for statistical disclosure control. J. Syst. Softw. 2007, 80, 1866–1878. [Google Scholar] [CrossRef]
  21. Yang, G.; Ye, X.; Fang, X.; Wu, R.; Wang, L. Associated attribute-aware differentially private data publishing via microaggregation. IEEE Access 2020, 8, 79158–79168. [Google Scholar] [CrossRef]
  22. Lin, J.L.; Wen, T.H.; Hsieh, J.C.; Chang, P.C. Density-based microaggregation for statistical disclosure control. Expert Syst. Appl. 2010, 37, 3256–3263. [Google Scholar] [CrossRef]
  23. Panagiotakis, C.; Tziritas, G. Successive group selection for microaggregation. IEEE Trans. Knowl. Data Eng. 2011, 25, 1191–1195. [Google Scholar] [CrossRef]
  24. Mortazavi, R.; Jalili, S.; Gohargazi, H. Multivariate microaggregation by iterative optimization. Appl. Intell. 2013, 39, 529–544. [Google Scholar] [CrossRef]
  25. Mortazavi, R.; Jalili, S. Fast data-oriented microaggregation algorithm for large numerical datasets. Knowl. Based Syst. 2014, 67, 195–205. [Google Scholar] [CrossRef]
  26. Fayyoumi, E.; Oommen, B.J. A survey on statistical disclosure control and micro-aggregation techniques for secure statistical databases. Softw. Pract. Exp. 2010, 40, 1161–1188. [Google Scholar] [CrossRef]
  27. Heaton, B.; Mukherjee, S. Record Ordering Heuristics for Disclosure Control through Microaggregation. In Proceedings of the International Conference on Advances in Communication and Information Technology, Amsterdam, The Netherlands, 1–2 December 2011. [Google Scholar]
  28. Templ, M. Statistical disclosure control for microdata using the R-package sdcMicro. Trans. Data Priv. 2008, 1, 67–85. [Google Scholar]
  29. OpenStreetMap Contributors. 2017. Available online: https://planet.osm.org (accessed on 1 May 2021).
  30. Nilsson, C. Heuristics for the Traveling Salesman Problem; Technical Report; Linköping University: Linköping, Sweden, 2003; Available online: http://www.ida.liu.se/~TDDB19/reports_2003/htsp.pdf (accessed on 1 May 2021).
  31. Morrison, D.R.; Jacobson, S.H.; Sauppe, J.J.; Sewell, E.C. Branch-and-bound algorithms: A survey of recent advances in searching, branching, and pruning. Discret. Optim. 2016, 19, 79–102. [Google Scholar] [CrossRef]
  32. Hahsler, M.; Hornik, K. TSP-Infrastructure for the Traveling Salesperson Problem. J. Stat. Softw. 2007, 23, 1–21. [Google Scholar] [CrossRef]
  33. Helsgaun, K. An effective implementation of the Lin–Kernighan traveling salesman heuristic. Eur. J. Oper. Res. 2000, 126, 106–130. [Google Scholar] [CrossRef] [Green Version]
Figure 1. A Hamiltonian path for the Eurodist dataset.
Figure 1. A Hamiltonian path for the Eurodist dataset.
Symmetry 13 00916 g001
Figure 2. Given a microdata dataset, we use a tour construction heuristic to generate a Hamiltonian path, which will be used as the input of the MHM method to generate the groups.
Figure 2. Given a microdata dataset, we use a tour construction heuristic to generate a Hamiltonian path, which will be used as the input of the MHM method to generate the groups.
Symmetry 13 00916 g002
Figure 3. Information Loss variability for each value of k over the Census dataset.
Figure 3. Information Loss variability for each value of k over the Census dataset.
Symmetry 13 00916 g003
Figure 4. Relation between SSE and Path Length for Madrid and k = 6 .
Figure 4. Relation between SSE and Path Length for Madrid and k = 6 .
Symmetry 13 00916 g004
Figure 5. Correlation between path length and SSE for each individual method (from top to bottom: Cheapest Insertion, Concorde, and Nearest Neighbor) for k = 3 over the Madrid dataset.
Figure 5. Correlation between path length and SSE for each individual method (from top to bottom: Cheapest Insertion, Concorde, and Nearest Neighbor) for k = 3 over the Madrid dataset.
Symmetry 13 00916 g005
Table 1. Comparing methods and features. For Concorde, M is a bound on the time to explore subproblems, b is a branching factor, and d is a search depth.
Table 1. Comparing methods and features. For Concorde, M is a bound on the time to explore subproblems, b is a branching factor, and d is a search depth.
MethodCardinalityComputational CostReference
MDAV fixed O ( n 2 / 2 k ) [17]
V-MDAV variable O ( n 2 ) [13]
MDAV-LK-MHM variable O ( n 2 / 2 k ) [27]
( H M ) 2 -MicroTSP Heuristic + MHM
Nearest Neighborvariable O ( n 2 ) [15]
Repetitive Nearest-Neighborvariable O ( n 2 log n ) [15]
Nearest Insertionvariable O ( n 2 ) [30]
Farthest Insertionvariable O ( n 2 ) [30]
Cheapest Insertionvariable O ( n 2 ) [30]
Arbitrary Insertionvariable O ( n 2 ) [30]
Concordevariable O ( M b d ) [31]
Table 2. Information Loss obtained on the Census dataset.
Table 2. Information Loss obtained on the Census dataset.
Census
k = 3 k = 4 k = 5 k = 6
MethodAverage σ minmaxAverage σ minmaxAverage σ minmaxAverage σ minmax
MDAV5.6922NANANA7.4947NANANA9.0884NANANA10.3847NANANA
V-MDAV5.6619NANANA7.4947NANANA9.0070NANANA10.2666NANANA
MDAV-LK-MHM5.10850.03985.02565.18776.91310.05266.77747.02278.51990.08428.31008.70309.97520.12849.767510.2527
Nearest Insertion-MHM5.65610.13695.35966.06957.48180.15797.19467.93188.96170.25398.51909.472710.30050.29279.762411.2086
Farthest Insertion-MHM5.56380.09565.33005.89957.34850.09907.17237.58538.82340.13228.57849.174810.12500.19329.697010.7363
Cheapest Insertion-MHM5.70440.07195.56695.87667.46250.11557.26747.80529.03400.12368.72129.384710.37870.130510.170610.9089
Arbitrary Insertion-MHM5.58830.09765.42355.87637.37230.14387.12727.82508.86960.17888.50729.286710.20110.24759.708110.7794
Nearest Neighbor-MHM6.97180.35086.19787.72919.24330.37028.674410.224611.32870.385410.523012.395813.13570.405312.471113.9421
Repetitive NN-MHM6.28880.21925.88116.68418.67790.27997.99419.334510.75180.247210.342111.455412.58820.314311.936013.2915
Concorde-MHM5.05630.03774.99175.11696.88460.05556.78957.02178.45760.09038.23728.66149.84400.12329.554210.2517
Table 3. Information Loss obtained on the EIA dataset.
Table 3. Information Loss obtained on the EIA dataset.
EIA
k = 3 k = 4 k = 5 k = 6
MethodAverage σ minmaxAverage σ minmaxAverage σ minmaxAverage σ minmax
MDAV0.4829NANANA0.6713NANANA1.6667NANANA1.3078NANANA
V-MDAV0.4829NANANA0.6713NANANA1.2771NANANA1.2320NANANA
MDAV-LK-MHM0.37410.00750.36590.40970.52510.01160.51170.56930.78900.03360.75020.89321.04300.02891.00331.1113
Nearest Insertion-MHM0.40610.01140.38310.42380.57810.02410.54410.61790.86210.04560.80320.97601.12540.08370.99761.3334
Farthest Insertion-MHM0.40700.01190.38720.42070.58780.02510.55240.62770.87640.05220.81900.97471.17760.03591.12451.2484
Cheapest Insertion-MHM0.52540.03580.46920.56510.73210.06410.63220.84771.08680.06890.99101.22641.40610.11471.26051.6329
Arbitrary Insertion-MHM0.42810.03000.39210.49440.60920.03760.55660.66990.90480.08400.81941.06211.19280.10771.06521.3476
Nearest Neighbor-MHM0.90280.14550.50891.10231.15100.16750.70561.37761.40150.17880.94511.67671.67920.11071.46351.9139
Repetitive NN-MHM0.51100.05320.47250.65990.71920.05570.66460.86191.00720.07010.92741.11261.31010.15211.15611.4825
Concorde-MHM0.38890.02030.36730.42100.52880.01700.50870.55760.78020.02670.75810.85011.04760.02821.00091.0904
Table 4. Information Loss obtained on the Tarragona dataset.
Table 4. Information Loss obtained on the Tarragona dataset.
Tarragona
k = 3 k = 4 k = 5 k = 6
MethodAverage σ minmaxAverage σ minmaxAverage σ minmaxAverage σ minmax
MDAV16.9326NANANA19.5460NANANA22.4619NANANA26.3252NANANA
V-MDAV16.6603NANANA19.5460NANANA22.4619NANANA26.3252NANANA
MDAV-LK-MHM18.79691.873815.059523.083022.85231.757619.119526.280626.24321.506623.042128.952228.52441.774225.170330.9656
Nearest Insertion-MHM15.96870.836015.110720.183519.36771.314117.803224.528623.73231.437621.836528.975326.90181.567424.653833.0785
Farthest Insertion-MHM15.76340.206215.474316.662319.03230.552118.106220.210522.83160.763621.331324.198825.76270.449624.900426.9613
Cheapest Insertion-MHM16.31421.486115.216922.027119.77841.606018.310325.891623.90171.715522.312130.082827.55721.661125.239432.7082
Arbitrary Insertion-MHM16.09180.752715.131018.966819.54611.343618.207225.857223.76851.398521.733329.186327.04191.687225.009333.2382
Nearest Neighbor-MHM22.30190.886619.962023.549627.10021.223424.252729.511730.44781.545527.702633.351334.54451.208831.330237.5350
Repetitive NN-MHM17.69811.215715.743520.998122.12321.913820.083928.739927.90891.794625.143432.572930.40851.921628.064835.2458
Concorde-MHM14.76770.085814.629414.963317.99570.124117.752818.221121.98950.216421.671222.347925.34590.206124.804525.6564
Table 5. Information Loss obtained on the Barcelona dataset.
Table 5. Information Loss obtained on the Barcelona dataset.
Barcelona
k = 3 k = 4 k = 5 k = 6
MethodAverage σ minmaxAverage σ minmaxAverage σ minmaxAverage σ minmax
MDAV2.5667NANANA3.5023NANANA4.2849NANANA5.1873NANANA
V-MDAV2.5667NANANA3.3193NANANA4.2849NANANA5.1873NANANA
MDAV-LK-MHM1.62510.03621.56371.74252.19130.03392.11702.27382.67980.06072.51562.80673.21200.06643.07313.3825
Nearest Insertion-MHM1.80220.06561.68571.94382.35260.08422.17542.50502.84050.10082.64173.04113.33160.10833.10933.5103
Farthest Insertion-MHM1.78380.05251.69671.89802.35750.06982.19192.46812.83860.07512.66542.96703.31890.11313.11123.6445
Cheapest Insertion-MHM1.81560.05651.68871.92932.38800.09122.23542.54732.88870.07922.78073.04053.41180.12383.19383.6247
Arbitrary Insertion-MHM1.80610.06351.68231.94692.35930.07492.18082.54142.82310.09112.63383.02513.33310.10853.10313.5584
Nearest Neighbor-MHM2.20190.12021.91652.44762.92740.17782.52763.33773.47330.21683.06113.93994.10530.25903.51594.6420
Repetitive NN-MHM2.00910.05631.88992.25472.74740.06112.61083.01303.23180.10013.11763.57013.88770.12203.71064.1982
Concorde-MHM1.68290.03751.62101.78482.21320.05342.11382.34262.67860.06272.49742.82683.10750.07182.95883.2348
Table 6. Information Loss obtained on the Madrid dataset.
Table 6. Information Loss obtained on the Madrid dataset.
Madrid
k = 3 k = 4 k = 5 k = 6
MethodAverage σ minmaxAverage σ minmaxAverage σ minmaxAverage σ minmax
MDAV3.1876NANANA4.3353NANANA5.2883NANANA5.8235NANANA
V-MDAV3.1876NANANA4.3353NANANA5.2883NANANA5.8235NANANA
MDAV-LK-MHM2.98720.12852.72003.19464.05360.13983.68044.33144.85410.16644.46805.18565.57030.21635.09316.0088
Nearest Insertion-MHM2.75110.08142.57822.91163.70390.11223.43043.96234.45220.15354.15334.84635.15440.15494.86615.5510
Farthest Insertion-MHM2.66830.05582.53192.82803.61870.07423.46053.77554.33380.11314.12604.56685.05980.11724.83915.3372
Cheapest Insertion-MHM2.78330.07492.65172.97893.75310.08043.52533.98304.47520.11404.31634.73565.24960.13455.01475.5609
Arbitrary Insertion-MHM2.74760.07572.60092.91603.71560.09863.52133.98284.41490.14204.05834.70785.10700.14374.76875.3754
Nearest Neighbor-MHM3.42570.17143.08163.90404.75530.21164.28235.37365.76710.21945.18076.31916.76150.25076.18717.4355
Repetitive NN-MHM3.12360.13452.87993.54304.41410.14824.12545.00125.39110.21275.08946.16766.48650.22236.17647.3492
Concorde-MHM2.48450.03362.40532.57283.43020.04663.32493.56644.11240.07743.98164.32284.80660.10654.65385.0534
Table 7. Information Loss obtained on the Tarraco dataset.
Table 7. Information Loss obtained on the Tarraco dataset.
Tarraco
k = 3 k = 4 k = 5 k = 6
MethodAverage σ minmaxAverage σ minmaxAverage σ minmaxAverage σ minmax
MDAV0.9988NANANA1.4180NANANA1.7683NANANA2.0260NANANA
V-MDAV0.9988NANANA1.3093NANANA1.7182NANANA2.0051NANANA
MDAV-LK-MHM1.13650.01541.09791.14651.42160.02031.41151.47231.72010.04011.69951.82572.02380.04042.00612.1247
Nearest Insertion-MHM0.91130.03450.84901.01001.26340.07451.10521.43061.59880.11601.42201.88391.91050.15171.70182.2870
Farthest Insertion-MHM0.91900.03680.85821.02681.22170.04901.11231.37551.50400.05811.39651.71181.83460.06121.75332.1299
Cheapest Insertion-MHM0.95000.04060.89751.09621.29510.05571.22701.46371.62000.08701.52251.86771.97040.10941.85842.2471
Arbitrary Insertion-MHM0.92580.04550.85891.02691.25300.07531.14191.45381.56950.09711.44541.83121.90510.12651.74752.3396
Nearest Neighbor-MHM1.50800.19371.16242.01892.13410.22321.58812.67252.64990.26712.08023.22713.30410.41232.65574.3884
Repetitive NN-MHM1.21770.12861.02761.59061.78060.15991.42442.11312.25450.18821.91462.73942.73840.22092.30733.4314
Concorde-MHM0.84820.01790.81670.90051.10310.03241.07391.23481.38050.05561.32751.68131.72800.06521.66102.1308
Table 8. Summary of the Pearson correlation between Path Length and SSE.
Table 8. Summary of the Pearson correlation between Path Length and SSE.
Dataset k = 3 k = 4 k = 5 k = 6
Census0.480.390.320.28
EIA0.620.670.740.76
Tarragona0.700.720.820.71
Barcelona0.830.810.810.80
Madrid0.840.810.800.78
Tarraco0.800.820.820.80
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Maya-López, A.; Casino, F.; Solanas, A. Improving Multivariate Microaggregation through Hamiltonian Paths and Optimal Univariate Microaggregation. Symmetry 2021, 13, 916. https://doi.org/10.3390/sym13060916

AMA Style

Maya-López A, Casino F, Solanas A. Improving Multivariate Microaggregation through Hamiltonian Paths and Optimal Univariate Microaggregation. Symmetry. 2021; 13(6):916. https://doi.org/10.3390/sym13060916

Chicago/Turabian Style

Maya-López, Armando, Fran Casino, and Agusti Solanas. 2021. "Improving Multivariate Microaggregation through Hamiltonian Paths and Optimal Univariate Microaggregation" Symmetry 13, no. 6: 916. https://doi.org/10.3390/sym13060916

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop