DSpace Collection:
http://hdl.handle.net/2117/639
Fri, 22 May 2015 22:14:09 GMT2015-05-22T22:14:09Zwebmaster.bupc@upc.eduUniversitat Politècnica de Catalunya. Servei de Biblioteques i DocumentaciónoAutomatic multi-partite graph generation from arbitrary data
http://hdl.handle.net/2117/26446
Title: Automatic multi-partite graph generation from arbitrary data
Authors: Álvarez García, Sandra; Baeza Yates, Ricardo; Brisaboa, Nieves R.; Larriba Pey, Josep; Pedreira, Oscar
Abstract: In this paper we present a generic model for automatic generation of basic multi-partite graphs obtained from collections of arbitrary input data following user indications. The paper also presents GraphGen, a tool that implements this model. The input data is a collection of complex objects composed by a set or list of heterogeneous elements. Our tool provides a simple interface for the user to specify the types of nodes that are relevant for the application domain in each case. The nodes and the relationships between them are derived from the input data through the application of a set of derivation rules specified by the user. The resulting graph can be exported in the standard GraphML format so that it can be further processed with other graph management and mining systems. We end by giving some examples in real scenarios that show the usefulness of this model.Fri, 20 Feb 2015 11:24:35 GMThttp://hdl.handle.net/2117/264462015-02-20T11:24:35ZÁlvarez García, Sandra; Baeza Yates, Ricardo; Brisaboa, Nieves R.; Larriba Pey, Josep; Pedreira, OscarnoGraph processing and analysis, Automatic graph generationIn this paper we present a generic model for automatic generation of basic multi-partite graphs obtained from collections of arbitrary input data following user indications. The paper also presents GraphGen, a tool that implements this model. The input data is a collection of complex objects composed by a set or list of heterogeneous elements. Our tool provides a simple interface for the user to specify the types of nodes that are relevant for the application domain in each case. The nodes and the relationships between them are derived from the input data through the application of a set of derivation rules specified by the user. The resulting graph can be exported in the standard GraphML format so that it can be further processed with other graph management and mining systems. We end by giving some examples in real scenarios that show the usefulness of this model.Two-way replacement selection
http://hdl.handle.net/2117/25227
Title: Two-way replacement selection
Authors: Martínez Palau, Xavier; Domínguez Sal, David; Larriba Pey, Josep
Abstract: The performance of external sorting using merge sort is highly dependent on the length of the runs generated. One of the most commonly used run generation strategies is Replacement Selection (RS) because, on average, it generates runs that are twice the size of the memory available. However, the length of the runs generated by RS is downsized for data with certain characteristics, like inputs sorted inversely with respect to the desired output order.
The goal of this paper is to propose and analyze two-way replacement selection (2WRS), which is a generalization of RS obtained by implementing two heaps instead of the single heap implemented by RS. The appropriate management of these two heaps allows generating runs larger than the memory available in a stable way, i.e. independent from the characteristics of the datasets. Depending on the changing characteristics of the input dataset, 2WRS assigns a new data record to one or the other heap, and grows or shrinks each heap, accommodating to the growing or decreasing tendency of the dataset. On average, 2WRS creates runs of at least the length generated by RS, and longer for datasets that combine increasing and decreasing data subsets. We tested both algorithms on large datasets with different characteristics and 2WRS achieves speedups at least similar to RS, and over 2.5 when RS fails to generate large runs.Mon, 12 Jan 2015 15:12:35 GMThttp://hdl.handle.net/2117/252272015-01-12T15:12:35ZMartínez Palau, Xavier; Domínguez Sal, David; Larriba Pey, JosepnoSorting, Out of core sorting, Heap sort, External sorting, Replacement selection, Merge sort, Run formationThe performance of external sorting using merge sort is highly dependent on the length of the runs generated. One of the most commonly used run generation strategies is Replacement Selection (RS) because, on average, it generates runs that are twice the size of the memory available. However, the length of the runs generated by RS is downsized for data with certain characteristics, like inputs sorted inversely with respect to the desired output order.
The goal of this paper is to propose and analyze two-way replacement selection (2WRS), which is a generalization of RS obtained by implementing two heaps instead of the single heap implemented by RS. The appropriate management of these two heaps allows generating runs larger than the memory available in a stable way, i.e. independent from the characteristics of the datasets. Depending on the changing characteristics of the input dataset, 2WRS assigns a new data record to one or the other heap, and grows or shrinks each heap, accommodating to the growing or decreasing tendency of the dataset. On average, 2WRS creates runs of at least the length generated by RS, and longer for datasets that combine increasing and decreasing data subsets. We tested both algorithms on large datasets with different characteristics and 2WRS achieves speedups at least similar to RS, and over 2.5 when RS fails to generate large runs.Using genetic algorithms for attribute grouping in multivariate microaggregation
http://hdl.handle.net/2117/25011
Title: Using genetic algorithms for attribute grouping in multivariate microaggregation
Authors: Balasch Masoliver, Jordi; Muntés Mulero, Víctor; Nin Guerrero, Jordi
Abstract: Anonymization techniques that provide k-anonymity suffer from loss of quality when data dimensionality is high. Microaggregation techniques are not an exception. Given a set of records, attributes are grouped into non-intersecting subsets and microaggregated independently. While this improves quality by reducing the loss of information, it usually leads to the loss of the k-anonymity property, increasing entity disclosure risk. In spite of this, grouping attributes is still a common practice for data sets containing a large number of records. Depending on the attributes chosen and their correlation, the amount of information loss and disclosure risk vary. However, there have not been serious attempts to propose a way to find the best way of grouping attributes. In this paper, we present GOMM, the Genetic Optimizer for Multivariate Microaggregation which, as far as we know, represents the first proposal using evolutionary algorithms for this problem. The goal of GOMM is finding the optimal, or near-optimal, attribute grouping taking into account both information loss and disclosure risk. We propose a way to map attribute subsets into a chromosome and a set of new mutation operations for this context. Also, we provide a comprehensive analysis of the operations proposed and we show that, after using our evolutionary approach for different real data sets, we obtain better quality in the anonymized data comparing it to previously used ad-hoc attribute grouping techniques. Additionally, we provide an improved version of GOMM called D-GOMM where operations are dynamically executed during the optimization process to reduce the GOMM execution time.Fri, 12 Dec 2014 12:11:36 GMThttp://hdl.handle.net/2117/250112014-12-12T12:11:36ZBalasch Masoliver, Jordi; Muntés Mulero, Víctor; Nin Guerrero, JordinoGenetic clustering algorithms, multivariate microaggregation, attribute selection, K-anonymity, Disclosure control, Record linkage, PrivacyAnonymization techniques that provide k-anonymity suffer from loss of quality when data dimensionality is high. Microaggregation techniques are not an exception. Given a set of records, attributes are grouped into non-intersecting subsets and microaggregated independently. While this improves quality by reducing the loss of information, it usually leads to the loss of the k-anonymity property, increasing entity disclosure risk. In spite of this, grouping attributes is still a common practice for data sets containing a large number of records. Depending on the attributes chosen and their correlation, the amount of information loss and disclosure risk vary. However, there have not been serious attempts to propose a way to find the best way of grouping attributes. In this paper, we present GOMM, the Genetic Optimizer for Multivariate Microaggregation which, as far as we know, represents the first proposal using evolutionary algorithms for this problem. The goal of GOMM is finding the optimal, or near-optimal, attribute grouping taking into account both information loss and disclosure risk. We propose a way to map attribute subsets into a chromosome and a set of new mutation operations for this context. Also, we provide a comprehensive analysis of the operations proposed and we show that, after using our evolutionary approach for different real data sets, we obtain better quality in the anonymized data comparing it to previously used ad-hoc attribute grouping techniques. Additionally, we provide an improved version of GOMM called D-GOMM where operations are dynamically executed during the optimization process to reduce the GOMM execution time.A discussion on the design of graph database benchmarks
http://hdl.handle.net/2117/24027
Title: A discussion on the design of graph database benchmarks
Authors: Domínguez Sal, David; Martínez Bazán, Norbert; Muntés Mulero, Víctor; Baleta Ferrer, Pedro; Larriba Pey, Josep
Abstract: Graph Database Management systems (GDBs) are gaining popularity. They are used to analyze huge graph datasets that are naturally appearing in many application areas to model interrelated data. The objective of this paper is to raise a new topic of discussion in the benchmarking community and allow practitioners having a set of basic guidelines for GDB benchmarking. We strongly believe that GDBs will become an important player in the market field of data analysis, and with that, their performance and capabilities will also become important. For this reason, we discuss those aspects that are important from our perspective, i.e. the characteristics of the graphs to be included in the benchmark, the characteristics of the queries that are important in graph analysis applications and the evaluation workbench.Wed, 10 Sep 2014 08:42:32 GMThttp://hdl.handle.net/2117/240272014-09-10T08:42:32ZDomínguez Sal, David; Martínez Bazán, Norbert; Muntés Mulero, Víctor; Baleta Ferrer, Pedro; Larriba Pey, JosepnoGraph Database Management systems (GDBs) are gaining popularity. They are used to analyze huge graph datasets that are naturally appearing in many application areas to model interrelated data. The objective of this paper is to raise a new topic of discussion in the benchmarking community and allow practitioners having a set of basic guidelines for GDB benchmarking. We strongly believe that GDBs will become an important player in the market field of data analysis, and with that, their performance and capabilities will also become important. For this reason, we discuss those aspects that are important from our perspective, i.e. the characteristics of the graphs to be included in the benchmark, the characteristics of the queries that are important in graph analysis applications and the evaluation workbench.Generalized median string computation by means of string embedding in vector spaces
http://hdl.handle.net/2117/19430
Title: Generalized median string computation by means of string embedding in vector spaces
Authors: Jiang, Xiaoyi; Wentker, Jöran; Ferrer Sumsi, Miquel
Abstract: In structural pattern recognition the median string has been established as a useful tool to represent a set of strings. However, its exact computation is complex and of high computational burden. In this paper we propose a new approach for the computation of median string based on string embedding. Strings are embedded into a vector space and the median is computed in the vector domain. We apply three different inverse transformations to go from the vector domain back to the string domain in order to obtain a final approximation of the median string. All of them are based on the weighted mean of a pair of strings. Experiments show that we succeed to compute good approximations of the median string.Tue, 28 May 2013 16:06:00 GMThttp://hdl.handle.net/2117/194302013-05-28T16:06:00ZJiang, Xiaoyi; Wentker, Jöran; Ferrer Sumsi, MiquelnoString, Generalized median, Embedding, Vector space, Lower boundIn structural pattern recognition the median string has been established as a useful tool to represent a set of strings. However, its exact computation is complex and of high computational burden. In this paper we propose a new approach for the computation of median string based on string embedding. Strings are embedded into a vector space and the median is computed in the vector domain. We apply three different inverse transformations to go from the vector domain back to the string domain in order to obtain a final approximation of the median string. All of them are based on the weighted mean of a pair of strings. Experiments show that we succeed to compute good approximations of the median string.Using Evolutive Summary Counters for Efficient Cooperative Caching in Search Engines
http://hdl.handle.net/2117/16552
Title: Using Evolutive Summary Counters for Efficient Cooperative Caching in Search Engines
Authors: Domínguez Sal, David; Aguilar Saborit, Josep; Surdeanu, Mihai; Larriba Pey, Josep
Description: We propose and analyze a distributed cooperative
caching strategy based on the Evolutive Summary Counters
(ESC), a new data structure that stores an approximated record
of the data accesses in each computing node of a search engine.
The ESC capture the frequency of accesses to the elements
of a data collection, and the evolution of the access patterns
for each node in a network of computers. The ESC can be
efficiently summarized into what we call ESC-summaries to
obtain approximate statistics of the document entries accessed
by each computing node.
We use the ESC-summaries to introduce two algorithms that
manage our distributed caching strategy, one for the distribution
of the cache contents, ESC-placement, and another one for the
search of documents in the distributed cache, ESC-search. While
the former improves the hit rate of the system and keeps a large
ratio of data accesses local, the latter reduces the network traffic
by restricting the number of nodes queried to find a document.
We show that our cooperative caching approach outperforms
state of the art models in both hit rate, throughput, and location
recall for multiple scenarios, i.e., different query distributions
and systems with varying degrees of complexity.Fri, 21 Sep 2012 09:29:32 GMThttp://hdl.handle.net/2117/165522012-09-21T09:29:32ZDomínguez Sal, David; Aguilar Saborit, Josep; Surdeanu, Mihai; Larriba Pey, JosepnoSocial based layouts for the increase of locality in graph operations
http://hdl.handle.net/2117/13533
Title: Social based layouts for the increase of locality in graph operations
Authors: Prat Pérez, Arnau; Domínguez Sal, David; Larriba Pey, Josep
Abstract: Graphs provide a natural data representation for analyzing the relationships among entities in many application areas. Since the
analysis algorithms perform memory intensive operations, it is important that the graph layout is adapted to take advantage of the memory hierarchy.
Here, we propose layout strategies based on community detection to improve the in-memory data locality of generic graph algorithms. We
conclude that the detection of communities in a graph provides a layout strategy that improves the performance of graph algorithms consistently over other state of the art strategies.Mon, 17 Oct 2011 11:15:59 GMThttp://hdl.handle.net/2117/135332011-10-17T11:15:59ZPrat Pérez, Arnau; Domínguez Sal, David; Larriba Pey, JosepnoGraph mining, BandwidthGraphs provide a natural data representation for analyzing the relationships among entities in many application areas. Since the
analysis algorithms perform memory intensive operations, it is important that the graph layout is adapted to take advantage of the memory hierarchy.
Here, we propose layout strategies based on community detection to improve the in-memory data locality of generic graph algorithms. We
conclude that the detection of communities in a graph provides a layout strategy that improves the performance of graph algorithms consistently over other state of the art strategies.Cooperative cache analysis for distributed search engines
http://hdl.handle.net/2117/13116
Title: Cooperative cache analysis for distributed search engines
Authors: Domínguez Sal, David; Pérez Casany, Marta; Larriba Pey, Josep
Abstract: In this paper, we study the performance of a distributed search engine from a data caching point of view using statistical tools on a varied set of configurations. We study two strategies to achieve better performance: cacheaware load balancing that issues the queries to nodes that store the computation in cache; and cooperative caching (CC) that stores and transfers the available computed contents from one node in the network to others. Since cache-aware
decisions depend on information about the recent history, we also analyse how the ageing of this information impacts the system performance. Our results show that the combination of both strategies yield better throughput than individually implementing cooperative cache or cache-aware load balancing strategies because
of a synergic improvement of the hit rate. Furthermore, the analysis concludes that the data structures to monitor the system need only moderate precision to achieve optimal throughput.Thu, 25 Aug 2011 10:59:55 GMThttp://hdl.handle.net/2117/131162011-08-25T10:59:55ZDomínguez Sal, David; Pérez Casany, Marta; Larriba Pey, JosepnoIn this paper, we study the performance of a distributed search engine from a data caching point of view using statistical tools on a varied set of configurations. We study two strategies to achieve better performance: cacheaware load balancing that issues the queries to nodes that store the computation in cache; and cooperative caching (CC) that stores and transfers the available computed contents from one node in the network to others. Since cache-aware
decisions depend on information about the recent history, we also analyse how the ageing of this information impacts the system performance. Our results show that the combination of both strategies yield better throughput than individually implementing cooperative cache or cache-aware load balancing strategies because
of a synergic improvement of the hit rate. Furthermore, the analysis concludes that the data structures to monitor the system need only moderate precision to achieve optimal throughput.