scispace - formally typeset
Search or ask a question

Showing papers on "Sorting published in 2009"


Book
31 Jul 2009
TL;DR: Pseudo-code explanation of the algorithms coupled with proof of their accuracy makes this book a great resource on the basic tools used to analyze the performance of algorithms.
Abstract: If you had to buy just one text on algorithms, Introduction to Algorithms is a magnificent choice. The book begins by considering the mathematical foundations of the analysis of algorithms and maintains this mathematical rigor throughout the work. The tools developed in these opening sections are then applied to sorting, data structures, graphs, and a variety of selected algorithms including computational geometry, string algorithms, parallel models of computation, fast Fourier transforms (FFTs), and more. This book's strength lies in its encyclopedic range, clear exposition, and powerful analysis. Pseudo-code explanation of the algorithms coupled with proof of their accuracy makes this book is a great resource on the basic tools used to analyze the performance of algorithms.

2,972 citations


Journal ArticleDOI
TL;DR: This paper presents a reversible or lossless watermarking algorithm for images without using a location map in most cases that employs prediction errors to embed data into an image.
Abstract: This paper presents a reversible or lossless watermarking algorithm for images without using a location map in most cases. This algorithm employs prediction errors to embed data into an image. A sorting technique is used to record the prediction errors based on magnitude of its local variance. Using sorted prediction errors and, if needed, though rarely, a reduced size location map allows us to embed more data into the image with less distortion. The performance of the proposed reversible watermarking scheme is evaluated using different images and compared with four methods: those of Kamstra and Heijmans, Thodi and Rodriguez, and Lee et al. The results clearly indicate that the proposed scheme can embed more data with less distortion.

773 citations


Journal ArticleDOI
TL;DR: Methods to reduce their computational cost are proposed in this paper and, on average, the enhanced KM algorithms can save about two iterations, which corresponds to more than a 39% reduction in computation time.
Abstract: The Karnik-Mendel (KM) algorithms are iterative procedures widely used in fuzzy logic theory. They are known to converge monotonically and superexponentially fast; however, several (usually two to six) iterations are still needed before convergence occurs. Methods to reduce their computational cost are proposed in this paper. Extensive simulations show that, on average, the enhanced KM algorithms can save about two iterations, which corresponds to more than a 39% reduction in computation time. An additional (at least) 23% computational cost can be saved if no sorting of the inputs is needed.

505 citations


Journal ArticleDOI
TL;DR: The existence of a single optimal acquisition and sorting policy with a simple structure is shown and it is shown that this policy is independent of production amount when acquisition costs are linear.
Abstract: The condition of the used items acquired by remanufacturers is often highly variable, and sorting is an important aspect of remanufacturing operations. Sorting policies—the rules specifying which used products should be remanufactured and which should be scrapped—have received limited attention in the literature. In this paper, we examine the case of a remanufacturer who acquires unsorted used products as needed from third party brokers. As more used items are acquired for a given demand, the remanufacturer can be more selective when sorting. Thus, two related decisions are made: how many used items to acquire, and how selective to be during the sorting process. We derive optimal acquisition and sorting policies in the presence of used product condition variability for a remanufacturer facing both deterministic and uncertain demand. We show the existence of a single optimal acquisition and sorting policy with a simple structure and show that this policy is independent of production amount when acquisition costs are linear.

278 citations


Proceedings ArticleDOI
29 Jun 2009
TL;DR: This paper proposes new data structures that efficiently support an order-preserving dictionary compression for (variablelength) string attributes with a large domain size that is likely to change over time and introduces a novel indexing approach that provides efficient access paths to such a dictionary while compressing the index data.
Abstract: Column-oriented database systems [19, 23] perform better than traditional row-oriented database systems on analytical workloads such as those found in decision support and business intelligence applications Moreover, recent work [1, 24] has shown that lightweight compression schemes significantly improve the query processing performance of these systems One such a lightweight compression scheme is to use a dictionary in order to replace long (variable-length) values of a certain domain with shorter (fixedlength) integer codes In order to further improve expensive query operations such as sorting and searching, column-stores often use order-preserving compression schemes In contrast to the existing work, in this paper we argue that orderpreserving dictionary compression does not only pay off for attributes with a small fixed domain size but also for long string attributes with a large domain size which might change over time Consequently, we introduce new data structures that efficiently support an order-preserving dictionary compression for (variablelength) string attributes with a large domain size that is likely to change over time The main idea is that we model a dictionary as a table that specifies a mapping from string-values to arbitrary integer codes (and vice versa) and we introduce a novel indexing approach that provides efficient access paths to such a dictionary while compressing the index data Our experiments show that our data structures are as fast as (or in some cases even faster than) other state-of-the-art data structures for dictionaries while being less memory intensive

202 citations


Patent
29 Dec 2009
TL;DR: In this paper, a method is disclosed wherein the webpages listed in the result set of a search engine is sorted according to the relevance of webpages to a list of prioritised search terms.
Abstract: A method is disclosed wherein the webpages listed in the result set of a search engine is sorted according to the relevance of the webpages to a list of prioritised search terms. Search terms which are phrases that are delimited by prepositions are considered search terms with high priority. Search terms which nouns are set to high priority. Search terms which are adjectives, verbs, auxiliary verbs, articles, conjunctions, pronouns and prepositions are set to low priority.

193 citations


Patent
31 Mar 2009
TL;DR: In this paper, the authors present a system and methods for locating digital media content, which include receiving search terms from a client device through a single interface, searching a plurality of digital media Content providers for digital media contents using the search terms to obtain search results, sorting and organizing the search results and returning the results to the client device.
Abstract: Systems and methods for locating digital media content. The systems and methods include receiving search terms from a client device through a single interface, searching a plurality of digital media content providers for digital media content using the search terms to obtain search results, sorting and organizing the search results, and returning the search results to the client device through the single interface.

191 citations


Patent
19 Aug 2009
TL;DR: A conveying assembly may have a generally continuous conveying track and is operable to automatically provide the trays to the induct of the at least one sorting assembly in a sequenced manner as discussed by the authors.
Abstract: A method and apparatus for sorting articles to a delivery point sequence includes at least one article sorter adapted to sort articles and a conveying system operable to arrange and convey containers of articles sorted in a first sort pass to an induct of the article sorter for a second sort pass or process of the articles. The system or method automatically sorts and conveys trays containing sorted articles from a discharge of at least one sorting assembly to an induct of the at least one sorting assembly. The conveying assembly may have a generally continuous conveying track and is operable to automatically provide the trays to the induct of the at least one sorting assembly in a sequenced manner.

155 citations


Proceedings ArticleDOI
16 Mar 2009
TL;DR: The experimental results demonstrate that this newly proposed algorithm yields noticeably better time and space efficiencies than all the currently published linear time algorithms for SA construction.
Abstract: We present a linear time and space suffix array (SA) construction algorithm called the SA-IS algorithm.The SA-IS algorithm is novel because of the LMS-substrings used for the problem reduction and the pure induced-sorting (specially coined for this algorithm)used to propagate the order of suffixes as well as that of LMS-substrings, which makes the algorithm almost purely relying on induced sorting at both its crucial steps.The pure induced-sorting renders the algorithm an elegant design and in turn a surprisingly compact implementation which consists of less than 100 lines of C code.The experimental results demonstrate that this newly proposed algorithm yields noticeably better time and space efficiencies than all the currently published linear time algorithms for SA construction.

154 citations


Book
29 Jun 2009
TL;DR: In this paper, Demanding Documents 2. Sorting systems 3. Card Cartel 4. Stretched Screens 5. Body Badges 6. Cyber-Citizens Bibliography Index
Abstract: Introduction 1. Demanding Documents 2. Sorting Systems 3. Card Cartel 4. Stretched Screens 5. Body Badges 6. Cyber-Citizens Bibliography Index

150 citations


Journal ArticleDOI
TL;DR: An approximation model based on queuing network theory is proposed to analyze the impact of order batching and picking area zoning on the mean order throughput time in a pick-and-pass order picking system and shows acceptable accuracy for practical purposes.

Proceedings ArticleDOI
15 Jun 2009
TL;DR: Three novel algorithms are presented, having different strengths, that combine template and predicate abstraction based formalisms to discover required sophisticated program invariants using SMT solvers, and can verify program properties that, to the knowledge, have not been automatically verified before.
Abstract: We address the problem of automatically generating invariants with quantified and boolean structure for proving the validity of given assertions or generating pre-conditions under which the assertions are valid. We present three novel algorithms, having different strengths, that combine template and predicate abstraction based formalisms to discover required sophisticated program invariants using SMT solvers.Two of these algorithms use an iterative approach to compute fixed-points (one computes a least fixed-point and the other computes a greatest fixed-point), while the third algorithm uses a constraint based approach to encode the fixed-point. The key idea in all these algorithms is to reduce the problem of invariant discovery to that of finding optimal solutions for unknowns (over conjunctions of some predicates from a given set) in a template formula such that the formula is valid.Preliminary experiments using our implementation of these algorithms show encouraging results over a benchmark of small but complicated programs. Our algorithms can verify program properties that, to our knowledge, have not been automatically verified before. In particular, our algorithms can generate full correctness proofs for sorting algorithms (which requires nested universally-existentially quantified invariants) and can also generate preconditions required to establish worst-case upper bounds of sorting algorithms. Furthermore, for the case of previously considered properties, in particular sortedness in sorting algorithms, our algorithms take less time than reported by previous techniques.

Journal ArticleDOI
TL;DR: According to the sorting model of education, the impact of education on civic engagement is relative, rather than absolute as discussed by the authors, and the analysis confirms the model, but considerably narrows its reach, since sorting applies only to one particular type of electoral activity, only when the educational environment accounts for variation across age and place, and only when one models the interactive relationship between education at the individual and environmental levels.
Abstract: According to the sorting model of education, the impact of education on civic engagement is relative, rather than absolute. Education correlates with greater engagement because it is a marker of social status; the degree of status conferred by your level of education is determined by the average level of education within your environment. This article tests the sorting model by paying strict heed to its assumptions. The analysis confirms the model, but considerably narrows its reach. Sorting applies only to one particular type (electoral activity), only when the educational environment accounts for variation across age and place, and only when one models the interactive relationship between education at the individual and environmental levels. Furthermore, sorting applies more to men than women. The same analytical framework demonstrates that being in a more highly educated environment amplifies the relationship between education and democratic enlightenment (political knowledge and tolerance).

Journal ArticleDOI
TL;DR: A new approach to analyze categorization data called FAST that stands for Factorial Approach for Sorting Task data, based on multiple correspondence analysis (MCA), provides an optimal representation of the products, an optimal representations of the consumers, which are to be interpreted jointly.

Journal ArticleDOI
TL;DR: This paper studies the impact of controls on the average customer order throughput time in a numerical way, for warehouses with time window batching and separate picking and sorting functions.

Book ChapterDOI
21 Aug 2009
TL;DR: It is shown that the working space for computing Burrows-Wheeler Transform directly in linear time is O(n log*** loglog *** n ) for any *** where *** is the alphabet size, which is the smallest among the known linear time algorithms.
Abstract: To compute Burrows-Wheeler Transform (BWT), one usually builds a suffix array (SA) first, and then obtains BWT using SA, which requires much redundant working space. In previous studies to compute BWT directly [5,12], one constructs BWT incrementally, which requires O(n logn ) time where n is the length of the input text. We present an algorithm for computing BWT directly in linear time by modifying the suffix array construction algorithm based on induced sorting [15]. We show that the working space is O(n log*** loglog *** n ) for any *** where *** is the alphabet size, which is the smallest among the known linear time algorithms.

Journal ArticleDOI
TL;DR: Simulation results showed that FL-NSGA2 outperformed other search methods in all of three various scenarios, and the role of fuzzy logic is to dynamically adjust the crossover rate and mutation rate after ten consecutive generations.
Abstract: In the field of supply chain management and logistics, using vehicles to deliver products from suppliers to customers is one of the major operations. Before transporting products, optimizing the routing of vehicles is required so as to provide a low-cost and efficient service for customers. This paper deals with the problem of optimization of vehicle routing in which multiple depots, multiple customers, and multiple products are considered. Since the total traveling time is not always restrictive as a time constraint, the objective considered in this paper comprises not only the total traveling distance, but also the total traveling time. We propose using a multi-objective evolutionary algorithm called the fuzzy logic guided non-dominated sorting genetic algorithm 2 (FL-NSGA2) to solve this multi-objective optimization problem. The role of fuzzy logic is to dynamically adjust the crossover rate and mutation rate after ten consecutive generations. In order to demonstrate the effectiveness of FL-NSGA2, we compared it with the following: non-dominated sorting genetic algorithms 2 (NSGA2) (without the guide of fuzzy logic), strength Pareto evolutionary algorithm 2 (SPEA2) (with and without the guide of fuzzy logic), and micro-genetic algorithm (MICROGA) (with and without the guide of fuzzy logic). Simulation results showed that FL-NSGA2 outperformed other search methods in all of three various scenarios.

Journal ArticleDOI
TL;DR: In this article, a conceptual model is proposed to identify the hydraulic processes involved in the hydrochory pathway that may lead to differential response of different seed types and hence their sorting.
Abstract: The species composition of riparian vegetation is determined by a variety of processes, including the dispersal of seeds. The seed types that end up at certain locations are determined by their particular characteristics as well as the prevailing hydrology and the physical characteristics of the river channel. A conceptual model is proposed to identify the hydraulic processes involved in the hydrochory pathway that may lead to differential response of different seed types and hence their sorting. The results of a series of laboratory experiments are presented to demonstrate the sorting between seed types by these processes. The behaviour of non-buoyant seeds during settling and entrainment is shown to be consistent with that of mineral sediments, and their sorting can therefore be described by existing sediment transport theory. Significant sorting of buoyant seed types takes place during transport under the influence of wind and during ingress into marginal emergent vegetation, trapping by vegetation and stranding during receding flows on sloping channel banks. Copyright © 2008 John Wiley & Sons, Ltd.

Journal ArticleDOI
TL;DR: The upper bound for the pancake problem is improved to (18/11)n+O(1)?(1.6363)n, which is closer to the minimum number of prefix reversals sufficient for sorting any permutation of length n.

Patent
23 Sep 2009
TL;DR: In this article, a method and system for sorting alert and offer messages on a mobile device is presented, where each alert message contains transaction data associated with a transaction conducted with a merchant.
Abstract: A method and system are presented for sorting alert and offer messages on a mobile device. One embodiment of the invention is directed to a method including receiving alert messages at a mobile device where each alert message contains transaction data associated with a transaction conducted with a merchant. Upon receiving the alert messages, the mobile device receives alert message sorting criteria from a user associated with the mobile device, sorts the alert messages according to the received offer sorting criteria, and displays the sorted alert messages.

Journal ArticleDOI
TL;DR: In this article, a non-dominated Sorting Genetic Algorithm (NSGA-II) was used to optimize the cutting conditions, yielding a nondominated solution set that is reported here.
Abstract: Optimization of cutting parameters is important to achieving high quality in the machining process, especially where more complex multiple performance optimization is required. The present investigation focuses on the multiple performance optimization on machining characteristics of glass fiber reinforced plastic (GFRP) composites. The cutting parameters used for the experiments, which were carried out according to Taguchi’s L27, 3-level orthogonal array, were cutting speed, feed and depth of cut. Statistical models based on second-order polynomial equations were developed for the different responses. The Non-dominated Sorting Genetic Algorithm (NSGA-II) tool was used to optimize the cutting conditions, yielding a non-dominated solution set that is reported here.

Patent
09 Apr 2009
TL;DR: In this paper, a target binding component, a scaffold component and an encoding component formed by standardized molecular units that can be coupled and decoupled in a controlled fashion are presented.
Abstract: Polynucleotide-encoded capture agents for target detection and in particular modular polynucleotide-capture agents comprising a target binding component, a scaffold component and an encoding component formed by standardized molecular units that can be coupled and decoupled in a controlled fashion, and related compositions methods and systems.

Book ChapterDOI
13 Sep 2009
TL;DR: A high-performance in-place implementation of Batcher's bitonic sorting networks for CUDA-enabled GPUs is presented, adapted bitonic sort for arbitrary input length and assigned compare/exchange-operations to threads in a way that decreases low-performance global-memory access and thereby greatly increases the performance of the implementation.
Abstract: State of the art graphics processors provide high processing power and furthermore, the high programmability of GPUs offered by frameworks like CUDA increases their usability as high-performance coprocessors for general-purpose computing. Sorting is well-investigated in Computer Science in general, but (because of this new field of application for GPUs) there is a demand for high-performance parallel sorting algorithms that fit to the characteristics of modern GPU-architecture. We present a high-performance in-place implementation of Batcher's bitonic sorting networks for CUDA-enabled GPUs. We adapted bitonic sort for arbitrary input length and assigned compare/exchange-operations to threads in a way that decreases low-performance global-memory access and thereby greatly increases the performance of the implementation.

Patent
11 Sep 2009
TL;DR: In this paper, a method and an apparatus for sorting mined material, such as mined ore, is disclosed, which comprises a chamber for exposing fragments of a material to be sorted to electromagnetic radiation, with the chamber comprising an inner wall for fragments to move downwardly and outwardly over from an upper inlet of the chamber to a lower outlet of the wall.
Abstract: A method and an apparatus for sorting mined material, such as mined ore, are disclosed. The apparatus comprises a chamber for exposing fragments of a material to be sorted to electromagnetic radiation, with the chamber comprising an inner wall for fragments to move downwardly and outwardly over from an upper inlet of the chamber to a lower outlet of the chamber. The apparatus also comprises a detection system for assessing one or more than one characteristic of the fragments. The apparatus also comprises a sorting means for separating the fragments into multiple streams in response to the assessment of the detection system.

Proceedings ArticleDOI
01 Dec 2009
TL;DR: This paper proposes to adopt mixtures of multivariate skew t distributions to handle highly asymmetric data and uses the EM algorithm to compute the maximum likelihood estimates of model parameters.
Abstract: In many applied problems in the context of pattern recognition, the data often involve highly asymmetric observations. Normal mixture models tend to overfit when additional components are included to capture the skewness of the data. Increased number of pseudo-components could lead to difficulties and inefficiencies in computations. Also, the contours of the fitted mixture components may be distorted. In this paper, we propose to adopt mixtures of multivariate skew t distributions to handle highly asymmetric data. The EM algorithm is used to compute the maximum likelihood estimates of model parameters. The method is illustrated using a flurorescence-activated cell sorting data.

Proceedings ArticleDOI
29 Mar 2009
TL;DR: A generic framework, termed TSS, for handling partially ordered domains using topological sorting is introduced and a novel dominance check is proposed that eliminates false hits/misses, further enhancing progressiveness and pruning ability.
Abstract: The vast majority of work on skyline queries considers totally ordered domains, whereas in many applications some attributes are partially ordered, as for instance, domains of set values, hierarchies, intervals and preferences. The only work addressing this issue has limited progressiveness and pruning ability, and it is only applicable to static skylines. This paper overcomes these problems with the following contributions. (i) We introduce a generic framework, termed TSS, for handling partially ordered domains using topological sorting. (ii) We propose a novel dominance check that eliminates false hits/misses, further enhancing progressiveness and pruning ability. (iii) We extend our methodology to dynamic skylines with respect to an input query. In this case, the dominance relationships change according to the query specification, and their computation is rather complex. We perform an extensive experimental evaluation demonstrating that TSS is up to 9 times and up to 2 orders of magnitude faster than existing methods in the static and the dynamic case, respectively.

Proceedings ArticleDOI
18 May 2009
TL;DR: With this rank-sum sorting and diversified selection, the speed of the algorithm has increased significantly, in particular by about twenty times on five objective problems when compared with the implementation using the non-domination sorting.
Abstract: In this paper, Multi-objective evolutionary programming (MOEP) using fuzzy rank-sum with diversified selection is introduced. The performances of this algorithm as well as MOEP with non-domination sorting on the set of benchmark functions provided for CEC2009 Special Session and competition on Multi-objective Optimization are reported. With this rank-sum sorting and diversified selection, the speed of the algorithm has increased significantly, in particular by about twenty times on five objective problems when compared with the implementation using the non-domination sorting. Beside this, the proposed approach has performed either comparable or better than the MOEP with non-domination sorting.

Patent
30 Sep 2009
TL;DR: In this paper, column based data encoded structures enabling efficient query processing over large scale data storage are queried with respect to complex queries implicating filter and/or sort operations for data over a defined window.
Abstract: The subject disclosure relates to querying of column based data encoded structures enabling efficient query processing over large scale data storage, and more specifically with respect to complex queries implicating filter and/or sort operations for data over a defined window. In this regard, in various embodiments, a method is provided that avoids scenarios involving expensive sorting of a high percentage of, or all, rows, either by not sorting any rows at all, or by sorting only a very small number of rows consistent with or smaller than a number of rows associated with the size of the requested window over the data. In one embodiment, this is achieved by splitting an external query request into two different internal sub-requests, a first one that computes statistics about distribution of rows for any specified WHERE clauses and ORDER BY columns, and a second one that selects only the rows that match the window based on the statistics.

Proceedings ArticleDOI
18 May 2009
TL;DR: This paper shows that there is no direct correlation between the time complexity of an algorithm and its energy consumption, and that different sorting algorithms consume different amounts of energy.
Abstract: Most mobile and embedded devices are battery powered. Hence, their uptime depends on the energy consumption of the used components. Developers made severe effort to optimize hardware components in order to reduce their energy consumption. However, in this paper we show that one also has to consider energy awareness in terms of software. In this study we focus on sorting algorithms, which are not only used directly by the user of a device but also very often implicitly by other algorithms. Our experiments show,that different sorting algorithms consume different amounts of energy. In detail, the experiments show that there is no direct correlation between the time complexity of an algorithm and its energy consumption.

Proceedings ArticleDOI
23 Jan 2009
TL;DR: This paper proposes partial yet efficient codebook search algorithm which uses sorting technique and uses only comparison and hence it is fastest as compared to other search methods ES, HOSM, DTPC.
Abstract: In this paper we propose partial yet efficient codebook search algorithm which uses sorting technique and uses only comparison. Our proposed algorithm does not use Euclidean distance computation and hence it is fastest as compared to other search methods ES, HOSM, DTPC. Form the results it is observed that proposed algorithm gives more MSE as compared to the exhaustive search method but with good execution speed. We also discuss codebook design methods LBG and FCG. The codebooks of different sizes 128, 256, 512 and 1024 are generated using LBG and FCG algorithm. Both the codebook generation algorithms are compared with respect to the execution speed. All the various search algorithms are implemented on the codebooks of different sizes 128, 256, 512 and 1024 obtained from LBG and FCG algorithms. From the results it is observed that FCG codebook gives better performance parameters MSE and PSNR as compared to LBG codebook and among the search algorithm proposed algorithm gives least time to encode the image with slight degradation in image quality.