scispace - formally typeset
Search or ask a question

Showing papers on "Interval tree published in 2014"


Proceedings ArticleDOI
06 Feb 2014
TL;DR: A new lock-free algorithm for concurrent manipulation of a binary search tree in an asynchronous shared memory system that supports search, insert and delete operations and significantly outperforms all other algorithms for a concurrent binarysearch tree in many cases.
Abstract: We present a new lock-free algorithm for concurrent manipulation of a binary search tree in an asynchronous shared memory system that supports search, insert and delete operations. In addition to read and write instructions, our algorithm uses (single-word) compare-and-swap (CAS) and bit-test-and-set (SETB) atomic instructions, both of which are commonly supported by many modern processors including Intel~64 and AMD64.In contrast to existing lock-free algorithms for a binary search tree, our algorithm is based on marking edges rather than nodes. As a result, when compared to other lock-free algorithms, modify (insert and delete) operations in our algorithm work on a smaller portion of the tree, thereby reducing conflicts, and execute fewer atomic instructions (one for insert and three for delete). Our experiments indicate that our lock-free algorithm significantly outperforms all other algorithms for a concurrent binary search tree in many cases, especially when contention is high, by as much as 100%.

191 citations


Journal ArticleDOI
TL;DR: This paper proposes a new method which significantly outperforms the McDiarmid tree algorithm and has a solid mathematical basis and ensures, with a high probability set by the user, that the best attribute chosen in the considered node using a finite data sample is the same as it would be in the case of the whole data stream.
Abstract: Since the Hoeffding tree algorithm was proposed in the literature, decision trees became one of the most popular tools for mining data streams. The key point of constructing the decision tree is to determine the best attribute to split the considered node. Several methods to solve this problem were presented so far. However, they are either wrongly mathematically justified (e.g., in the Hoeffding tree algorithm) or time-consuming (e.g., in the McDiarmid tree algorithm). In this paper, we propose a new method which significantly outperforms the McDiarmid tree algorithm and has a solid mathematical basis. Our method ensures, with a high probability set by the user, that the best attribute chosen in the considered node using a finite data sample is the same as it would be in the case of the whole data stream.

156 citations


Journal ArticleDOI
TL;DR: A new efficient edge-preserving filter-“tree filter”-to achieve strong image smoothing, which can smooth out high-contrast details while preserving major edges, which is not achievable for bilateral-filter-like techniques.
Abstract: We present a new efficient edge-preserving filter-“tree filter”-to achieve strong image smoothing. The proposed filter can smooth out high-contrast details while preserving major edges, which is not achievable for bilateral-filter-like techniques. Tree filter is a weighted-average filter, whose kernel is derived by viewing pixel affinity in a probabilistic framework simultaneously considering pixel spatial distance, color/intensity difference, as well as connectedness. Pixel connectedness is acquired by treating pixels as nodes in a minimum spanning tree (MST) extracted from the image. The fact that an MST makes all image pixels connected through the tree endues the filter with the power to smooth out high-contrast, fine-scale details while preserving major image structures, since pixels in small isolated region will be closely connected to surrounding majority pixels through the tree, while pixels inside large homogeneous region will be automatically dragged away from pixels outside the region. The tree filter can be separated into two other filters, both of which turn out to have fast algorithms. We also propose an efficient linear time MST extraction algorithm to further improve the whole filtering speed. The algorithms give tree filter a great advantage in low computational complexity (linear to number of image pixels) and fast speed: it can process a 1-megapixel 8-bit image at ~ 0.25 s on an Intel 3.4 GHz Core i7 CPU (including the construction of MST). The proposed tree filter is demonstrated on a variety of applications.

124 citations


Journal ArticleDOI
TL;DR: This work presents an in-depth comparison of the existing algorithms in a unique framework, as well as variations of some of them that improve their efficiency.
Abstract: Connected operators are morphological tools that have the property of filtering images without creating new contours and without moving the contours that are preserved. Those operators are related to the max-tree and min-tree representations of images, and many algorithms have been proposed to compute those trees. However, no exhaustive comparison of these algorithms has been proposed so far, and the choice of an algorithm over another depends on many parameters. Since the need for fast algorithms is obvious for production code, we present an in-depth comparison of the existing algorithms in a unique framework, as well as variations of some of them that improve their efficiency. This comparison involves both sequential and parallel algorithms, and execution times are given w.r.t. the number of threads, the input image size, and the pixel value quantization. Eventually, a decision tree is given to help the user choose the most appropriate algorithm with respect to the user requirements. To favor reproducible research, an online demo allows the user to upload an image and bench the different algorithms, and the source code of every algorithms has been made available.

115 citations


Proceedings ArticleDOI
06 Feb 2014
TL;DR: In this paper, the authors describe a general technique for obtaining provably correct, non-blocking implementations of a large class of tree data structures where pointers are directed from parents to children, where updates are permitted to modify any contiguous portion of the tree atomically.
Abstract: We describe a general technique for obtaining provably correct, non-blocking implementations of a large class of tree data structures where pointers are directed from parents to children. Updates are permitted to modify any contiguous portion of the tree atomically. Our non-blocking algorithms make use of the LLX, SCX and VLX primitives, which are multi-word generalizations of the standard LL, SC and VL primitives and have been implemented from single-word CAS. To illustrate our technique, we describe how it can be used in a fairly straightforward way to obtain a non-blocking implementation of a chromatic tree, which is a relaxed variant of a red-black tree. The height of the tree at any time is O(c + log n), where n is the number of keys and c is the number of updates in progress. We provide an experimental performance analysis which demonstrates that our Java implementation of a chromatic tree rivals, and often significantly outperforms, other leading concurrent dictionaries.

105 citations


Proceedings ArticleDOI
06 Feb 2014
TL;DR: The experimental results show that the algorithms with lock-free contains and on-time deletion are practical and often comparable to the state-of-the-art.
Abstract: We present practical, concurrent binary search tree (BST) algorithms that explicitly maintain logical ordering information in the data structure, permitting clean separation from its physical tree layout. We capture logical ordering using intervals, with the property that an item belongs to the tree if and only if the item is an endpoint of some interval. We are thus able to construct efficient, synchronization-free and intuitive lookup operations. We present (i) a concurrent non-balanced BST with a lock-free lookup, and (ii) a concurrent AVL tree with a lock-free lookup that requires no synchronization with any mutating operations, including balancing operations. Our algorithms apply on-time deletion; that is, every request for removal of a node, results in its immediate removal from the tree. This new feature did not exist in previous concurrent internal tree algorithms.We implemented our concurrent BST algorithms and evaluated them against several state-of-the-art concurrent tree algorithms. Our experimental results show that our algorithms with lock-free contains and on-time deletion are practical and often comparable to the state-of-the-art.

91 citations


Posted Content
TL;DR: This paper discusses an alternative algorithm that builds a balanced k-d tree by presorting the data in each of k dimensions prior to building the tree and preserves the order of these k sorts during tree construction and thereby avoids the requirement for any further sorting.
Abstract: The original description of the k-d tree recognized that rebalancing techniques, such as are used to build an AVL tree or a red-black tree, are not applicable to a k-d tree. Hence, in order to build a balanced k-d tree, it is necessary to find the median of the data for each recursive partition. The choice of selection or sort that is used to find the median for each subdivision strongly influences the computational complexity of building a k-d tree. This paper discusses an alternative algorithm that builds a balanced k-d tree by presorting the data in each of k dimensions prior to building the tree. It then preserves the order of these k sorts during tree construction and thereby avoids the requirement for any further sorting. Moreover, this algorithm is amenable to parallel execution via multiple threads. Compared to an algorithm that finds the median for each recursive subdivision, this presorting algorithm has equivalent performance for four dimensions and better performance for three or fewer dimensions.

73 citations


Journal ArticleDOI
TL;DR: The user study suggests that Tree Colors are useful, not only for improving node-link diagrams, but also for unveiling tree structure in non-hierarchical visualizations.
Abstract: We present a method to map tree structures to colors from the Hue-Chroma-Luminance color model, which is known for its well balanced perceptual properties The Tree Colors method can be tuned with several parameters, whose effect on the resulting color schemes is discussed in detail We provide a free and open source implementation with sensible parameter defaults Categorical data are very common in statistical graphics, and often these categories form a classification tree We evaluate applying Tree Colors to tree structured data with a survey on a large group of users from a national statistical institute Our user study suggests that Tree Colors are useful, not only for improving node-link diagrams, but also for unveiling tree structure in non-hierarchical visualizations

55 citations


Patent
19 Aug 2014
TL;DR: In this article, a decision tree representation is generated and stored on persistent storage in depth-first order, and a respective predictive utility metric (PUM) value is determined for one or more nodes, indicating expected contributions of the nodes to a prediction.
Abstract: During a training phase of a machine learning model, representations of at least some nodes of a decision tree are generated and stored on persistent storage in depth-first order. A respective predictive utility metric (PUM) value is determined for one or more nodes, indicating expected contributions of the nodes to a prediction of the model. A particular node is selected for removal from the tree based at least partly on its PUM value. A modified version of the tree, with the particular node removed, is stored for obtaining a prediction.

54 citations


Proceedings ArticleDOI
15 Jul 2014
TL;DR: This work improves upon an existing non-blocking implementation of a binary search tree from single-word compare-and-swap instructions and shows that the worst-case amortized step complexity of performing a Find, Insert or Delete operation op on the tree is O(h(op)+c(op).
Abstract: We improve upon an existing non-blocking implementation of a binary search tree from single-word compare-and-swap instructions. We show that the worst-case amortized step complexity of performing a Find, Insert or Delete operation op on the tree is O(h(op)+c(op)) where h(op) is the height of the tree at the beginning of op and c(op) is the maximum number of operations accessing the tree at any one time during op. This is the first bound on the complexity of a non-blocking implementation of a search tree.

44 citations


Proceedings Article
01 Jan 2014
TL;DR: This paper adapts the B + -Tree and prefix B-Tree by changing the search algorithm on inner nodes from binary search to k-ary search, and introduces two tree adaptations that satisfy the specific constraints of SIMD instructions.
Abstract: In this paper, we accelerate the processing of tree-based index structures by using SIMD instructions. We adapt the B + -Tree and prefix B-Tree (trie) by changing the search algorithm on inner nodes from binary search to k-ary search. The k-ary search enables the use of SIMD instructions, which are commonly available on most modern processors today. The main challenge for using SIMD instructions on CPUs is their inherent requirement for consecutive memory loads. The data for one SIMD load instruction must be located in consecutive memory locations and cannot be scattered over the entire memory. The original layout of tree-based index structures does not satisfy this constraint and must be adapted to enable SIMD usage. Thus, we introduce two tree adaptations that satisfy the specific constraints of SIMD instructions. We present two di↵erent algorithms for transforming the original tree layout into a SIMD-friendly layout. Additionally, we introduce two SIMD-friendly search algorithms designed for the new layout. Our adapted B + -Tree speeds up search processes by a factor of up to eight for small data types compared to the original B + -Tree using binary search. Furthermore, our adapted prefix B-Tree enables a high search performance even for larger data types. We report a constant 14 fold speedup and an 8 fold reduction in memory consumption compared to the original B + -Tree.

Journal ArticleDOI
TL;DR: This paper presents a binary tree search algorithm for the three dimensional container loading problem (3D-CLP), which achieves good results for the well-known 3D- CLP instances suggested by Bischoff and Ratcliff with reasonable computing time.

Proceedings ArticleDOI
01 Jun 2014
TL;DR: A new tree kernel, called feature-enriched tree kernel (FTK), is proposed, which can enhance the traditional tree kernel by refining the syntactic tree representation by annotating each tree node with a set of discriminant features.
Abstract: Tree kernel is an effective technique for relation extraction. However, the traditional syntactic tree representation is often too coarse or ambiguous to accurately capture the semantic relation information between two entities. In this paper, we propose a new tree kernel, called feature-enriched tree kernel (FTK), which can enhance the traditional tree kernel by: 1) refining the syntactic tree representation by annotating each tree node with a set of discriminant features; and 2) proposing a new tree kernel which can better measure the syntactic tree similarity by taking all features into consideration. Experimental results show that our method can achieve a 5.4% F-measure improvement over the traditional convolution tree kernel.

Proceedings ArticleDOI
05 Jan 2014
TL;DR: An online algorithm is given that maintains a Steiner tree under only deletions, and an algorithm that changes only a constant number of edges upon each request, and maintains a constant-competitive tree at all times is given.
Abstract: In the online Steiner tree problem, the input is a set of vertices that appear one-by-one, and we have to maintain a Steiner tree on the current set of vertices The cost of the tree is the total length of edges in the tree, and we want this cost to be close to the cost of the optimal Steiner tree at all points in time If we are allowed to only add edges, a tight bound of Θ(log n) on the competitiveness has been known for two decades Recently it was shown that if we can add one new edge and make one edge swap upon every vertex arrival, we can still maintain a constant-competitive tree onlineBut what if the set of vertices sees both additions and deletions? Again, we would like to obtain a low-cost Steiner tree with as few edge changes as possible The original paper of Imase and Waxman (SIAM J Disc Math, 4(3): 369--384, 1991) had also considered this model, and it gave an algorithm that made at most O(n3/2) edge changes for the first n requests, and maintained a constant-competitive tree online In this paper we improve on these results:• We give an online algorithm that maintains a Steiner tree under only deletions: we start off with a set of vertices, and at each time one of the vertices is removed from this set---our Steiner tree no longer has to span this vertex We give an algorithm that changes only a constant number of edges upon each request, and maintains a constant-competitive tree at all times Our algorithm uses the primal-dual framework and a global charging argument to carefully make these constant number of changes• We also give an algorithm that maintains a Steiner tree in the fully-dynamic model (where each request either adds or deletes a vertex) Our algorithm for this setting makes a constant number of changes per request in an amortized sense

Proceedings ArticleDOI
01 Jan 2014
TL;DR: This is the first example of a grammar-based tree compression algorithm with an approximation ratio polynomial in g, where g is the size of a minimal linear context-free tree grammar for T, and r is the maximal rank of symbols in T.
Abstract: A simple linear-time algorithm for constructing a linear context-free tree grammar of size O(r^2.g.log(n)) for a given input tree T of size n is presented, where g is the size of a minimal linear context-free tree grammar for T, and r is the maximal rank of symbols in T (which is a constant in many applications). This is the first example of a grammar-based tree compression algorithm with an approximation ratio polynomial in g. The analysis of the algorithm uses an extension of the recompression technique (used in the context of grammar-based string compression) from strings to trees.

Journal ArticleDOI
TL;DR: Simulation results show that the pruning method based on t-test outperforms the method which prunes the binary tree using a fixed threshold.
Abstract: Network tomography is a promising inference technique for network topology from end-to-end measurements. In this letter, we propose a novel binary tree pruning algorithm based on t-test to infer the network topology. A binary tree topology is first inferred using the existing Agglomerative Likelihood Tree (ALT) method, and then two samples t-test is applied to prune the binary tree, thus a general tree corresponding to the real topology is obtained. A lower bound on the correctly identified probability of the proposed method is derived. Simulation results show that the pruning method based on t-test outperforms the method which prunes the binary tree using a fixed threshold.

Patent
10 Apr 2014
TL;DR: In this paper, a scalable data structure (SDF) is proposed to provide sequential access to data using scalable data structures, which includes a first data structure, e.g., hash map, and a second data structure.
Abstract: The technology is directed to providing sequential access to data using scalable data structures. In some embodiments, the scalable data structures include a first data structure, e.g., hash map, and a second data structure, e.g., tree data structure (“tree”). The technology receives multiple key-value pairs representing data associated with an application. A key includes a prefix and a suffix. While the suffixes of the keys are distinct, some of the keys can have the same prefix. The technology stores the keys having the same prefix in a tree, and stores the root node of the tree in the first data structure. To retrieve values of a set of input keys with a given prefix, the technology retrieves a root node of a tree corresponding to the given prefix from the first data structure using the given prefix, and traverses the tree to obtain the values in a sequence.

Journal ArticleDOI
TL;DR: This paper focuses on the variations tractable by the algorithms including the submodule of a network algorithm, either the minimum cost maximum flow algorithm or the maximum weighted bipartite matching algorithm, and shows that both network algorithms are replaceable.
Abstract: In this paper, we investigate the problem of computing structural sensitive variations of an unordered tree edit distance. First, we focus on the variations tractable by the algorithms including the submodule of a network algorithm, either the minimum cost maximum flow algorithm or the maximum weighted bipartite matching algorithm. Then, we show that both network algorithms are replaceable, and hence the time complexity of computing these variations can be reduced to O(nmd) time, where n is the number of nodes in a tree, m is the number of nodes in another tree and d is the minimum degree of given two trees. Next, we show that the problem of computing the bottom-up distance is MAX SNP-hard. Note that the well-known linear-time algorithm for the bottom-up distance designed by Valiente (2001) computes just a bottom-up indel (insertion-deletion) distance allowing no substitutions.

Posted ContentDOI
TL;DR: Understanding the left-most and real minimal search graphs leads to some new ideas for enhancing Alpha-Beta search, and one of them, enhanced transposition cutoffs, is shown to significantly reduce search tree size.
Abstract: Knuth and Moore presented a theoretical lower bound on the number of leaves that any fixed-depth minimax tree-search algorithm traversing a uniform tree must explore, the so-called minimal tree. Since real-life minimax trees are not uniform, the exact size of this tree is not known for most applications. Further, most games have transpositions, implying that there exists a minimal graph which is smaller than the minimal tree. For three games (chess, Othello and checkers) we compute the size of the minimal tree and the minimal graph. Empirical evidence shows that in all three games, enhanced Alpha-Beta search is capable of building a tree that is close in size to that of the minimal graph. Hence, it appears game-playing programs build nearly optimal search trees. However, the conventional definition of the minimal graph is wrong. There are ways in which the size of the minimal graph can be reduced: by maximizing the number of transpositions in the search, and generating cutoffs using branches that lead to smaller search trees. The conventional definition of the minimal graph is just a left-most approximation. Calculating the size of the real minimal graph is too computationally intensive. However, upper bound approximations show it to be significantly smaller than the left-most minimal graph. Hence, it appears that game-playing programs are not searching as efficiently as is widely believed. Understanding the left-most and real minimal search graphs leads to some new ideas for enhancing Alpha-Beta search. One of them, enhanced transposition cutoffs, is shown to significantly reduce search tree size.

Posted Content
Alina Beygelzimer1, John Langford2, Yuri Lifshits2, Gregory B. Sorkin1, Alexander Strehl2 
TL;DR: This work proposes the first online algorithm which provably constructs a logarithmic depth tree on the set of labels to solve the problem of estimating the conditional probability of a label in time O(log n), where n is the number of possible labels.
Abstract: We consider the problem of estimating the conditional probability of a label in time O(log n), where n is the number of possible labels. We analyze a natural reduction of this problem to a set of binary regression problems organized in a tree structure, proving a regret bound that scales with the depth of the tree. Motivated by this analysis, we propose the first online algorithm which provably constructs a logarithmic depth tree on the set of labels to solve this problem. We test the algorithm empirically, showing that it works succesfully on a dataset with roughly 106 labels.

01 Jan 2014
TL;DR: Better performance is revealed using the k-d tree, compared to the brute-force approach to make faster, more accurate and efficient data structure primarily depends on a particular data set.
Abstract: Data mining may be viewed as the extraction of the hidden predictive information from large databases, is a powerful new technology with great potential to analyze important information in the data warehouse. Nearest neighbor search (NNS), also known as proximity search, similarity search or closest point search, is an optimization problem for finding closest points in metric spaces. Brute-force search is a very general problem-solving technique that consists of systematically enumerating all possible candidates for the solution and checking whether each candidate satisfies the problem's statement. A Brute-force algorithm for string matching problem has two inputs to be considered: pattern and text. A k-d tree, or k-dimensional tree, is a data structure used for organizing some number of points in a space with k dimensions. K-d trees are very useful for range and nearest neighbour searches. In this paper, we studied and compared k-d tree algorithm and brute force algorithm on various levels. The use of the approximate k-nearest neighbour with K-d Tree data structure and comparing its performance attributes to the brute- force approach. In approximate nearest neighbour to evaluate and compare the efficiency of the data structure when applied on a particular number of points, distance and execution time. The work performed between two techniques and select the best one. The result of the work performed in this paper revealed better performance using the k-d tree, compared to the brute-force approach. The aim of the algorithm is to make faster, more accurate and efficient data structure primarily depends on a particular data set. It can be further expanded as by changing the k-d tree traversal technique. We have proposed a new modified traversal technique for k-d tree.

Patent
20 Nov 2014
TL;DR: In this article, a restart method of traversing a binary tree in a ray tracing system is proposed, which traverses a tree consisting of a hierarchical acceleration structure using one-bit stacks respectively assigned to levels of the binary tree and restarting at the highest level at which a corresponding one bit stack has a value indicating that a child node has not yet been traversed.
Abstract: A restart method of traversing a binary tree in a ray tracing system includes traversing a tree consisting of a hierarchical acceleration structure using one-bit stacks respectively assigned to levels of the binary tree and restarting at a highest level at which a corresponding one-bit stack has a value indicating that a child node has not yet been traversed, thereby decreasing the capacity of a stack memory required.

Patent
John M. Boyer1
28 Feb 2014
TL;DR: In this article, a parse tree data structure that is representative of a document object model (DOM) data structure is received and an assignment of index values for the DOM nodes consisting of distinct index value for each existing DOM node is received by the processor.
Abstract: An illustrative embodiment includes processing parse tree data. A parse tree data structure that is representative of a document object model (DOM) tree data structure is received. Concomitant to receiving the parse tree data structure, an assignment of index values for the DOM nodes consisting of distinct index values for each existing DOM node is received by the processor. Requests to manipulate the parse tree data structure that include node inserts and document order comparisons are also performed.

Journal ArticleDOI
TL;DR: The diameter is investigated into, and it is demonstrated that the average diameters between a gene tree and a species tree grow significantly slower than their naive upper bounds, suggesting that the exact bounds can significantly improve on assessing DC costs when using diameters.
Abstract: The minimizing deep coalescence (MDC) problem seeks a species tree that reconciles the given gene trees with the minimum number of deep coalescence events, called deep coalescence (DC) cost. To better assess MDC species trees we investigate into a basic mathematical property of the DC cost, called the diameter. Given a gene tree, a species tree, and a leaf labeling function that assigns leaf-genes of the gene tree to a leaf-species in the species tree from which they were sampled, the DC cost describes the discordance between the trees caused by deep coalescence events. The diameter of a gene tree and a species tree is the maximum DC cost across all leaf labelings for these trees. We prove fundamental mathematical properties describing precisely these diameters for bijective and general leaf labelings, and present efficient algorithms to compute the diameters and their corresponding leaf labelings. In particular, we describe an optimal, i.e., linear time, algorithm for the bijective case. Finally, in an experimental study we demonstrate that the average diameters between a gene tree and a species tree grow significantly slower than their naive upper bounds, suggesting that our exact bounds can significantly improve on assessing DC costs when using diameters.

Book ChapterDOI
13 Jul 2014
TL;DR: An algorithm to solve the ordinal classification problems, where both the condition attributes with interval values and the decision attributes meet the monotonic requirement is presented.
Abstract: Traditional decision tree algorithms for interval valued data only can deal with non-ordinal classification problems. In this paper, we presented an algorithm to solve the ordinal classification problems, where both the condition attributes with interval values and the decision attributes meet the monotonic requirement. The algorithm uses the rank mutual information to select extended attributes, which guarantees that the outputted decision tree is monotonic. The proposed algorithm is illustrated by a numerical example, and a monotonically consistent decision tree is generated. The design of algorithm can provide some useful guidelines for extending real-vauled to interval-valued attributes in ordinal decision tree induction.

Journal ArticleDOI
TL;DR: In this article, the authors present techniques for speeding up the tree edit distance computation which are applicable to a family of algorithms based on closely related recursion strategies, which aim to reduce repetitious steps in the original algorithms by exploring certain structural features in the tree.
Abstract: An ordered labeled tree is a tree in which the nodes are labeled and the left-to-right order among siblings is relevant. The edit distance between two ordered labeled trees is the minimum cost of transforming one tree into the other through a sequence of edit operations. We present techniques for speeding up the tree edit distance computation which are applicable to a family of algorithms based on closely related recursion strategies. These techniques aim to reduce repetitious steps in the original algorithms by exploring certain structural features in the tree. When these features exist in a large portion of the tree, the speedup due to our techniques would be significant. Viable examples for application include RNA secondary structure comparison and structured text comparison.

Patent
Alexandra Manevitch1
23 Dec 2014
TL;DR: In this article, the authors describe an apparatus and method for implementing a nearest neighbor search on a graphics processing unit (GPU) based on the NN tree search on the host processor.
Abstract: An apparatus and method are described for implementing a nearest neighbor search on a graphics processing unit For example, one embodiment of an apparatus comprises: a host processor; and a graphics processor unit (GPU) to execute a nearest neighbor (NN) tree build operation and a NN tree search operation in response to signals from the host processor; the NN tree build operation comprising dividing a first set of data points to define each tree node to form a plurality of buckets, each bucket comprising at least one of the first set of data points, the number of buckets to be limited based on a threshold value, resulting in at least some buckets with multiple data points; and the NN tree search operation comprising traversing the tree by comparing each of a second set of input data points to data points in the tree to identify a bucket and then comparing one of the second data points to data points in the bucket to identify a nearest neighbor data point in the bucket

Journal ArticleDOI
TL;DR: This paper devises a novel technique to deal with the degree-dependent node costs and applies the technique to develop an approximation algorithm for thedegree-dependent nodes-weighted Steiner tree problem.
Abstract: This paper studies the problem of constructing a minimum-cost multicast tree (or Steiner tree) in which each node is associated with a cost that is dependent on its degree in the multicast tree. The cost of a node may depend on its degree in the multicast tree due to a number of reasons. For example, a node may need to perform various processing for sending messages to each of its neighbors in the multicast tree. Thus, the overhead for processing the messages increases as the number of neighbors increases. This paper devises a novel technique to deal with the degree-dependent node costs and applies the technique to develop an approximation algorithm for the degree-dependent node-weighted Steiner tree problem. The bound on the cost of the tree constructed by the proposed approximation algorithm is derived to be 2((ln (k/2))+1)(W_T∗ + B), where k is the size of the set of multicast members, W_T∗ is the cost of a minimum-cost Steiner tree T∗, and B is related to the degree-dependent node costs. Simulations are carried out to study the performance of the proposed algorithm. A distributed implementation of the proposed algorithm is presented. In addition, the proposed algorithm is generalized to solve the degree-dependent node-weighted constrained forest problem.

Journal ArticleDOI
TL;DR: Three different types of data sets, for which the uniquely most parsimonious tree can be known exactly but is hard to find with heuristic tree search methods, are studied.

Journal ArticleDOI
TL;DR: This paper presents methods by which a SOM with an underlying Binary Search Tree (BST) structure can be adaptively re-structured using Conditional Rotations (CONROT), and introduces the pioneering concept referred to as Neural Promotion, where neurons gain prominence in the Neural Network as their significance increases.