scispace - formally typeset
Search or ask a question
Author

Shai Avidan

Bio: Shai Avidan is an academic researcher from Tel Aviv University. The author has contributed to research in topics: Pixel & Template matching. The author has an hindex of 50, co-authored 138 publications receiving 15378 citations. Previous affiliations of Shai Avidan include Mitsubishi Electric Research Laboratories & Mitsubishi.


Papers
More filters
Patent
06 Apr 2017
TL;DR: In this article, the authors proposed a method for dehazing a digital image and restoring an underwater digital image by clustering pixels of the digital image into haze lines, where each of the haze-lines is comprised of a sub-group of the pixels that are scattered non-locally over the image.
Abstract: Methods for dehazing a digital image and for restoring an underwater digital image. The methods include the following steps: First, clustering pixels of a digital image into haze-lines, wherein each of the haze-lines is comprised of a sub-group of the pixels that are scattered non-locally over the digital image. Second, estimating, based on the haze-lines, a transmission map of the digital image, wherein the transmission map encodes scene depth information for each pixel of the digital image. Then, for a hazy image, calculating a dehazed digital image based on the transmission map. For an underwater image, calculating a restored image based on the transmission map and also based on attenuation coefficient ratios. An optional addition to the underwater image restoration takes into account different attenuation coefficients for different color channels, when the image depicts a scene characterized by wavelength-dependent transmission, such as under water. Further disclosed are methods for airlight estimation and for veiling-light estimation, which may be utilized for the dehazing and restoration, or for other purposes.

16 citations

Journal ArticleDOI
TL;DR: A probabilistic algorithm for finding correspondences across multiple images in systems with large numbers of cameras and considerable overlap is presented, and a heuristic method for discarding false matches is proposed, demonstrating its effectiveness in reducing errors.
Abstract: A probabilistic algorithm is presented for finding correspondences across multiple images in systems with large numbers of cameras and considerable overlap. The algorithm employs the theory of random graphs to provide an efficient probabilistic algorithm that performs Wide-baseline Stereo (WBS) comparisons on a small number of image pairs, and then propagates correspondence information among the cameras. A concrete mathematical analysis of its performance is given. The algorithm is extended to handle false-positive and false-negative failures of the WBS computations. We characterize the detectability of the existence of such failures, and propose an efficient method for this detection. Based on this, we propose a heuristic method for discarding false matches, and demonstrate its effectiveness in reducing errors. Since in many multi-camera applications cameras are attached to processors that handle local processing and communication, it is natural to consider distributed solutions that make use of the local processors and do not use a central computer. Our algorithm is especially suited to run in a distributed setting. If the local processors are sufficiently powerful, this allows an order of magnitude increase in computational efficiency. More importantly, a distributed implementation provides strong robustness guarantees, and eliminates the existence of a single point of failure that is inherent when the application is coordinated by a central computer. We show how to efficiently overcome processor crashes and communication failures with a minimal reduction in the quality of the algorithm's results.

16 citations

Proceedings ArticleDOI
Shai Avidan1
19 Jul 2004
TL;DR: The masking matrix help extend feature and basis selection methods while blurring the lines between them and offers a sub-optimal probabilistic method to find it.
Abstract: We treat feature selection and basis selection in a unified framework by introducing the masking matrix. If one considers feature selection as finding a binary mask vector that determines which features participate in the learning process, and similarly, basis selection as finding a binary mask vector that determines which basis vectors are needed for the learning process, then the masking matrix is, in particular, the outer product of the feature masking vector and the basis masking vector. This representation allows for a joint estimation of both features and basis. In addition, it allows one to select features that appear in only part of the basis functions. This joint selection of feature/basis subset is not possible when using feature selection and basis selection algorithms independently, thus, the masking matrix help extend feature and basis selection methods while blurring the lines between them. The problem of searching for an optimal masking matrix is NP-hand and we offer a sub-optimal probabilistic method to find it. In particular we demonstrate our ideas on the problem of feature and basis selection for SVM classification and show results for the problem of image classification on faces and vehicles.

14 citations

Posted Content
TL;DR: This work is the first to consider the problem of adversarial examples at a geometric level, and demonstrates the robustness of the attack in the case of defense, where it is shown that remnant characteristics of the target shape are still present at the output after applying the defense to the adversarial input.
Abstract: Deep neural networks are prone to adversarial examples that maliciously alter the network's outcome. Due to the increasing popularity of 3D sensors in safety-critical systems and the vast deployment of deep learning models for 3D point sets, there is a growing interest in adversarial attacks and defenses for such models. So far, the research has focused on the semantic level, namely, deep point cloud classifiers. However, point clouds are also widely used in a geometric-related form that includes encoding and reconstructing the geometry. In this work, we are the first to consider the problem of adversarial examples at a geometric level. In this setting, the question is how to craft a small change to a clean source point cloud that leads, after passing through an autoencoder model, to the reconstruction of a different target shape. Our attack is in sharp contrast to existing semantic attacks on 3D point clouds. While such works aim to modify the predicted label by a classifier, we alter the entire reconstructed geometry. Additionally, we demonstrate the robustness of our attack in the case of defense, where we show that remnant characteristics of the target shape are still present at the output after applying the defense to the adversarial input. Our code is publicly available at this https URL.

14 citations

Book ChapterDOI
26 Jun 2000
TL;DR: Two algorithms for recovering multiview geometry by integrating multiple local affine joint images into the global projective joint image are described, including one that directly recovers the image epipoles without recovering the fundamental matrix as an intermediate step.
Abstract: The fundamental matrix defines a nonlinear 3D variety in the joint image space of multiple projective (or "uncalibrated perspective") images. We show that, in the case of two images, this variety is a 4D cone whose vertex is the joint epipole (namely the 4D point obtained by stacking the two epipoles in the two images). Affine (or "para-perspective") projection approximates this nonlinear variety with a linear subspace, both in two views and in multiple views. We also show that the tangent to the projective joint image at any point on that image is obtained by using local affine projection approximations around the corresponding 3D point. We use these observations to develop a new approach for recovering multiview geometry by integrating multiple local affine joint images into the global projective joint image. Given multiple projective images, the tangents to the projective joint image are computed using local affine approximations for multiple image patches. The affine parameters from different patches are combined to obtain the epipolar geometry of pairs of projective images. We describe two algorithms for this purpose, including one that directly recovers the image epipoles without recovering the fundamental matrix as an intermediate step.

13 citations


Cited by
More filters
01 Jan 2001
TL;DR: This book is referred to read because it is an inspiring book to give you more chance to get experiences and also thoughts and it will show the best book collections and completed collections.
Abstract: Downloading the book in this website lists can give you more advantages. It will show you the best book collections and completed collections. So many books can be found in this website. So, this is not only this multiple view geometry in computer vision. However, this book is referred to read because it is an inspiring book to give you more chance to get experiences and also thoughts. This is simple, read the soft file of the book and you get it.

14,282 citations

Book
24 Aug 2012
TL;DR: This textbook offers a comprehensive and self-contained introduction to the field of machine learning, based on a unified, probabilistic approach, and is suitable for upper-level undergraduates with an introductory-level college math background and beginning graduate students.
Abstract: Today's Web-enabled deluge of electronic data calls for automated methods of data analysis. Machine learning provides these, developing methods that can automatically detect patterns in data and then use the uncovered patterns to predict future data. This textbook offers a comprehensive and self-contained introduction to the field of machine learning, based on a unified, probabilistic approach. The coverage combines breadth and depth, offering necessary background material on such topics as probability, optimization, and linear algebra as well as discussion of recent developments in the field, including conditional random fields, L1 regularization, and deep learning. The book is written in an informal, accessible style, complete with pseudo-code for the most important algorithms. All topics are copiously illustrated with color images and worked examples drawn from such application domains as biology, text processing, computer vision, and robotics. Rather than providing a cookbook of different heuristic methods, the book stresses a principled model-based approach, often using the language of graphical models to specify models in a concise and intuitive way. Almost all the models described have been implemented in a MATLAB software package--PMTK (probabilistic modeling toolkit)--that is freely available online. The book is suitable for upper-level undergraduates with an introductory-level college math background and beginning graduate students.

8,059 citations

Journal ArticleDOI
TL;DR: A new superpixel algorithm is introduced, simple linear iterative clustering (SLIC), which adapts a k-means clustering approach to efficiently generate superpixels and is faster and more memory efficient, improves segmentation performance, and is straightforward to extend to supervoxel generation.
Abstract: Computer vision applications have come to rely increasingly on superpixels in recent years, but it is not always clear what constitutes a good superpixel algorithm. In an effort to understand the benefits and drawbacks of existing methods, we empirically compare five state-of-the-art superpixel algorithms for their ability to adhere to image boundaries, speed, memory efficiency, and their impact on segmentation performance. We then introduce a new superpixel algorithm, simple linear iterative clustering (SLIC), which adapts a k-means clustering approach to efficiently generate superpixels. Despite its simplicity, SLIC adheres to boundaries as well as or better than previous methods. At the same time, it is faster and more memory efficient, improves segmentation performance, and is straightforward to extend to supervoxel generation.

7,849 citations

Proceedings ArticleDOI
22 Jan 2006
TL;DR: Some of the major results in random graphs and some of the more challenging open problems are reviewed, including those related to the WWW.
Abstract: We will review some of the major results in random graphs and some of the more challenging open problems. We will cover algorithmic and structural questions. We will touch on newer models, including those related to the WWW.

7,116 citations

Journal ArticleDOI
TL;DR: The goal of this article is to review the state-of-the-art tracking methods, classify them into different categories, and identify new trends to discuss the important issues related to tracking including the use of appropriate image features, selection of motion models, and detection of objects.
Abstract: The goal of this article is to review the state-of-the-art tracking methods, classify them into different categories, and identify new trends. Object tracking, in general, is a challenging problem. Difficulties in tracking objects can arise due to abrupt object motion, changing appearance patterns of both the object and the scene, nonrigid object structures, object-to-object and object-to-scene occlusions, and camera motion. Tracking is usually performed in the context of higher-level applications that require the location and/or shape of the object in every frame. Typically, assumptions are made to constrain the tracking problem in the context of a particular application. In this survey, we categorize the tracking methods on the basis of the object and motion representations used, provide detailed descriptions of representative methods in each category, and examine their pros and cons. Moreover, we discuss the important issues related to tracking including the use of appropriate image features, selection of motion models, and detection of objects.

5,318 citations