scispace - formally typeset
Search or ask a question
Author

Tanuja Sarode

Bio: Tanuja Sarode is an academic researcher from University of Mumbai. The author has contributed to research in topics: Wavelet transform & Vector quantization. The author has an hindex of 18, co-authored 140 publications receiving 1264 citations. Previous affiliations of Tanuja Sarode include Narsee Monjee Institute of Management Studies & Thadomal Shahani Engineering College.


Papers
More filters
01 Jan 2009
TL;DR: A novel technique for image retrieval using the color- texture features extracted from images based on vector quantization with Kekre's fast codebook generation is proposed, which gives better discrimination capability for Content Based Image Retrieval (CBIR).
Abstract: novel technique for image retrieval using the color- texture features extracted from images based on vector quantization with Kekre's fast codebook generation is proposed. This gives better discrimination capability for Content Based Image Retrieval (CBIR). Here the database image is divided into 2x2 pixel windows to obtain 12 color descriptors per window (Red, Green and Blue per pixel) to form a vector. Collection of all such vectors is a training set. Then the Kekre's Fast Codebook Generation (KFCG) is applied on this set to get 16 codevectors. The Discrete Cosine Transform (DCT) is applied on these codevectors by converting them to column vector. This transform vector is used as the image signature (feature vector) for image retrieval. The method takes lesser computations as compared to conventional DCT applied on complete image. The method gives the color-texture features of the image database at reduced feature set size. Proposed method avoids resizing of images which is required for any transform based feature extraction method.

91 citations

Proceedings ArticleDOI
16 Jul 2008
TL;DR: The proposed algorithm uses sorting method to generate codebook and the codevectors are obtained by using median approach and it gives less MSE as compared to the LBG for the codebooks of sizes 128, 256, 512 & 1024 respectively.
Abstract: In this paper we present a very simple and yet effective algorithm to generate codebook. The algorithm uses sorting method to generate codebook and the codevectors are obtained by using median approach. The proposed algorithm was experimented on six different images each of size 512 x 512 and four different codebooks of sizes 128, 256, 512 and 1024 are generated. The proposed algorithm is found to be much faster than the LBG and KPE algorithm. The performance of this algorithm is better than LBG and KPE algorithms considering MSE, PSNR and execution time. The proposed algorithm gives less MSE as compared to the LBG for the codebooks of sizes 128, 256, 512 & 1024 respectively. It also gives higher PSNR as compared to LBG for the codebooks of various sizes.

64 citations

Journal Article
TL;DR: A new performance parameter Average Fractional Change in Speech Sample (AFCSS) is introduced and the FCG algorithm gives far better performance considering mean absolute error, AFCSS and complexity as compared to others.
Abstract: Mostly transforms are used for speech data compressions which are lossy algorithms. Such algorithms are tolerable for speech data compression since the loss in quality is not perceived by the human ear. However the vector quantization (VQ) has a potential to give more data compression maintaining the same quality. In this paper we propose speech data compression algorithm using vector quantization technique. We have used VQ algorithms LBG, KPE and FCG. The results table shows computational complexity of these three algorithms. Here we have introduced a new performance parameter Average Fractional Change in Speech Sample (AFCSS). Our FCG algorithm gives far better performance considering mean absolute error, AFCSS and complexity as compared to others. Keywords—Vector Quantization, Data Compression, Encoding,, Speech coding.

54 citations

Proceedings ArticleDOI
23 Jan 2009
TL;DR: This paper proposes partial yet efficient codebook search algorithm which uses sorting technique and uses only comparison and hence it is fastest as compared to other search methods ES, HOSM, DTPC.
Abstract: In this paper we propose partial yet efficient codebook search algorithm which uses sorting technique and uses only comparison. Our proposed algorithm does not use Euclidean distance computation and hence it is fastest as compared to other search methods ES, HOSM, DTPC. Form the results it is observed that proposed algorithm gives more MSE as compared to the exhaustive search method but with good execution speed. We also discuss codebook design methods LBG and FCG. The codebooks of different sizes 128, 256, 512 and 1024 are generated using LBG and FCG algorithm. Both the codebook generation algorithms are compared with respect to the execution speed. All the various search algorithms are implemented on the codebooks of different sizes 128, 256, 512 and 1024 obtained from LBG and FCG algorithms. From the results it is observed that FCG codebook gives better performance parameters MSE and PSNR as compared to LBG codebook and among the search algorithm proposed algorithm gives least time to encode the image with slight degradation in image quality.

50 citations

Journal Article
TL;DR: A novel technique for image retrieval using the color-texture features extracted from images based on the color indexing using vector quantization to give better discrimination capability for CBIR.
Abstract: Image retrieval has become imperative area of research because of vide range of applications needing the image data search facility. Most of the research approaches in the area are either database based indexing or image processing based CBIR. The hours need is to combine these parallel going approaches of research to have better image retrieval techniques. The paper proposes a novel technique for image retrieval using the color-texture features extracted from images based on the color indexing using vector quantization. This gives better discrimination capability for CBIR. Here we are dividing the database image into 2x2 pixel windows to obtain 12 color descriptors (Per pixel Red, Green and Blue) per row of window table. Then the Kekre’s Median Codebook Generation (KMCG) is applied on window table to get 256 centre rows. The DCT is applied on this centre row vector to obtain feature set of size 256x12, which is user for image retrieval. The method takes fewer computations as compared to conventional DCT applied on complete image. The method gives the color-texture features of the image database at reduced feature set size.

50 citations


Cited by
More filters
01 Jan 1979
TL;DR: This special issue aims at gathering the recent advances in learning with shared information methods and their applications in computer vision and multimedia analysis and addressing interesting real-world computer Vision and multimedia applications.
Abstract: In the real world, a realistic setting for computer vision or multimedia recognition problems is that we have some classes containing lots of training data and many classes contain a small amount of training data. Therefore, how to use frequent classes to help learning rare classes for which it is harder to collect the training data is an open question. Learning with Shared Information is an emerging topic in machine learning, computer vision and multimedia analysis. There are different level of components that can be shared during concept modeling and machine learning stages, such as sharing generic object parts, sharing attributes, sharing transformations, sharing regularization parameters and sharing training examples, etc. Regarding the specific methods, multi-task learning, transfer learning and deep learning can be seen as using different strategies to share information. These learning with shared information methods are very effective in solving real-world large-scale problems. This special issue aims at gathering the recent advances in learning with shared information methods and their applications in computer vision and multimedia analysis. Both state-of-the-art works, as well as literature reviews, are welcome for submission. Papers addressing interesting real-world computer vision and multimedia applications are especially encouraged. Topics of interest include, but are not limited to: • Multi-task learning or transfer learning for large-scale computer vision and multimedia analysis • Deep learning for large-scale computer vision and multimedia analysis • Multi-modal approach for large-scale computer vision and multimedia analysis • Different sharing strategies, e.g., sharing generic object parts, sharing attributes, sharing transformations, sharing regularization parameters and sharing training examples, • Real-world computer vision and multimedia applications based on learning with shared information, e.g., event detection, object recognition, object detection, action recognition, human head pose estimation, object tracking, location-based services, semantic indexing. • New datasets and metrics to evaluate the benefit of the proposed sharing ability for the specific computer vision or multimedia problem. • Survey papers regarding the topic of learning with shared information. Authors who are unsure whether their planned submission is in scope may contact the guest editors prior to the submission deadline with an abstract, in order to receive feedback.

1,758 citations

Journal ArticleDOI
Alan R. Jones1

1,349 citations

Journal ArticleDOI
TL;DR: Experimental results show that the learning-based method proposed can accurately predict CT images in various scenarios, even for the images undergoing large shape variation, and also outperforms two state-of-the-art methods.
Abstract: Computed tomography (CT) imaging is an essential tool in various clinical diagnoses and radiotherapy treatment planning. Since CT image intensities are directly related to positron emission tomography (PET) attenuation coefficients, they are indispensable for attenuation correction (AC) of the PET images. However, due to the relatively high dose of radiation exposure in CT scan, it is advised to limit the acquisition of CT images. In addition, in the new PET and magnetic resonance (MR) imaging scanner, only MR images are available, which are unfortunately not directly applicable to AC. These issues greatly motivate the development of methods for reliable estimate of CT image from its corresponding MR image of the same subject. In this paper, we propose a learning-based method to tackle this challenging problem. Specifically, we first partition a given MR image into a set of patches. Then, for each patch, we use the structured random forest to directly predict a CT patch as a structured output, where a new ensemble model is also used to ensure the robust prediction. Image features are innovatively crafted to achieve multi-level sensitivity, with spatial information integrated through only rigid-body alignment to help avoiding the error-prone inter-subject deformable registration. Moreover, we use an auto-context model to iteratively refine the prediction. Finally, we combine all of the predicted CT patches to obtain the final prediction for the given MR image. We demonstrate the efficacy of our method on two datasets: human brain and prostate images. Experimental results show that our method can accurately predict CT images in various scenarios, even for the images undergoing large shape variation, and also outperforms two state-of-the-art methods.

238 citations

Journal ArticleDOI
TL;DR: Pynamical, an open-source Python package to easily visualize and explore nonlinear dynamical systems’ behavior and introduce the foundations of nonlinear dynamics, chaos, fractals, self-similarity and the limits of prediction is presented.
Abstract: Nearly all nontrivial real-world systems are nonlinear dynamical systems. Chaos describes certain nonlinear dynamical systems that have a very sensitive dependence on initial conditions. Chaotic systems are always deterministic and may be very simple, yet they produce completely unpredictable and divergent behavior. Systems of nonlinear equations are difficult to solve analytically, and scientists have relied heavily on visual and qualitative approaches to discover and analyze the dynamics of nonlinearity. Indeed, few fields have drawn as heavily from visualization methods for their seminal innovations: from strange attractors, to bifurcation diagrams, to cobweb plots, to phase diagrams and embedding. Although the social sciences are increasingly studying these types of systems, seminal concepts remain murky or loosely adopted. This article has three aims. First, it argues for several visualization methods to critically analyze and understand the behavior of nonlinear dynamical systems. Second, it uses these visualizations to introduce the foundations of nonlinear dynamics, chaos, fractals, self-similarity and the limits of prediction. Finally, it presents Pynamical, an open-source Python package to easily visualize and explore nonlinear dynamical systems' behavior.

230 citations