scispace - formally typeset
Search or ask a question
Author

Jing-Ming Guo

Bio: Jing-Ming Guo is an academic researcher from National Taiwan University of Science and Technology. The author has contributed to research in topics: Halftone & Digital watermarking. The author has an hindex of 34, co-authored 211 publications receiving 3499 citations. Previous affiliations of Jing-Ming Guo include Chang Gung University & National Taiwan University.


Papers
More filters
01 Jan 2008
TL;DR: This study presents a robust and semi-blind watermarking by embedding information into low frequency AC coefficients of Discrete Cosine Transform (DCT) by utilizing the DC values of the neighboring blocks to predict the AC coefficient of the center block.
Abstract: This study presents a robust and semi-blind watermarking by embedding information into low frequency AC coefficients of Discrete Cosine Transform (DCT). Since the imperceptibility is the most significant issue in watermarking, the DC value is maintained unchanged. The proposed methods utilize the DC values of the neighboring blocks to predict the AC coefficients of the center block. The low frequency AC coefficients are modified to carry watermark information. The Least Mean Squares (LMS) is employed to yield the intermediate filters, cooperating with the neighboring DC coefficients to precisely predict the original AC coefficients. Two watermarking methods, namely Watermark Embedding in the Predicted AC Coefficient (WEPAC) and Watermark Embedding in the Original AC coefficient (WEOAC), are presented in this study. Moreover, many attacks are addressed to show the robustness of the proposed methods.

270 citations

Journal ArticleDOI
TL;DR: This study dedicates itself in License Plate Localization and Character Segmentation, and proposes a hybrid- binarization technique to effectively segment the characters in the dirt LP.
Abstract: License plate localization (LPL) and character segmentation (CS) play key roles in the license plate (LP) recognition system. In this paper, we dedicate ourselves to these two issues. In LPL, histogram equalization is employed to solve the low-contrast and dynamic-range problems; the texture properties, e.g., aspect ratio, and color similarity are used to locate the LP; and the Hough transform is adopted to correct the rotation problem. In CS, the hybrid binarization technique is proposed to effectively segment the characters in the dirt LP. The feedback self-learning procedure is also employed to adjust the parameters in the system. As documented in the experiments, good localization and segmentation results are achieved with the proposed algorithms.

151 citations

Journal ArticleDOI
TL;DR: This work overcomes the problem by adding an additional color feature, namely Color Information Feature (CIF), along with the LBP-based feature in the image retrieval and classification systems, which adequately represent the color and texture features.

139 citations

Journal ArticleDOI
TL;DR: Experimental results show that the proposed method is superior to the block truncation coding image retrieval systems and the other earlier methods, and thus prove that the ODBTC scheme is not only suited for image compression, because of its simplicity, but also offers a simple and effective descriptor to index images in CBIR system.
Abstract: This paper presents a technique for content-based image retrieval (CBIR) by exploiting the advantage of low-complexity ordered-dither block truncation coding (ODBTC) for the generation of image content descriptor. In the encoding step, ODBTC compresses an image block into corresponding quantizers and bitmap image. Two image features are proposed to index an image, namely, color co-occurrence feature (CCF) and bit pattern features (BPF), which are generated directly from the ODBTC encoded data streams without performing the decoding process. The CCF and BPF of an image are simply derived from the two ODBTC quantizers and bitmap, respectively, by involving the visual codebook. Experimental results show that the proposed method is superior to the block truncation coding image retrieval systems and the other earlier methods, and thus prove that the ODBTC scheme is not only suited for image compression, because of its simplicity, but also offers a simple and effective descriptor to index images in CBIR system.

120 citations

Journal ArticleDOI
TL;DR: This paper presents an effective image retrieval method by combining high-level features from convolutional neural network (CNN) model and low- level features from dot-diffused block truncation coding (DDBTC) to improve the overall retrieval rate.
Abstract: This paper presents an effective image retrieval method by combining high-level features from convolutional neural network (CNN) model and low-level features from dot-diffused block truncation coding (DDBTC). The low-level features, e.g., texture and color, are constructed by vector quantization -indexed histogram from DDBTC bitmap, maximum, and minimum quantizers. Conversely, high-level features from CNN can effectively capture human perception. With the fusion of the DDBTC and CNN features, the extended deep learning two-layer codebook features is generated using the proposed two-layer codebook, dimension reduction, and similarity reweighting to improve the overall retrieval rate. Two metrics, average precision rate and average recall rate (ARR), are employed to examine various data sets. As documented in the experimental results, the proposed schemes can achieve superior performance compared with the state-of-the-art methods with either low-or high-level features in terms of the retrieval rate. Thus, it can be a strong candidate for various image retrieval related applications.

118 citations


Cited by
More filters
Proceedings Article
01 Jan 1994
TL;DR: The main focus in MUCKE is on cleaning large scale Web image corpora and on proposing image representations which are closer to the human interpretation of images.
Abstract: MUCKE aims to mine a large volume of images, to structure them conceptually and to use this conceptual structuring in order to improve large-scale image retrieval. The last decade witnessed important progress concerning low-level image representations. However, there are a number problems which need to be solved in order to unleash the full potential of image mining in applications. The central problem with low-level representations is the mismatch between them and the human interpretation of image content. This problem can be instantiated, for instance, by the incapability of existing descriptors to capture spatial relationships between the concepts represented or by their incapability to convey an explanation of why two images are similar in a content-based image retrieval framework. We start by assessing existing local descriptors for image classification and by proposing to use co-occurrence matrices to better capture spatial relationships in images. The main focus in MUCKE is on cleaning large scale Web image corpora and on proposing image representations which are closer to the human interpretation of images. Consequently, we introduce methods which tackle these two problems and compare results to state of the art methods. Note: some aspects of this deliverable are withheld at this time as they are pending review. Please contact the authors for a preview.

2,134 citations

Journal ArticleDOI
TL;DR: This paper provides a state-of-the-art review and analysis of the different existing methods of steganography along with some common standards and guidelines drawn from the literature and some recommendations and advocates for the object-oriented embedding mechanism.

1,572 citations

Journal ArticleDOI
TL;DR: In this article, the authors provide a short overview of recent advances and some associated challenges in machine learning applied to medical image processing and image analysis, and provide a starting point for people interested in experimenting and perhaps contributing to the field of machine learning for medical imaging.
Abstract: What has happened in machine learning lately, and what does it mean for the future of medical image analysis? Machine learning has witnessed a tremendous amount of attention over the last few years. The current boom started around 2009 when so-called deep artificial neural networks began outperforming other established models on a number of important benchmarks. Deep neural networks are now the state-of-the-art machine learning models across a variety of areas, from image analysis to natural language processing, and widely deployed in academia and industry. These developments have a huge potential for medical imaging technology, medical data analysis, medical diagnostics and healthcare in general, slowly being realized. We provide a short overview of recent advances and some associated challenges in machine learning applied to medical image processing and image analysis. As this has become a very broad and fast expanding field we will not survey the entire landscape of applications, but put particular focus on deep learning in MRI. Our aim is threefold: (i) give a brief introduction to deep learning with pointers to core references; (ii) indicate how deep learning has been applied to the entire MRI processing chain, from acquisition to image retrieval, from segmentation to disease prediction; (iii) provide a starting point for people interested in experimenting and perhaps contributing to the field of machine learning for medical imaging by pointing out good educational resources, state-of-the-art open-source code, and interesting sources of data and problems related medical imaging.

991 citations

Journal ArticleDOI
TL;DR: This paper categorizes different ALPR techniques according to the features they used for each stage, and compares them in terms of pros, cons, recognition accuracy, and processing speed.
Abstract: Automatic license plate recognition (ALPR) is the extraction of vehicle license plate information from an image or a sequence of images. The extracted information can be used with or without a database in many applications, such as electronic payment systems (toll payment, parking fee payment), and freeway and arterial monitoring systems for traffic surveillance. The ALPR uses either a color, black and white, or infrared camera to take images. The quality of the acquired images is a major factor in the success of the ALPR. ALPR as a real-life application has to quickly and successfully process license plates under different environmental conditions, such as indoors, outdoors, day or night time. It should also be generalized to process license plates from different nations, provinces, or states. These plates usually contain different colors, are written in different languages, and use different fonts; some plates may have a single color background and others have background images. The license plates can be partially occluded by dirt, lighting, and towing accessories on the car. In this paper, we present a comprehensive review of the state-of-the-art techniques for ALPR. We categorize different ALPR techniques according to the features they used for each stage, and compare them in terms of pros, cons, recognition accuracy, and processing speed. Future forecasts of ALPR are given at the end.

682 citations