scispace - formally typeset
Search or ask a question
Author

Jeng-Shyang Pan

Bio: Jeng-Shyang Pan is an academic researcher from Shandong University of Science and Technology. The author has contributed to research in topics: Digital watermarking & Watermark. The author has an hindex of 50, co-authored 789 publications receiving 11645 citations. Previous affiliations of Jeng-Shyang Pan include National Kaohsiung Normal University & Technical University of Ostrava.


Papers
More filters
Journal ArticleDOI
TL;DR: Wang et al. as discussed by the authors proposed a unified framework of deep learning-based MRI super resolution, which has five current deep learning methods with the best super-resolution effect, and constructed a high-low resolution MR image dataset with the scales of ×2, ×3, and ×4.
Abstract: Magnetic resonance imaging (MRI) is widely used in the detection and diagnosis of diseases. High-resolution MR images will help doctors to locate lesions and diagnose diseases. However, the acquisition of high-resolution MR images requires high magnetic field intensity and long scanning time, which will bring discomfort to patients and easily introduce motion artifacts, resulting in image quality degradation. Therefore, the resolution of hardware imaging has reached its limit. Based on this situation, a unified framework based on deep learning super resolution is proposed to transfer state-of-the-art deep learning methods of natural images to MRI super resolution. Compared with the traditional image super-resolution method, the deep learning super-resolution method has stronger feature extraction and characterization ability, can learn prior knowledge from a large number of sample data, and has a more stable and excellent image reconstruction effect. We propose a unified framework of deep learning -based MRI super resolution, which has five current deep learning methods with the best super-resolution effect. In addition, a high-low resolution MR image dataset with the scales of ×2, ×3, and ×4 was constructed, covering 4 parts of the skull, knee, breast, and head and neck. Experimental results show that the proposed unified framework of deep learning super resolution has a better reconstruction effect on the data than traditional methods and provides a standard dataset and experimental benchmark for the application of deep learning super resolution in MR images.

4 citations

Journal ArticleDOI
TL;DR: Gaussian normal basis (GNB) multiplication based on the proposed architecture can be used to reduce the space complexity of the proposed STMVP scheme.
Abstract: Toeplitz matrix–vector product (TMVP) decomposition is one of the high-precision multiplication algorithms. A symmetric TMVP (STMVP) decomposition is presented and theoretical analysis shows that the space complexity of the proposed STMVP scheme is less compared with the traditional TMVP approach. Gaussian normal basis (GNB) multiplication based on the proposed architecture can be used to reduce the space complexity.

4 citations

Journal ArticleDOI
TL;DR: In this article , the authors present a comprehensively state-of-the-art investigation of the engineering applications utilized by binary metaheuristic algorithms, and identify current troubles and challenges by the conducted review, and discusses that novel binary algorithm, transfer function, benchmark function, time-consuming problem and application integration are need to be resolved in future.
Abstract: This article presents a comprehensively state-of-the-art investigation of the engineering applications utilized by binary metaheuristic algorithms. Surveyed work is categorized based on application scenarios and solution encoding, and describes these algorithms in detail to help researchers choose appropriate methods to solve related applications. It is seen that transfer function is the main binary coding of metaheuristic algorithms, which usually adopts Sigmoid function. Among the contributions presented, there were different implementations and applications of metaheuristic algorithms, or the study of engineering applications by different objective functions such as the single- and multi-objective problems of feature selection, scheduling, layout and engineering structure optimization. The article identifies current troubles and challenges by the conducted review, and discusses that novel binary algorithm, transfer function, benchmark function, time-consuming problem and application integration are need to be resolved in future.

4 citations

Journal ArticleDOI
01 Dec 2015-Optik
TL;DR: Experimental results prove that the proposed L1-norm plus L2-norm sparse parameter (L1L2-SP) classifier achieves better recognition rate than the SRC classifiers, SFR classifier, and several other classifiers.

4 citations

Proceedings ArticleDOI
20 Jun 2010
TL;DR: The statistical results from simulating data indicate the validity of modeling approach and the effects of three parameters on overhead and efficiency of system running come out in compared analysis both of different size swarms and of specific size swarm.
Abstract: As one of valid modeling and coordinated controlling tools, the particle swarm optimization algorithm can be extended for applying to task of swarm robotic search. To gain an insight into effects of key parameters on distributed swarm search, a series of simulations are conducted. In such control algorithm, main parameters including communication range, detection radius, and swarm size are paid close attentions. Similar to ideal “particles”, swarm robots are modeled at an abstract level with the extended particle swarm optimization method. Also, individual robots are designed to be controlled under three-state finite state machine mechanism having SingleSearch, SwarmSearch, and Declaration states. Then, control strategy and algorithm are developed. The statistical results from simulating data indicate the validity of modeling approach. Further, the effects of three parameters on overhead and efficiency of system running come out in compared analysis both of different size swarms and of specific size swarm.

4 citations


Cited by
More filters
Journal ArticleDOI
TL;DR: It is proved the convergence of a recursive mean shift procedure to the nearest stationary point of the underlying density function and, thus, its utility in detecting the modes of the density.
Abstract: A general non-parametric technique is proposed for the analysis of a complex multimodal feature space and to delineate arbitrarily shaped clusters in it. The basic computational module of the technique is an old pattern recognition procedure: the mean shift. For discrete data, we prove the convergence of a recursive mean shift procedure to the nearest stationary point of the underlying density function and, thus, its utility in detecting the modes of the density. The relation of the mean shift procedure to the Nadaraya-Watson estimator from kernel regression and the robust M-estimators; of location is also established. Algorithms for two low-level vision tasks discontinuity-preserving smoothing and image segmentation - are described as applications. In these algorithms, the only user-set parameter is the resolution of the analysis, and either gray-level or color images are accepted as input. Extensive experimental results illustrate their excellent performance.

11,727 citations

Book
24 Oct 2001
TL;DR: Digital Watermarking covers the crucial research findings in the field and explains the principles underlying digital watermarking technologies, describes the requirements that have given rise to them, and discusses the diverse ends to which these technologies are being applied.
Abstract: Digital watermarking is a key ingredient to copyright protection. It provides a solution to illegal copying of digital material and has many other useful applications such as broadcast monitoring and the recording of electronic transactions. Now, for the first time, there is a book that focuses exclusively on this exciting technology. Digital Watermarking covers the crucial research findings in the field: it explains the principles underlying digital watermarking technologies, describes the requirements that have given rise to them, and discusses the diverse ends to which these technologies are being applied. As a result, additional groundwork is laid for future developments in this field, helping the reader understand and anticipate new approaches and applications.

2,849 citations

Proceedings Article
01 Jan 1999

2,010 citations

Posted Content
TL;DR: This paper defines and explores proofs of retrievability (PORs), a POR scheme that enables an archive or back-up service to produce a concise proof that a user can retrieve a target file F, that is, that the archive retains and reliably transmits file data sufficient for the user to recover F in its entirety.
Abstract: In this paper, we define and explore proofs of retrievability (PORs). A POR scheme enables an archive or back-up service (prover) to produce a concise proof that a user (verifier) can retrieve a target file F, that is, that the archive retains and reliably transmits file data sufficient for the user to recover F in its entirety.A POR may be viewed as a kind of cryptographic proof of knowledge (POK), but one specially designed to handle a large file (or bitstring) F. We explore POR protocols here in which the communication costs, number of memory accesses for the prover, and storage requirements of the user (verifier) are small parameters essentially independent of the length of F. In addition to proposing new, practical POR constructions, we explore implementation considerations and optimizations that bear on previously explored, related schemes.In a POR, unlike a POK, neither the prover nor the verifier need actually have knowledge of F. PORs give rise to a new and unusual security definition whose formulation is another contribution of our work.We view PORs as an important tool for semi-trusted online archives. Existing cryptographic techniques help users ensure the privacy and integrity of files they retrieve. It is also natural, however, for users to want to verify that archives do not delete or modify files prior to retrieval. The goal of a POR is to accomplish these checks without users having to download the files themselves. A POR can also provide quality-of-service guarantees, i.e., show that a file is retrievable within a certain time bound.

1,783 citations