scispace - formally typeset
Search or ask a question

Showing papers on "JPEG 2000 published in 2021"


Journal ArticleDOI
Tong Chen1, Haojie Liu1, Zhan Ma1, Qiu Shen1, Xun Cao1, Yao Wang2 
TL;DR: An end-to-end learnt lossy image compression approach, which is built on top of the deep nerual network (DNN)-based variational auto-encoder (VAE) structure with Non-Local Attention optimization and Improved Context modeling (NLAIC).
Abstract: This article proposes an end-to-end learnt lossy image compression approach, which is built on top of the deep nerual network (DNN)-based variational auto-encoder (VAE) structure with Non-Local Attention optimization and Improved Context modeling (NLAIC). Our NLAIC 1) embeds non-local network operations as non-linear transforms in both main and hyper coders for deriving respective latent features and hyperpriors by exploiting both local and global correlations, 2) applies attention mechanism to generate implicit masks that are used to weigh the features for adaptive bit allocation, and 3) implements the improved conditional entropy modeling of latent features using joint 3D convolutional neural network (CNN)-based autoregressive contexts and hyperpriors. Towards the practical application, additional enhancements are also introduced to speed up the computational processing (e.g., parallel 3D CNN-based context prediction), decrease the memory consumption (e.g., sparse non-local processing) and reduce the implementation complexity (e.g., a unified model for variable rates without re-training). The proposed model outperforms existing learnt and conventional (e.g., BPG, JPEG2000, JPEG) image compression methods, on both Kodak and Tecnick datasets with the state-of-the-art compression efficiency, for both PSNR and MS-SSIM quality measurements. We have made all materials publicly accessible at https://njuvision.github.io/NIC for reproducible research.

142 citations


Journal ArticleDOI
TL;DR: A novel lossy-to-lossless data compression scheme with a compression throughput well above 4 GB/s and compression rates and rate-distortion curves competitive with those achieved by JPEG2000 and JP3D is presented.
Abstract: The rapid increase in medical and biomedical image acquisition rates has opened up new avenues for image analysis, but has also introduced formidable challenges. This is evident, for example, in selective plane illumination microscopy where acquisition rates of about 1–4 GB/s sustained over several days have redefined the scale of I/O bandwidth required by image analysis tools. Although the effective bandwidth could, principally, be increased by lossy-to-lossless data compression, this is of limited value in practice due to the high computational demand of current schemes such as JPEG2000 that reach compression throughput of one order of magnitude below that of image acquisition. Here we present a novel lossy-to-lossless data compression scheme with a compression throughput well above 4 GB/s and compression rates and rate-distortion curves competitive with those achieved by JPEG2000 and JP3D.

15 citations


Journal ArticleDOI
TL;DR: The main aim of this article is to analyse the most advanced lossless image compression algorithms from each point of view, and evaluate the strength of each algorithm for each kind of image.
Abstract: A great deal of information is produced daily, due to advances in telecommunication, and the issue of storing it on digital devices or transmitting it over the Internet is challenging Data compression is essential in managing this information well Therefore, research on data compression has become a topic of great interest to researchers, and the number of applications in this area is increasing Over the last few decades, international organisations have developed many strategies for data compression, and there is no specific algorithm that works well on all types of data The compression ratio, as well as encoding and decoding times, are mainly used to evaluate an algorithm for lossless image compression However, although the compression ratio is more significant for some applications, others may require higher encoding or decoding speeds or both; alternatively, all three parameters may be equally important The main aim of this article is to analyse the most advanced lossless image compression algorithms from each point of view, and evaluate the strength of each algorithm for each kind of image We develop a technique regarding how to evaluate an image compression algorithm that is based on more than one parameter The findings that are presented in this paper may be helpful to new researchers and to users in this area

12 citations


Journal ArticleDOI
TL;DR: Zhang et al. as mentioned in this paper proposed a semantics-to-signal scalable image compression method, where partial bitstreams are decodeable for machine vision and the entire bitstream is decodeable by human vision.
Abstract: Image/video compression and communication need to serve both human vision and machine vision. To address this need, we propose a scalable image compression solution. We assume that machine vision needs less information that is related to semantics, whereas human vision needs more information that is to reconstruct signal. We then propose semantics-to-signal scalable compression, where partial bitstream is decodeable for machine vision and the entire bitstream is decodeable for human vision. Our method is inspired by the scalable image coding standard, JPEG2000, and similarly adopts subband-wise representations. We first design a trainable and revertible transform based on the lifting structure, which converts an image into a pyramid of multiple subbands; the transform is trained to make the partial representations useful for multiple machine vision tasks. We then design an end-to-end optimized encoding/decoding network for compressing the multiple subbands, to jointly optimize compression ratio, semantic analysis accuracy, and signal reconstruction quality. We experiment with two datasets: CUB200-2011 and FGVC-Aircraft, taking coarse-to-fine image classification tasks as an example. Experimental results demonstrate that our proposed method achieves semantics-to-signal scalable compression, and outperforms JPEG2000 in compression efficiency. The proposed method sheds light on a generic approach for image/video coding for human and machines.

11 citations


Journal ArticleDOI
TL;DR: This work aims to study the quality evaluation of digital hologram reconstructions presented on regular 2D displays in the presence of compression distortions, and a set of state-of-the-art compression codecs were used for compression of the digital holograms on the object plane.
Abstract: Recently, more interest in the different plenoptic formats, including digital holograms, has emerged. Aside from other challenges that several steps of the holographic pipeline, from digital acquisition to display, have to face, visual quality assessment of compressed holograms is particularly demanding due to the distinct nature of this 3D image modality when compared to regular 2D imaging. There are few studies on holographic data quality assessment, particularly with respect to the perceptual effects of lossy compression. This work aims to study the quality evaluation of digital hologram reconstructions presented on regular 2D displays in the presence of compression distortions. As there is no established or generally agreed on compression methodology for digital hologram compression on the hologram plane with available implementations, a set of state-of-the-art compression codecs, namely, HEVC, AV1, and JPEG2000, were used for compression of the digital holograms on the object plane. Both computer-generated and optically generated holograms were considered. Two subjective tests were conducted to evaluate distortions caused by compression. The first subjective test was conducted on the reconstructed amplitude images of central views, while the second test was conducted on pseudovideos generated from the reconstructed amplitudes of different views. The subjective quality assessment was based on mean opinion scores. A selection of objective quality metrics was evaluated, and their correlations with mean opinion scores were computed. The VIFp metrics appeared to have the highest correlation.

10 citations


Journal ArticleDOI
TL;DR: In this paper, two auxiliary adversarial networks are incorporated to make the intermediate grayscale images and final restored color images indistinguishable from normal gray-scale and color images, and the JPEG simulator is utilized to simulate real JPEG compression during the online training so that the hiding and restoring sub-networks can automatically learn to be JPEG robust.
Abstract: Invertible grayscale is a special kind of grayscale from which the original color can be recovered. Given an input color image, this seminal work tries to hide the color information into its grayscale counterpart while making it hard to recognize any anomalies. This powerful functionality is enabled by training a hiding sub-network and restoring sub-network in an end-to-end way. Despite its expressive results, two key limitations exist: 1) The restored color image often suffers from some noticeable visual artifacts in the smooth regions. 2) It is very sensitive to JPEG compression, i.e., the original color information cannot be well recovered once the intermediate grayscale image is compressed by JPEG. To overcome these two limitations, this paper introduces adversarial training and JPEG simulator respectively. Specifically, two auxiliary adversarial networks are incorporated to make the intermediate grayscale images and final restored color images indistinguishable from normal grayscale and color images. And the JPEG simulator is utilized to simulate real JPEG compression during the online training so that the hiding and restoring sub-networks can automatically learn to be JPEG robust. Extensive experiments demonstrate that the proposed method is superior to the original invertible grayscale work both qualitatively and quantitatively while ensuring the JPEG robustness. We further show that the proposed framework can be applied under different types of grayscale constraints and achieve excellent results.

9 citations


Journal ArticleDOI
TL;DR: A novel approach to achieve scene classification in Joint Photographic Experts Group (JPEG) 2000 compressed RS images by models the multiresolution paradigm given in the JPEG 2000 compression algorithm in an end-to-end trainable unified neural network.
Abstract: To reduce the storage requirements, remote-sensing (RS) images are usually stored in compressed format. Existing scene classification approaches using deep neural networks (DNNs) require to fully decompress the images, which is a computationally demanding task in operational applications. To address this issue, in this article, we propose a novel approach to achieve scene classification in Joint Photographic Experts Group (JPEG) 2000 compressed RS images. The proposed approach consists of two main steps: 1) approximation of the finer resolution subbands of reversible biorthogonal wavelet filters used in JPEG 2000 and 2) characterization of the high-level semantic content of approximated wavelet subbands and scene classification based on the learned descriptors. This is achieved by taking codestreams associated with the coarsest resolution wavelet subband as input to approximate finer resolution subbands using a number of transposed convolutional layers. Then, a series of convolutional layers models the high-level semantic content of the approximated wavelet subband. Thus, the proposed approach models the multiresolution paradigm given in the JPEG 2000 compression algorithm in an end-to-end trainable unified neural network. In the classification stage, the proposed approach takes only the coarsest resolution wavelet subbands as input, thereby reducing the time required to apply decoding. Experimental results performed on two benchmark aerial image archives demonstrate that the proposed approach significantly reduces the computational time with similar classification accuracies when compared with traditional RS scene classification approaches (which requires full image decompression).

8 citations


Proceedings ArticleDOI
02 Apr 2021
TL;DR: In this article, the output quality of the image is analyzed using quality metrics like PSNR (Peak Signal-to-Noise Ratio) and SSIM (Structural Similarity Index Metric).
Abstract: Online videos and image blogging are some of the major sources for internet traffic and are estimated to have exponential growth in the future. Efforts are made to introduce advanced compression formats for images and included in the future codecs. This paper focuses on analyzing and comparing the existing codecs with the potential codecs to obtain very good quality images with substantially high compression rate. Using the same test conditions and same test content-set for both image and video codecs, like JPEG, JPEG LS, JPEG XR, JPEG 2000, JPEG XT, HEVC, VVC and EVC, the output quality of the image is analyzed using quality metrics like PSNR (Peak Signal-to-Noise Ratio) and SSIM (Structural Similarity Index Metric). Fitting properties of each codec, conditions of testing and performance comparison between the codecs are analyzed. The test results indicate that JPEG XR is best suited for low resolution and HD image compressions while JPEG 2000 and JPEG XT are best suited for grey-scale and 4K image compressions respectively. The paper concludes with detailed analysis of codec performance including suggestion over its applications and scope for future work.

8 citations


Journal ArticleDOI
TL;DR: MAGIC, a novel machine learning (ML)-guided image compression framework that judiciously sacrifices the visual quality to achieve much higher compression when compared to traditional techniques, while maintaining accuracy for coarse-grained vision tasks is proposed.
Abstract: The emergent ecosystems of intelligent edge devices in diverse Internet-of-Things (IoT) applications, from automatic surveillance to precision agriculture, increasingly rely on recording and processing a variety of image data. Due to resource constraints, e.g., energy and communication bandwidth requirements, these applications require compressing the recorded images before transmission. For these applications, image compression commonly requires: 1) maintaining features for coarse-grain pattern recognition instead of the high-level details for human perception due to machine-to-machine communications; 2) high compression ratio that leads to improved energy and transmission efficiency; and 3) large dynamic range of compression and an easy tradeoff between compression factor and quality of reconstruction to accommodate a wide diversity of IoT applications as well as their time-varying energy/performance needs. To address these requirements, we propose, MAGIC, a novel machine learning (ML)-guided image compression framework that judiciously sacrifices the visual quality to achieve much higher compression when compared to traditional techniques, while maintaining accuracy for coarse-grained vision tasks. The central idea is to capture application-specific domain knowledge and efficiently utilize it in achieving high compression. We demonstrate that the MAGIC framework is configurable across a wide range of compression/quality and is capable of compressing beyond the standard quality factor limits of both JPEG 2000 and WebP. We perform experiments on representative IoT applications using two vision data sets and show $42.65\times $ compression at similar accuracy with respect to the source. We highlight low variance in compression rate across images using our technique as compared to JPEG 2000 and WebP.

7 citations


Journal ArticleDOI
TL;DR: Using the concept of the Guided Grad-CAM (Gradient-weighted Class Activation Mapping) technique to produce heat maps with the help of a heat map generator trained on ResNet-50, a saliency-guided encoding–decoding algorithm is developed, in which a wider multi-scale saliency guided convolutional neural network is designed.

7 citations


Journal ArticleDOI
TL;DR: These joint models provide image compression tailored for the specific task of 3D reconstruction, and show that this can be done highly efficiently at almost no additional cost to obtain compression on top of the computation already required for performing the3D reconstruction task.

Journal ArticleDOI
TL;DR: Wang et al. as mentioned in this paper employed the big data mining to set up the image codebook and proposed a soft compression algorithm for multi-component medical images, which can exactly reflect the fundamental structure of images.
Abstract: In disease diagnosis, medical image plays an important part. Its lossless compression is pretty critical, which directly determines the requirement of local storage space and communication bandwidth of remote medical systems, so as to help the diagnosis and treatment of patients. There are two extraordinary properties related to medical images: lossless and similarity. How to take advantage of these two properties to reduce the information needed to represent an image is the key point of compression. In this paper, we employ the big data mining to set up the image codebook. That is, to find the basic components of images. We propose a soft compression algorithm for multi-component medical images, which can exactly reflect the fundamental structure of images. A general representation framework for image compression is also put forward and the results indicate that our developed soft compression algorithm can outperform the popular benchmarks PNG and JPEG2000 in terms of compression ratio.

Journal ArticleDOI
TL;DR: In this article, the authors proposed a method to estimate the upscaling factors on double JPEG compressed images in the presence of image up-scaling between the two compressions by analyzing the spectrum of scaled images and giving an overall formulation of how the scaling factors along with the parameters of JPEG compression and image contents influence the appearance of tampering artifacts.
Abstract: As one of the most important topics in image forensics, resampling detection has developed rapidly in recent years. However, the robustness to JPEG compression is still challenging for most classical spectrum-based methods, since JPEG compression severely degrades the image contents and introduces block artifacts in the boundary of the compression grid. In this article, we propose a method to estimate the upscaling factors on double JPEG compressed images in the presence of image upscaling between the two compressions. We first analyze the spectrum of scaled images and give an overall formulation of how the scaling factors along with the parameters of JPEG compression and image contents influence the appearance of tampering artifacts. The expected positions of five kinds of characteristic peaks are analytically derived. Then, we analyze the features of double JPEG compressed images in the block discrete cosine transform (BDCT) domain and present an inverse scaling strategy for the upscaling factor estimation with a detailed proof. Finally, a fusion method is proposed that through frequency-domain analysis, a candidate set of upscaling factors is given, and through analysis in the BDCT domain, the optimal estimation from all candidates is determined. The experimental results demonstrate that the proposed method outperforms other state-of-the-art methods.

Journal ArticleDOI
05 Jan 2021-Symmetry
TL;DR: This study evaluates the coding performance of various projection formats, including recently-proposed formats adapting to the input of JPEG and JPEG 2000 content, and proposes an evaluation framework for reducing the bias toward the native equi-rectangular (ERP) format.
Abstract: Recently, 360° content has emerged as a new method for offering real-life interaction. Ultra-high resolution 360° content is mapped to the two-dimensional plane to adjust to the input of existing generic coding standards for transmission. Many formats have been proposed, and tremendous work is being done to investigate 360° videos in the Joint Video Exploration Team using projection-based coding. However, the standardization activities for quality assessment of 360° images are limited. In this study, we evaluate the coding performance of various projection formats, including recently-proposed formats adapting to the input of JPEG and JPEG 2000 content. We present an overview of the nine state-of-the-art formats considered in the evaluation. We also propose an evaluation framework for reducing the bias toward the native equi-rectangular (ERP) format. We consider the downsampled ERP image as the ground truth image. Firstly, format conversions are applied to the ERP image. Secondly, each converted image is subjected to the JPEG and JPEG 2000 image coding standards, then decoded and converted back to the downsampled ERP to find the coding gain of each format. The quality metrics designed for 360° content and conventional 2D metrics have been used for both end-to-end distortion measurement and codec level, in two subsampling modes, i.e., YUV (4:2:0 and 4:4:4). Our evaluation results prove that the hybrid equi-angular format and equatorial cylindrical format achieve better coding performance among the compared formats. Our work presents evidence to find the coding gain of these formats over ERP, which is useful for identifying the best image format for a future standard.

Journal ArticleDOI
TL;DR: Results obtained from challenging cross-database experiments in which the analyzed retouching technique is unknown during training yield interesting findings, including the finding that in some cases, the application of image compression might as well improve detection performance.
Abstract: In the past years, numerous methods have been introduced to reliably detect digital face image manipulations. Lately, the generalizability of these schemes has been questioned in particular with respect to image post-processing. Image compression represents a post-processing which is frequently applied in diverse biometric application scenarios. Severe compression might erase digital traces of face image manipulation and hence hamper a reliable detection thereof. In this work, the effects of image compression on face image manipulation detection are analyzed. In particular, a case study on facial retouching detection under the influence of image compression is presented. To this end, ICAO-compliant subsets of two public face databases are used to automatically create a database containing more than 9,000 retouched reference images together with unconstrained probe images. Subsequently, reference images are compressed applying JPEG and JPEG 2000 at compression levels recommended for face image storage in electronic travel documents. Novel detection algorithms utilizing texture descriptors and deep face representations are proposed and evaluated in a single image and differential scenario. Results obtained from challenging cross-database experiments in which the analyzed retouching technique is unknown during training yield interesting findings: (1) most competitive detection performance is achieved for differential scenarios employing deep face representations; (2) image compression severely impacts the performance of face image manipulation detection schemes based on texture descriptors while methods utilizing deep face representations are found to be highly robust; (3) in some cases, the application of image compression might as well improve detection performance.

Proceedings ArticleDOI
14 Jun 2021
TL;DR: In this article, a subset of 8 holograms, selected from the JPEG Pleno Database, is compressed both in hologram and object plane at three different bit rates, chosen bit rates vary per hologram depending on their content characteristics.
Abstract: In preparation of the Call for Proposals on JPEG Pleno Holography, multiple exploration studies are ongoing to define the general procedure for performance assessment of proposed coding solutions. The performance of proposals will be compared against JPEG 2000 and H.265/HEVC intra coding, both serving as anchor codecs. In this paper, we report on the results of a dynamic subjective visual quality assessment procedure. A subset of 8 holograms, selected from the JPEG Pleno Database, is compressed both in hologram and object plane at three different bit rates. Chosen bit rates vary per hologram depending on their content characteristics. Then, for each hologram and its compressed versions, pseudo-video sequences are generated from the reconstructed views along a scan path that involves focus and viewing angle changes and that is hologram specific. A double stimulus simultaneous test, combined with a 5-level impairment-scale scoring protocol is deployed where videos created from the reference and the decoded holograms are visualized side by side on a professional 4K display. Results demonstrate that this test procedure requires deep scenes and sufficient scene complexity throughout the depth stack to allow for adequate stress testing of the codecs under test, particularly those solutions that adhere to compression in the object plane.

Journal ArticleDOI
TL;DR: In this article, an objective and subjective assessment of HEIC compression method on dermatological color images and benchmarking the performance of High-Efficiency Image Coding (HEIC) with different algorithms to a feasibility study of the method for teledermatology.
Abstract: BACKGROUND Currently, teledermatology assumes a progressively greater role in the modern healthcare system, especially in consultation, diagnosis, or examining lesions and skin cancers. One of the major challenges facing teledermatology systems is determining the optimal image compression method to efficiently reduce the space needed for electronic storage and data transmission. OBJECTIVE To the objective and subjective assessment of HEIC compression method on dermatological color images and benchmarking the performance of High-Efficiency Image Coding (HEIC) with different algorithms to a feasibility study of the method for teledermatology. METHODS Twenty-five clinical and five skin histopathology images were taken in department of dermatology, Imam Reza Hospital, Mashhad, Iran. For each image, a set of 24 compressed images with different compression rates, which is composed of eight JPEG, eight JPEG2000, and eight HEIC images, has been prepared. Compressed and original images were shown simultaneously to three dermatologists and one dermatopathologist with different experiences. Each dermatologist scored quality and suitability of compressed images for diagnostic, as well as educational/scientific purposes. An objective evaluation was performed by calculating the mean "distance" of pixel colors and peak signal-to-noise ratio (PSNR). RESULTS All compression rates for HEIC were objectively better than JPEG and JPEG2000, particularly at PSNR. Moreover, mean "color distance" per pixel for compressed images using HEIC was lower than others. The subjective image quality assessment also confirms the results of objective evaluation. In both educational and clinical diagnostic applications, HEIC compressed images have the highest score. CONCLUSION In consideration of objective and subjective evaluation, the HEIC algorithm represents an optimal performance in dermatology images compression compared with JPEG and JPEG2000.

Journal ArticleDOI
TL;DR: Experimental results show that the Gaussian Mixture Model constrained by Markov Random Field framework for image compression performs better than the previous work, HEVC, JPEG 2000 and Better Portable Graphics (BPG) which is an improved version of HEVC.

Journal ArticleDOI
TL;DR: An end-to-end approach using Convolutional Neural Networks is presented to classify images into six categories of bad lighting, Gaussian blur, motion blur, JPEG 2000, white-noise, and high quality reference images that can be easily re-calibrated for other applications with only a small sample of training images.
Abstract: The detection of poor quality images for reasons such as focus, lighting, compression, and encoding is of great importance in the field of computer vision. The ability to quickly and automatically classify an image as poor quality creates opportunities for a multitude of applications such as digital cameras, phones, self-driving cars, and web search technologies. In this paper an end-to-end approach using Convolutional Neural Networks (CNN) is presented to classify images into six categories of bad lighting, Gaussian blur, motion blur, JPEG 2000, white-noise, and high quality reference images. A new dataset of images was produced and used to train and validate the model. Finally, the application of the developed model was evaluated using images from the German Traffic Sign Recognition Benchmark. The results show that the trained CNN can detect and correctly classify images into the aforementioned categories with high accuracy and the model can be easily re-calibrated for other applications with only a small sample of training images.

Journal ArticleDOI
TL;DR: In this paper, a semi-fragile watermarking scheme for JPEG2000 image self-authentication is proposed to ensure the robustness of the watermark against the compression attacks generated by the JPEG2000 encoder itself.
Abstract: The increasing performances of personal computers, as well as software of image processing, enable the easy manipulation of digital media content. Unfortunately, this easy manipulation makes the detection of changes in the multimedia content a very difficult task. Digital watermarking is among the most appropriate techniques to verify the integrity of multimedia content. In this paper, we propose a semi-fragile watermarking scheme for JPEG2000 image self-authentication that ensuring the robustness of the watermark against the compression attacks generated by the JPEG2000 encoder itself. The scheme is combined with the JPEG2000 encoder by embedding the generated watermark into the host image during the JPEG2000 compression process. To generate the watermark we suppose to use a perceptual hash function (PHF) operating on discrete wavelet coefficients of the host image. The proposed watermark generation process leads the system to verify the integrity of the image without the need to any file except for the watermarked image. The watermark is embedded during the JPEG2000 compression process after Discrete Wavelet Transform (DWT) step into the approximation sub-band coefficients of the five wavelet decomposition using Index Modulation Quantification (QIM), and can be extracted during image decoding. To prove the authentication of the image, the system compares the extracted watermark with the new watermark generated from the received image. Experimental results show that our proposed approach has not only an extremely high accuracy of tampering detection but also a relatively very high resistance against JPEG2000 compression attacks.

Journal ArticleDOI
TL;DR: Results show clear superiority of the proposed scheme over the conventional separate compression approach involving two codecs: JPEG-2000 for images and ECG SPIHT-1D as well as other competing multimodal compression schemes in terms of both PRD and SNR at the signal decompression stage while maintaining good image quality and exhibiting a reduced computational complexity.
Abstract: In this paper, a wavelet-based multimodal compression method is proposed. The method jointly compresses a medical image and an ECG signal within a single codec, i.e., JPEG-2000 in an effective and simple way. The multimodal scheme operates in two main stages: the first stage, consists of the encoder and involves a mixing function, aiming at inserting the samples of the signal in the image according to a predefined insertion pattern in the wavelet domain. The second stage represented by a separation function, consists of the extraction process of the ECG signal from the image after performing the decoding stage. Both the cubic spline and the median edge detection (MED) predictor have been adopted to conduct the interpolation process for estimating image pixels. Intensive experiments have been conducted to evaluate the performance of the multimodal scheme using objective distortion criteria. Results show clear superiority of the proposed scheme over the conventional separate compression approach involving two codecs: JPEG-2000 for images and ECG SPIHT-1D as well as other competing multimodal compression schemes in terms of both PRD and SNR at the signal decompression stage while maintaining good image quality and exhibiting a reduced computational complexity. Improvements in terms of average PRD and SNR values are as significant as 0.7 and 6 dB at low bit rates and 0.06 and 2 dB at higher bit rates on a number of test ECG signals and medical images.

Journal ArticleDOI
Yi Qin1, Yuhong Wan1, Shujia Wan, Chao Liu1, Wei Liu 
TL;DR: Wang et al. as discussed by the authors proposed a simple and universal scheme to compress and decompress the ciphertext of an optical cryptosystem by the aid of deep learning (DL), which can achieve a far smaller size of the compressed file (SCF) while offering comparable decompression quality.
Abstract: The compression of the ciphertext of a cryptosystem is desirable considering the dramatic increase in secure data transfer via Internet. In this paper, we propose a simple and universal scheme to compress and decompress the ciphertext of an optical cryptosystem by the aid of deep learning (DL). For compression, the ciphertext is first resized to a relatively small dimension by bilinear interpolation and thereafter condensed by the JPEG2000 standard. For decompression, a well-trained deep neural network (DNN) can be employed to perfectly recover the original ciphertext, in spite of the severe information loss suffered by the compressed file. In contrast with JPEG2000 and JPEG, our proposal can achieve a far smaller size of the compressed file (SCF) while offering comparable decompression quality. In addition, the SCF can be further reduced by compromising the quality of the recovered plaintext. It is also shown that the compression procedure can provide an additional security level, and this may offer new insight into the compressive encryption in optical cryptosystems. Both simulation and experimental results are presented to demonstrate the proposal.

Journal ArticleDOI
TL;DR: An optimized JPEG-XT method (OPT_JPEG-XT) is represented that better compresses 16-bit depth medical images by amplifying (N times) discrete cosine transform (DCT) coefficients to realize lossless compression of medical images.

Journal ArticleDOI
TL;DR: An enhanced binary MQ arithmetic coder to make use of look-up table (LUT) for (A × Qe) using quantization skill to improve the coding efficiency is proposed in this paper.
Abstract: Binary MQ arithmetic coding is widely used as a basic entropy coder in multimedia coding system. MQ coder esteems high in compression efficiency to be used in JBIG2 and JPEG2000. The importance of arithmetic coding is increasing after it is adopted as a unique entropy coder in HEVC standard. In the binary MQ coder, arithmetic approximation without multiplication is used in the process of recursive subdivision of range interval. Because of the MPS/LPS exchange activity that happens in the MQ coder, the output byte tends to increase. This paper proposes an enhanced binary MQ arithmetic coder to make use of look-up table (LUT) for (A × Qe) using quantization skill to improve the coding efficiency. Multi-level quantization using 2-level, 4-level and 8-level look-up tables is proposed in this paper. Experimental results applying to binary documents show about 3% improvement for basic context-free binary arithmetic coding. In the case of JBIG2 bi-level image compression standard, compression efficiency improved about 0.9%. In addition, in the case of lossless JPEG2000 compression, compressed byte decreases 1.5% using 8-level LUT. For the lossy JPEG2000 coding, this figure is a little lower, about 0.3% improvement of PSNR at the same rate.

Journal ArticleDOI
TL;DR: Improved Joint Source Channel (JSC) decoding using turbo codes is employed that are unified into JPEG 2000 decoder architecture that provides reduced BER and complexity.
Abstract: In this paper, improved Joint Source Channel (JSC) decoding using turbo codes is employed that are unified into JPEG 2000 decoder architecture. Rather than using conventional decoding approach, the proposed system provides reduced BER and complexity. This work deliberates Soft Output Viterbi Algorithm (SOVA) for its real time implementation features. Nevertheless, the algorithm suffers from performance degradation due to optimistic and correlation effects. The objective of this work is to enhance the performance of SOVA by using appropriate reduction factors that integrates and eliminates both these distortions. Acquired Integrated Factor is incorporated in the turbo decoders of the image transmission system. Bit Error rate and Peak Signal to Noise Ratio (PSNR) analysis is made by considering Additive White Gaussian Noise and Rayleigh fading channel through MATLAB simulation. To illustrate the visual quality improvement of the proposed scheme, comparison with other FEC codes is done with significant PSNR gains.

Journal ArticleDOI
24 Sep 2021
TL;DR: A new improved IVW algorithm for copyright protection that can deliver additional information to the image content is presented and overcomes several drawbacks reported in previous algorithms, including geometric and image processing attacks such as JPEG and JPEG2000.
Abstract: Digital image watermarking algorithms have been designed for intellectual property, copyright protection, medical data management, and other related fields; furthermore, in real-world applications such as official documents, banknotes, etc., they are used to deliver additional information about the documents’ authenticity. In this context, the imperceptible–visible watermarking (IVW) algorithm has been designed as a digital reproduction of the real-world watermarks. This paper presents a new improved IVW algorithm for copyright protection that can deliver additional information to the image content. The proposed algorithm is divided into two stages: in the embedding stage, a human visual system-based strategy is used to embed an owner logotype or a 2D quick response (QR) code as a watermark into a color image, maintaining a high watermark imperceptibility and low image-quality degradation. In the exhibition, a new histogram binarization function approach is introduced to exhibit any watermark with enough quality to be recognized or decoded by any application, which is focused on reading QR codes. The experimental results show that the proposed algorithm can embed one or more watermark patterns, maintaining the high imperceptibility and visual quality of the embedded and the exhibited watermark. The performance evaluation shows that the method overcomes several drawbacks reported in previous algorithms, including geometric and image processing attacks such as JPEG and JPEG2000.

Proceedings ArticleDOI
10 Jan 2021
TL;DR: In this paper, the authors proposed a lossless compression method that integrates a Recurrent Neural Network (RNN) as a 3D sequence prediction model, which learns the long dependencies of the voxel's neighbourhood in 3D using Long Short-Term Memory (LSTM) network then compress the residual error using arithmetic coding.
Abstract: As scanners produce higher-resolution and more densely sampled images, this raises the challenge of data storage, transmission and communication within healthcare systems. Since the quality of medical images plays a crucial role in diagnosis accuracy, medical imaging compression techniques are desired to reduce scan bitrate while guaranteeing lossless reconstruction. This paper presents a lossless compression method that integrates a Recurrent Neural Network (RNN) as a 3D sequence prediction model. The aim is to learn the long dependencies of the voxel's neighbourhood in 3D using Long Short-Term Memory (LSTM) network then compress the residual error using arithmetic coding. Experiential results reveal that our method obtains a higher compression ratio achieving 15% saving compared to the state-of-the-art lossless compression standards, including JPEG-LS, JPEG2000, JP3D, HEVC, and PPMd. Our evaluation demonstrates that the proposed method generalizes well to unseen modalities CT and MRI for the lossless compression scheme. To the best of our knowledge, this is the first lossless compression method that uses LSTM neural network for 16-bit volumetric medical image compression.

Journal ArticleDOI
TL;DR: In this paper, a variable-size 2D-block extraction and encoding method with built-in bi-level coding for color image is developed to decrease the entropy of information and improve the compression ratio.
Abstract: With IoT development, it becomes more popular that image data is transmitted via wireless communication systems. If bit errors occur during transmission, the recovered image will become useless. To solve this problem, a bit-error aware lossless image compression based on bi-level coding is proposed for gray image compression. But bi-level coding has not considered the inherent statistical correlation in 2D context region. To resolve this shortage, a novel variable-size 2D-block extraction and encoding method with built-in bi-level coding for color image is developed to decrease the entropy of information and improve the compression ratio. A lossless color transformation from RGB to the YCrCb color space is used for the decorrelation of color components. Particularly, the layer-extraction method is proposed to keep the Laplacian distribution of the data in 2D blocks which is suitable for bi-level coding. In addition, optimization of 2D-block start bits is used to improve the performance. To evaluate the performance of our proposed method, many experiments including the comparison with state-of-the-art methods, the effects with different color space, etc. are conducted. The comparison experiments under a bit-error environment show that the average compression rate of our method is better than bi-level, Jpeg2000, WebP, FLIF, and L3C (deep learning method) with hamming code. Also, our method achieves the same image quality with the bi-level method. Other experiments illustrate the positive effect of built-in bi-level encoding and encoding with zero-mean values, which can maintain high image quality. At last, the results of the decrease of entropy and the procedure of our method are given and discussed.

Journal ArticleDOI
TL;DR: In this paper, a generative adversarial network (GAN) based prediction method called MultiTempGAN was proposed for multispectral (MS) image compression, where the generator parameters are saved for the reconstruction purpose in the receiver system.

Journal ArticleDOI
TL;DR: This proposed method of compression can be used efficiently for the medical image in order to store and retrieve in healthcare industry.
Abstract: Medical images are generated in a huge number in the research centers and hospitals every day. Working with the medical images and maintain the storage needs an efficient fund and huge storage space. Retaining the quality of the medical image is also very essential. Image compression without losing its quality is the only term to achieve the desired task. Achieving the desired task using the integer Karhunen Loeve transform attains a quality output and also with less storage space. JPEG and JPEG 2000 are also challenging to the integer transform based compression. Resulting the compression quality in terms of peak signal noise ratio, compression ratio is attained. Proposed method of compression is compared with the other efficient algorithms. Thus this proposed method can be used efficiently for the medical image in order to store and retrieve in healthcare industry.