A scheme for image compression that takes into account psychovisual features both in the space and frequency domains is proposed. This method involves two steps. First, a wavelet transform used in order to obtain a set of biorthogonal subclasses of images: the original image is decomposed at different scales using a pyramidal algorithm architecture. The decomposition is along the vertical and horizontal directions and maintains constant the number of pixels required to describe the image. Second, according to Shannon's rate distortion theory, the wavelet coefficients are vector quantized using a multiresolution codebook. To encode the wavelet coefficients, a noise shaping bit allocation procedure which assumes that details at high resolution are less visible to the human eye is proposed. In order to allow the receiver to recognize a picture as quickly as possible at minimum cost, a progressive transmission scheme is presented. It is shown that the wavelet transform is particularly well adapted to progressive transmission.
Introduces a new image coding scheme using lattice vector quantization. The proposed method involves two steps: biorthogonal wavelet transform of the image, and lattice vector quantization of wavelet coefficients. In order to obtain a compromise between minimum distortion and bit rate, we must truncate and scale the lattice suitably. To meet this goal, we need to know how many lattice points lie within the truncated area. We investigate the case of Laplacian sources where surfaces of equal probability are spheres for the L(1) metric (pyramids) for arbitrary lattices. We give explicit generating functions for the codebook sizes for the most useful lattices like Z(n), D(n), E(s), wedge(16).
We propose a new wavelet compression algorithm based on the rate-distortion optimization for densely sampled triangular meshes. This algorithm includes the normal remesher, a wavelet transform, and an original bit allocation optimizing the quantization of the wavelet coefficients. The allocation process minimizes the reconstruction error for a given bit budget. As distortion measure, we use the mean square error of the normal mesh quantization, expressed according to the quantization error of each subband. We show that this metric is a suitable criterion to evaluate the reconstruction error, i.e. the geometric distance between the input mesh and the quantized normal one. Moreover, to design a useful bit allocation, we propose a model-based approach, depending on the wavelet coefficient distributions. The proposed algorithm achieves results better than state-of-the-art methods, up to +2.5 dB compared to the original zerotree coder for normal meshes.
This paper proposes a new method for the design of lifting filters to compute a multidimensional nonseparable wavelet transform. Our approach is stated in the general case, and is illustrated for the 2-D separable and for the quincunx images. Results are shown for the JPEG2000 database and for satellite images acquired on a quincunx sampling grid. The design of efficient quincunx filters is a difficult challenge which has already been addressed for specific cases. Our approach enables the design of less expensive filters adapted to the signal statistics to enhance the compression efficiency in a more general case. It is based on a two-step lifting scheme and joins the lifting theory with Wiener's optimization. The prediction step is designed in order to minimize the variance of the signal, and the update step is designed in order to minimize a reconstruction error. Application for lossy compression shows the performances of the method.
Presents a new scheme for fractal image compression based on adaptive Delaunay triangulation. Such a partition is computed on an initial set of points obtained with a split and merge algorithm in a grey level dependent way. The triangulation is thus fully flexible and returns a limited number of blocks allowing good compression ratios. Moreover, a second original approach is the integration of a classification step based on a modified version of the Lloyd algorithm (vector quantization) in order to reduce the encoding complexity. The vector quantization algorithm is implemented on pixel histograms directly generated from the triangulation. The aim is to reduce the number of comparisons between the two sets of blocks involved in fractal image compression by keeping only the best representative triangles in the domain blocks set. Quality coding results are achieved at rates between 0.25-0.5 b/pixel depending on the nature of the original image and on the number of triangles retained.
We present an efficient compression scheme for animated sequences of triangular meshes of the same connectivity. The proposed algorithm exploits the temporal coherence of the geometry component by using a temporal wavelet filtering. The quantization of the resulting wavelet coefficients is then optimized by a bit allocation process. This process dispatches the bit budget across the coefficient subbands according to their influence on the quality of the reconstructed sequence for one specific user-given bitrate. The proposed scheme is simple, fast, flexible, and scalable in frame rate and bitrate. Moreover, simulation results show that our approach is competitive for any kind of animated models, whatever the characteristics (parametrically coherent or not, fine/coarse meshes...), contrary to the related works.
Digital holography allows the recording, storage and subsequent reconstruction of both amplitude and phase of the light field scattered by an object. This is accomplished by recording interference patterns that preserve the properties of the original object field essential for 3D visualization, the so-called holograms. Digital holography refers to the acquisition of holograms with a digital sensor, typically a CCD or a CMOS camera, and to the reconstruction of the 3D object field using numerical methods. In the current work, the different representations of digital holographic information in the hologram and in the object planes are studied. The coding performance of the different complex field representations, notably Amplitude-Phase and Real-Imaginary, in both the hologram plane and the object plane, is assessed using both computer generated and experimental holograms. The HEVC intra main coding profile is used for the compression of the different representations in both planes, either for experimental holograms or computer generated holograms. The HEVC intra compression in the object plane outperforms encoding in the hologram plane. Furthermore, encoding computer generated holograms in the object plane has a larger benefit than the same encoding over the experimental holograms. This difference was expected, since experimental holograms are affected by a larger negative influence of speckle noise, resulting in a loss of compression efficiency. This work emphasizes the possibility of holographic coding on the object plane, instead of the common encoding in the hologram plane approach. Moreover, this possibility allows direct visualization of the Object Plane Amplitude in a regular 2D display without any transformation methods. The complementary phase information can easily be used to render 3D features such as depth map, multi-view or even holographic interference patterns for further 3D visualization depending on the display technology.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.