Ambient assisted living (AAL) has the ambitious goal of improving the quality of life and maintaining independence of older and vulnerable people through the use of technology. Most of the western world will see a very large increase in the number of older people within the next 50 years with limited resources to care for them. AAL is seen as a promising alternative to the current care models and consequently has attracted lots of attention. Recently, a number of researchers have developed solutions based on video cameras and computer vision systems with promising results. However, for the domain to reach maturity, several challenges need to be faced, including the development of systems that are robust in the real-world and are accepted by users, carers and society. In this literature review paper we present a comprehensive survey of the scope of the domain, the existing technical solutions and the challenges to be faced.
Deep learning-based algorithms provide an efficient and reliable diagnosis for medical imaging. This paper proposes COVID-19 diagnosis based on analysis of Computerised tomography (CT) chest scans. In recent years, deep learningbased analysis of CT chest scans has demonstrated competitive sensitivity for pneumonia prognosis. This paper presents our submission for the 2021 ICASSP Signal Processing Grand Challenge (SPGC). We exploit a 3D Networkbased transfer learning approach to classify volumetric CT scans with a novel pre-processing method to render the volume with salient features. This work uses the pre-trained 3D ResNet50 as the backbone network. The 3D network is trained on a dataset consisting of 3 classes: Community Acquired Pneumonia (CAP), COVID-19 and Normal patient. The final testing results have shown an overall accuracy of 85.56% with the COVID-19 sensitivity attaining 82.86%.
Scalable coding-based content adaptation poses serious challenges to traditional watermarking algorithms, which do not consider the scalable coding structure and hence cannot guarantee correct watermark extraction in media consumption chain. In this paper, we propose a novel concept of scalable blind watermarking that ensures more robust watermark extraction at various compression ratios while not effecting the visual quality of host media. The proposed algorithm generates scalable and robust watermarked image code-stream that allows the user to constrain embedding distortion for target content adaptations. The watermarked image code-stream consists of hierarchically nested joint distortion-robustness coding atoms. The code-stream is generated by proposing a new wavelet domain blind watermarking algorithm guided by a quantization based binary tree. The code-stream can be truncated at any distortion-robustness atom to generate the watermarked image with the desired distortion-robustness requirements. A blind extractor is capable of extracting watermark data from the watermarked images. The algorithm is further extended to incorporate a bit-plane discarding-based quantization model used in scalable coding-based content adaptation, e.g., JPEG2000. This improves the robustness against quality scalability of JPEG2000 compression. The simulation results verify the feasibility of the proposed concept, its applications, and its improved robustness against quality scalable content adaptation. Our proposed algorithm also outperforms existing methods showing 35% improvement. In terms of robustness to quality scalable video content adaptation using Motion JPEG2000 and wavelet-based scalable video coding, the proposed method shows major improvement for video watermarking.
A fast compressed sensing reconstruction using least squares method with the signal correlation is presented in this paper. It is well known that the complexity of 𝑙 ! -minimisation is very high and is undesirable for many practical applications. The least squares method, on the other hand, has a much lower complexity. However, least squares does not promote the sparsity of signal and therefore cannot provide acceptable reconstructed results. The main contribution of this paper is to show that by exploiting signal correlation, the reconstruction error of least squares is greatly improved. Moreover, the correlated reference used in this method is very flexible, and can contain many kinds of correlation, such as spatial or temporal correlation. Experimental results show that the performance of this method is comparable to the state-of-theart algorithms, whilst having a much lower complexity. It also shows that this method can be applied to both sparse and redundant signal reconstruction.
This is the unspecified version of the paper.This version of the publication may differ from the final published version. ABSTRACT Over the recent years there has been growing interest to propose a robust and efficient hand gesture recognition (HGR) system, using real-time depth sensors like Microsoft Kinect. The performance of such HGR systems have been affected by the low resolution, noise and quantization error in the depth stream. In this paper, we propose a method to pre-process Kinect depth stream in order to overcome some of these limitations. The design approach utilizes the hand tracker from OpenNI SDK to perform distance invariant segmentation of hand region depth stream. This is followed by the construction of three different projections of hand in XY, ZX and ZY planes. These projections are then further enhanced using a combination of morphological closing and simple averaging based interpolation. The evaluation results show above 80% similarity with ground truth, and 1.45-5.35% increase in accuracy for gestures with recognition accuracy less than 90%. Permanent repository link
With the emergence of new scalable coding standards, such as JPEG2000, multimedia is stored as scalable coded bit streams that may be adapted to cater network, device and usage preferences in multimedia usage chains providing universal multimedia access. These adaptations include quality, resolution, frame rate and region of interest scalability and achieved by discarding least significant parts of the bit stream according to the scalability criteria. Such content adaptations may also effect the content protection data, such as watermarks, hidden in the original content. Many wavelet-based robust watermarking techniques robust to such JPEG2000 compression attacks are proposed in the literature. In this paper, we have categorized and evaluated the robustness of such wavelet based image watermarking techniques against JPEG2000 compression, in terms of algorithmic choices, wavelet kernel selection, subband selection, or watermark selection using a new modular framework. As most of the algorithms uses a different set of parametric combination, this analysis is particularly useful to understand the effect of various parameters on the robustness under a common platform and helpful to design any such new algorithm. The analysis also considers the imperceptibility performance of the watermark embedding, as robustness and imperceptibility are two main watermarking properties, complementary to each other.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.