SignificanceThe earliest way humans can learn what their body is and where the outside world begins is through the tactile sense, especially through touch between parent and baby. In this study, we demonstrated differential processing of touch from self and others at cortical and spinal levels. Our results support top-down modulation of dorsal horn somatosensory processing, as recently shown in animal studies. We provide evidence that the individual self-concept relates to differential self- vs. other-processing in the tactile domain. Self- vs. other-distinction is necessary for successful social interaction with others and for establishing a coherent self. Our results suggest an association between impaired somatosensory processing and a dysfunctional self-concept, as seen in many psychiatric disorders.
Grasping and manipulating an object requires us to perceive its material compliance. Compliance is thought to be encoded by relationships of force, displacement, and contact area at the finger pad. Prior work suggests that objects must be sufficiently deformed to become discriminable, but the utility of time-dependent cues has not been fully explored. The studies herein find that the availability of force-rate cues improve compliance discriminability so as to require less deformation of stimulus and finger pad. In particular, we tested the impact of controlling force-rate and displacement-rate cues in passive touch psychophysical experiments. An ink-based method to mark the finger pad was used to measure contact area per stimulus, simultaneously with displacement and force. Compliances spanned a range harder and softer than the finger pad. The results indicated harder compliances were discriminable at lower peak forces when the stimulus control mode was displacement-rate (0.5 N) compared to force-rate (1.3 N). That is, when displacement-rate was controlled to be equal between the two compliances, the resultant force-rate psychophysical cues could be more readily discriminated. In extending prior studies, while some magnitude of finger pad deformation may be sufficient for discriminability, temporal cues tied to force afford more efficient judgments.
Applications of machine learning are subject to three major components that contribute to the final performance metrics. Within the category of neural networks, and deep learning specifically, the first two are the architecture for the model being trained and the training approach used. This work focuses on the third component, the data used during training. The primary questions that arise are “what is in the data” and “what within the data matters?” looking into the radio frequency machine learning (RFML) field of automatic modulation classification (AMC) as an example of a tool used for situational awareness, the use of synthetic, captured, and augmented data are examined and compared to provide insights about the quantity and quality of the available data necessary to achieve desired performance levels. Three questions are discussed within this work: (1) how useful a synthetically trained system is expected to be when deployed without considering the environment within the synthesis, (2) how can augmentation be leveraged within the RFML domain, and, lastly, (3) what impact knowledge of degradations to the signal caused by the transmission channel contributes to the performance of a system. In general, the examined data types each make useful contributions to a final application, but captured data germane to the intended use case will always provide more significant information and enable the greatest performance. Despite the benefit of captured data, the difficulties and costs that arise from live collection often make the quantity of data needed to achieve peak performance impractical. This paper helps quantify the balance between real and synthetic data, offering concrete examples where training data is parametrically varied in size and source.
In our ability to discriminate compliant, or ‘soft,’ objects, we rely upon information acquired from interactions at the finger pad. We have yet to resolve the most pertinent perceptual cues. However, doing so is vital for building effective, dynamic displays. By introducing psychophysical illusions through spheres of various size and elasticity, we investigate the utility of contact area cues, thought to be key in encoding compliance. For both active and passive touch, we determine finger pad-to-stimulus contact areas, using an ink-based procedure, as well as discrimination thresholds. The findings indicate that in passive touch, participants cannot discriminate certain small compliant versus large stiff spheres, which generate similar contact areas. In active touch, however, participants easily discriminate these spheres, though contact areas remain similar. Supplementary cues based on stimulus rate and/or proprioception seem vital. One cue that does differ for illusion cases is finger displacement given a volitionally applied force.
Understanding how we perceive differences in material compliance, or 'softness,' is a central topic in the field of haptics. The intrinsic elasticity of an object is the primary factor thought to influence our perceptual estimates. Therefore, most studies test and report the elasticity of their stimuli, typically as stiffness or modulus. However, many reported estimates are of very high magnitude for silicone-elastomers, which may be due to artifacts in characterization technique. This makes it very difficult to compare the perceptual results between the studies. The work herein defines a standardized and easy-to-implement way to characterize test stimuli. The procedure involves the unconstrained, uniaxial compression of a plate into cylindrical substrates 10 mm tall by 10 mm diameter. The resultant force-displacement data are straightforwardly converted into stress-strain data, from which a modulus is readily derived. This procedure was used to re-characterize stimuli from prior studies. The revised results from the validated method herein are 200-1,100% lower than modulus values either reported and/or approximated from stiffness. This is practically significant when differences of 10-15% are perceptually discriminable. The re-characterized estimates are useful in comparing prior studies and designing new studies. Furthermore, this characterization methodology may help more readily bridge studies on perception with those designing technology.
Touch is a powerful communication tool, but we have a limited understanding of the role played by particular physical features of interpersonal touch communication. In this study, adults living in Sweden performed a task in which messages (attention, love, happiness, calming, sadness, and gratitude) were conveyed by a sender touching the forearm of a receiver, who interpreted the messages. Two experiments ( N = 32, N = 20) showed that within close relationships, receivers could identify the intuitive touch expressions of the senders, and we characterized the physical features of the touches associated with successful communication. Facial expressions measured with electromyography varied by message but were uncorrelated with communication performance. We developed standardized touch expressions and quantified the physical features with 3D hand tracking. In two further experiments ( N = 20, N = 16), these standardized expressions were conveyed by trained senders and were readily understood by strangers unacquainted with the senders. Thus, the possibility emerges of a standardized, intuitively understood language of social touch.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.