This paper introduces a new challenge and datasets to foster research toward designing systems that can understand medical videos and provide visual answers to natural language questions. We believe medical videos may provide the best possible answers to many first aid, medical emergency, and medical education questions. Toward this, we created the MedVidCL and MedVidQA datasets and introduce the tasks of Medical Video Classification (MVC) and Medical Visual Answer Localization (MVAL), two tasks that focus on cross-modal (medical language and medical video) understanding. The proposed tasks and datasets have the potential to support the development of sophisticated downstream applications that can benefit the public and medical practitioners. Our datasets consist of 6, 117 annotated videos for the MVC task and 3, 010 annotated questions and answers timestamps from 899 videos for the MVAL task. These datasets have been verified and corrected by medical informatics experts. We have also benchmarked 1 each task with the created MedVidCL and MedVidQA datasets and propose the multimodal learning methods that set competitive baselines for future research.
Variants in the gene encoding ankyrin repeat and SOCS box–containing 4 ( ASB4 ) are linked to human obesity. Here, we characterized the pathways underlying the metabolic functions of ASB4. Hypothalamic Asb4 expression was suppressed by fasting in wild-type mice but not in mice deficient in AgRP , which encodes Agouti-related protein (AgRP), an appetite-stimulating hormone, suggesting that ASB4 is a negative target of AgRP. Many ASB4 neurons in the brain were adjacent to AgRP terminals, and feeding induced by AgRP neuronal activation was disrupted in Asb4 -deficient mice. Acute knockdown of Asb4 in the brain caused marked hyperphagia due to increased meal size, and Asb4 deficiency led to increased meal size and food intake at the onset of refeeding, when very large meals were consumed. Asb4 -deficient mice were resistant to the meal-terminating effects of exogenously administered calcitonin and showed decreased neuronal expression of Calcr , which encodes the calcitonin receptor. Pro-opiomelanocortin (POMC) neurons in the arcuate nucleus in mice are involved in glucose homeostasis, and Asb4 deficiency specifically in POMC neurons resulted in glucose intolerance that was independent of obesity. Furthermore, individuals with type 2 diabetes showed reduced ASB4 abundance in the infundibular nuclei, the human equivalent of the arcuate nucleus. Together, our results indicate that ASB4 acts in the brain to improve glucose homeostasis and to induce satiety after substantial meals, particularly those after food deprivation.
Objective Plain language in medicine has long been advocated as a way to improve patient understanding and engagement. As the field of Natural Language Processing has progressed, increasingly sophisticated methods have been explored for the automatic simplification of existing biomedical text for consumers. We survey the literature in this area with the goals of characterizing approaches and applications, summarizing existing resources, and identifying remaining challenges. Materials and Methods We search English language literature using lists of synonyms for both the task (eg, “text simplification”) and the domain (eg, “biomedical”), and searching for all pairs of these synonyms using Google Scholar, Semantic Scholar, PubMed, ACL Anthology, and DBLP. We expand search terms based on results and further include any pertinent papers not in the search results but cited by those that are. Results We find 45 papers that we deem relevant to the automatic simplification of biomedical text, with data spanning 7 natural languages. Of these (nonexclusively), 32 describe tools or methods, 13 present data sets or resources, and 9 describe impacts on human comprehension. Of the tools or methods, 22 are chiefly procedural and 10 are chiefly neural. Conclusions Though neural methods hold promise for this task, scarcity of parallel data has led to continued development of procedural methods. Various low-resource mitigations have been proposed to advance neural methods, including paragraph-level and unsupervised models and augmentation of neural models with procedural elements drawing from knowledge bases. However, high-quality parallel data will likely be crucial for developing fully automated biomedical text simplification.
This paper introduces a new challenge and datasets to foster research toward designing systems that can understand medical videos and provide visual answers to natural language questions. We believe medical videos may provide the best possible answers to many first aid, medical emergency, and medical education questions. Toward this, we created the and datasets and introduce the tasks of Medical Video Classification (MVC) and Medical Visual Answer Localization (MVAL), two tasks that focus on cross-modal (medical language and medical video) understanding. The proposed tasks and datasets have the potential to support the development of sophisticated downstream applications that can benefit the public and medical practitioners. Our datasets consist of 6,117 fine-grained annotated videos for the MVC task and 3,010 questions and answers timestamps from 899 videos for the MVAL task. These datasets have been verified and corrected by medical informatics experts. We have also benchmarked each task with the created and datasets and propose the multimodal learning methods that set competitive baselines for future research.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.