Photorealistic image stylization concerns transferring style of a reference photo to a content photo with the constraint that the stylized photo should remain photorealistic. While several photorealistic image stylization methods exist, they tend to generate spatially inconsistent stylizations with noticeable artifacts. In this paper, we propose a method to address these issues. The proposed method consists of a stylization step and a smoothing step. While the stylization step transfers the style of the reference photo to the content photo, the smoothing step ensures spatially consistent stylizations. Each of the steps has a closedform solution and can be computed efficiently. We conduct extensive experimental validations. The results show that the proposed method generates photorealistic stylization outputs that are more preferred by human subjects as compared to those by the competing methods while running much faster. Source code and additional results are available at https://github.com/NVIDIA/FastPhotoStyle.
Given a random pair of images, an arbitrary style transfer method extracts the feel from the reference image to synthesize an output based on the look of the other content image. Recent arbitrary style transfer methods transfer second order statistics from reference image onto content image via a multiplication between content image features and a transformation matrix, which is computed from features with a pre-determined algorithm. These algorithms either require computationally expensive operations, or fail to model the feature covariance and produce artifacts in synthesized images. Generalized from these methods, in this work, we derive the form of transformation matrix theoretically and present an arbitrary style transfer approach that learns the transformation matrix with a feed-forward network. Our algorithm is highly efficient yet allows a flexible combination of multi-level styles while preserving content affinity during style transfer process. We demonstrate the effectiveness of our approach on four tasks: artistic style transfer, video and photo-realistic style transfer as well as domain adaptation, including comparisons with the stateof-the-art methods.
In recent years, rechargeable aqueous zinc‐ion batteries (ZIBs) have received much attention. However, the disproportionation effect of Mn2+ seriously affects the capacity retention of ZIBs during cycling. Here, the capacity retention of the Mn3O4 cathode is improved by effective valence engineering. The valence engineering of Mn3O4 is caused by bulk oxygen defects, which are in situ derived from the Mn‐metal organic framework during carbonization. Bulk oxygen defects can change the (MnO6) octahedral structure, which improves structural stability and inhibits the dissolution of Mn2+. The ZIB assembled from bulk oxygen defects Mn3O4@C nanorod arrays (Od‐Mn3O4@C NA/CC) exhibits an ultra‐long cycle life, reaching 84.1 mAh g−1 after 12 000 cycles at 5 A g−1 (up to 95.7% of the initial capacity). Furthermore, the battery has a high specific capacity of 396.2 mAh g−1 at 0.2 A g−1. Ex situ characterization results show that initial Mn3O4 is converted to ramsdellite MnO2 for insertion and extraction of H+ and Zn2+. First‐principles calculations show that the charge density of Mn3+ increases greatly, which improves the conductivity. In addition, the flexible quasi‐solid‐state ZIB is successfully assembled using Od‐Mn3O4 @ C NA/CC. Valence engineering induced by bulk oxygen defects can help develop advanced cathodes for aqueous ZIB.
Magnetic resonance imaging (MRI) provides valuable data for understanding the human mind and brain disorders, but in-scanner head motion introduces systematic and spurious biases. For example, differences in MRI measures (e.g., network strength, white matter integrity) between patient and control groups may be due to the differences in their head motion. To determine whether head motion is an important variable in itself, or just simply a confounding variable, we explored individual differences in psychological traits that may predispose some people to move more than others during an MRI scan. In the first two studies, we demonstrated in both children (N = 245) and adults (N = 581) that head motion, estimated from resting-state functional MRI and diffusion tensor imaging, was reliably correlated with impulsivity scores. Further, the difference in head motion between children with attention deficit hyperactivity disorder (ADHD) and typically developing children was largely due to differences in impulsivity. Finally, in the third study, we confirmed the observation that the regression approach, which aims to deal with motion issues by regressing out motion in the group analysis, would underestimate the effect of interest. Taken together, the present findings provide empirical evidence that links in-scanner head motion to psychological traits.
Creativity is crucial to the progression of human civilization and has led to important scientific discoveries. Especially, individuals are more likely to have scientific discoveries if they possess certain personality traits of creativity (trait creativity), including imagination, curiosity, challenge and risk-taking. This study used voxel-based morphometry to identify the brain regions underlying individual differences in trait creativity, as measured by the Williams creativity aptitude test, in a large sample (n ¼ 246). We found that creative individuals had higher gray matter volume in the right posterior middle temporal gyrus (pMTG), which might be related to semantic processing during novelty seeking (e.g. novel association, conceptual integration and metaphor understanding). More importantly, although basic personality factors such as openness to experience, extroversion, conscientiousness and agreeableness (as measured by the NEO Personality Inventory) all contributed to trait creativity, only openness to experience mediated the association between the right pMTG volume and trait creativity. Taken together, our results suggest that the basic personality trait of openness might play an important role in shaping an individuals trait creativity.
Affordance 1 modeling plays an important role in visual understanding. In this paper, we aim to predict affordances of 3D indoor scenes, specifically what human poses are afforded by a given indoor environment, such as sitting on a chair or standing on the floor. In order to predict valid affordances and learn possible 3D human poses in indoor scenes, we need to understand the semantic and geometric structure of a scene as well as its potential interactions with a human. To learn such a model, a large-scale dataset of 3D indoor affordances is required. In this work, we build a fully automatic 3D pose synthesizer that fuses semantic knowledge from a large number of 2D poses extracted from TV shows as well as 3D geometric knowledge from voxel representations of indoor scenes. With the data created by the synthesizer, we introduce a 3D pose generative model to predict semantically plausible and physically feasible human poses within a given scene (provided as a single RGB, RGB-D, or depth image). We demonstrate that our human affordance prediction method consistently outperforms existing state-of-the-art methods. The project website can be found at https
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
334 Leonard St
Brooklyn, NY 11211
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.