In this paper, we propose a parallel and scalable approach for geodesic distance computation on triangle meshes. Our key observation is that the recovery of geodesic distance with the heat method [1] can be reformulated as optimization of its gradients subject to integrability, which can be solved using an efficient first-order method that requires no linear system solving and converges quickly. Afterward, the geodesic distance is efficiently recovered by parallel integration of the optimized gradients in breadth-first order. Moreover, we employ a similar breadth-first strategy to derive a parallel Gauss-Seidel solver for the diffusion step in the heat method. To further lower the memory consumption from gradient optimization on faces, we also propose a formulation that optimizes the projected gradients on edges, which reduces the memory footprint by about 50%. Our approach is trivially parallelizable, with a low memory footprint that grows linearly with respect to the model size. This makes it particularly suitable for handling large models. Experimental results show that it can efficiently compute geodesic distance on meshes with more than 200 million vertices on a desktop PC with 128GB RAM, outperforming the original heat method and other state-of-the-art geodesic distance solvers.
The local barycentric coordinates (LBC), proposed in Zhang et al. (2014), demonstrate good locality and can be used for local control on function value interpolation and shape deformation. However, it has no closedform expression and must be computed by solving an optimization problem, which can be time-consuming especially for high-resolution models. In this paper, we propose a new technique to compute LBC efficiently. The new solver is developed based on two key insights. First, we prove that the non-negativity constraints in the original LBC formulation is not necessary, and can be removed without affecting the solution of the optimization problem. Furthermore, the removal of this constraint allows us to reformulate the computation of LBC as a convex constrained optimization for its gradients, followed by a fast integration to recover the coordinate values. The reformulated gradient optimization problem can be solved using ADMM, where each step is trivially parallelizable and does not involve global linear system solving, making it much more scalable and efficient than the original LBC solver. Numerical experiments verify the effectiveness of our technique on a large variety of models.
Anderson acceleration (AA) is a popular method for accelerating fixed-point iterations, but may suffer from instability and stagnation. We propose a globalization method for AA to improve its stability and achieve global and local convergence. Unlike existing AA globalization approaches that often rely on safeguarding operations and might hinder fast local convergence, we adopt a nonmonotone trust-region framework and introduce an adaptive quadratic regularization together with a tailored acceptance mechanism. We prove global convergence and show that our algorithm attains the same local convergence as AA under appropriate assumptions. The effectiveness of our method is demonstrated in several numerical experiments.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.