The extension of conventional clustering to hypergraph clustering, which involves higher order similarities instead of pairwise similarities, is increasingly gaining attention in computer vision. This is due to the fact that many clustering problems require an affinity measure that must involve a subset of data of size more than two. In the context of hypergraph clustering, the calculation of such higher order similarities on data subsets gives rise to hyperedges. Almost all previous work on hypergraph clustering in computer vision, however, has considered the smallest possible hyperedge size, due to a lack of study into the potential benefits of large hyperedges and effective algorithms to generate them. In this paper, we show that large hyperedges are better from both a theoretical and an empirical standpoint. We then propose a novel guided sampling strategy for large hyperedges, based on the concept of random cluster models. Our method can generate large pure hyperedges that significantly improve grouping accuracy without exponential increases in sampling costs. We demonstrate the efficacy of our technique on various higher-order grouping problems. In particular, we show that our approach improves the accuracy and efficiency of motion segmentation from dense, long-term, trajectories.
Here, we illustrate what happens inside the catalytic cleft of an enzyme when substrate or ligand binds on single-millisecond timescales. The initial phase of the enzymatic cycle is observed with near-atomic resolution using the most advanced X-ray source currently available: the European XFEL (EuXFEL). The high repetition rate of the EuXFEL combined with our mix-and-inject technology enables the initial phase of ceftriaxone binding to the Mycobacterium tuberculosis β-lactamase to be followed using time-resolved crystallography in real time. It is shown how a diffusion coefficient in enzyme crystals can be derived directly from the X-ray data, enabling the determination of ligand and enzyme–ligand concentrations at any position in the crystal volume as a function of time. In addition, the structure of the irreversible inhibitor sulbactam bound to the enzyme at a 66 ms time delay after mixing is described. This demonstrates that the EuXFEL can be used as an important tool for biomedically relevant research.
As the next wave of productivity, Industry 4.0 aims to enhance the competitiveness and efficiency of manufacturers by bridging the gap between industrial manufacturing and information technology. Through digitalization, it provides the advantage of enabling the real-time/near-real-time monitoring of manufacturing. This digital information allows monitoring tools such as Value stream mapping (VSM) to help the decision makers efficiently capture the non-value-adding processes on the factory floor. However, the application of VSM into small and medium sized enterprises (SMEs), including diverse manufacturing environments, is not an easy task. It is even more challenging especially when the product processing is more complicated and requires improvements to labour management and facility utilization. Conventional VSM is not competent to handle the contemporary rapid dynamic manufacturing environment, complex material flow or efficiency of machine and labour performance. These three are the most important resources on the shop floor to bring transparency to the decision maker. We present a multi-agent system composed of several cost effective embedded Arduino systems as agents and a Raspberry-Pi ® as a core agent. Equipped with Cyber-Physical System (CPS) technology, these agents, placed on or near the station, could reflect the non-linear material value flow without modelling the process or using RFID tags. Moreover, through the sensor node installed in each machine and by knowing the staff ID, the agents could send the relevant information in the form of dynamic value stream mapping (DVSM) in near-real-time for storage, analysis and visualization. We present a suitable visualization tool based in Node-RED ® to carry out DVSM.
A peak-finding algorithm for serial crystallography (SX) data analysis based on the principle of `robust statistics' has been developed. Methods which are statistically robust are generally more insensitive to any departures from model assumptions and are particularly effective when analysing mixtures of probability distributions. For example, these methods enable the discretization of data into a group comprising inliers (i.e. the background noise) and another group comprising outliers (i.e. Bragg peaks). Our robust statistics algorithm has two key advantages, which are demonstrated through testing using multiple SX data sets. First, it is relatively insensitive to the exact value of the input parameters and hence requires minimal optimization. This is critical for the algorithm to be able to run unsupervised, allowing for automated selection or `vetoing' of SX diffraction data. Secondly, the processing of individual diffraction patterns can be easily parallelized. This means that it can analyse data from multiple detector modules simultaneously, making it ideally suited to real-time data processing. These characteristics mean that the robust peak finder (RPF) algorithm will be particularly beneficial for the new class of MHz X-ray free-electron laser sources, which generate large amounts of data in a short period of time.
Identifying the underlying models in a set of data points that is contaminated by noise and outliers leads to a highly complex multi-model fitting problem. This problem can be posed as a clustering problem by the projection of higher-order affinities between data points into a graph, which can be clustered using spectral clustering. Calculating all possible higher-order affinities is computationally expensive. Hence, in most cases, only a subset is used. In this paper, we propose an effective sampling method for obtaining a highly accurate approximation of the full graph, which is required to solve multi-structural model fitting problems in computer vision. The proposed method is based on the observation that the usefulness of a graph for segmentation improves as the distribution of the hypotheses that are used to build the graph approaches the distribution of the actual parameters for the given data. In this paper, we approximate this actual parameter distribution by using a th-order statistics-based cost function, and the samples are generated using a greedy algorithm that is coupled with a data sub-sampling strategy. The experimental analysis shows that the proposed method is both accurate and computationally efficient compared with the state-of-the-art robust multi-model fitting techniques. The implementation of the method is publicly available from https://github.com/RuwanT/model-fitting-cbs.
X-ray crystallography has witnessed a massive development over the past decade, driven by large increases in the intensity and brightness of X-ray sources and enabled by employing high-frame-rate X-ray detectors. The analysis of large data sets is done via automatic algorithms that are vulnerable to imperfections in the detector and noise inherent with the detection process. By improving the model of the behaviour of the detector, data can be analysed more reliably and data storage costs can be significantly reduced. One major requirement is a software mask that identifies defective pixels in diffraction frames. This paper introduces a methodology and program based upon concepts of machine learning, called robust mask maker (RMM), for the generation of bad-pixel masks for large-area X-ray pixel detectors based on modern robust statistics. It is proposed to discriminate normally behaving pixels from abnormal pixels by analysing routine measurements made with and without X-ray illumination. Analysis software typically uses a Bragg peak finder to detect Bragg peaks and an indexing method to detect crystal lattices among those peaks. Without proper masking of the bad pixels, peak finding methods often confuse the abnormal values of bad pixels in a pattern with true Bragg peaks and flag such patterns as useful regardless, leading to storage of enormous uninformative data sets. Also, it is computationally very expensive for indexing methods to search for crystal lattices among false peaks and the solution may be biased. This paper shows how RMM vastly improves peak finders and prevents them from labelling bad pixels as Bragg peaks, by demonstrating its effectiveness on several serial crystallography data sets.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.