Social media websites are becoming more prevalent on the Internet. Sites, such as Twitter, Facebook, and Instagram, spend significantly more of their time on users online. People in social media share thoughts, views, and facts and create new acquaintances. Social media sites supply users with a great deal of useful information. This enormous quantity of social media information invites hackers to abuse data. These hackers establish fraudulent profiles for actual people and distribute useless material. The material on spam might include commercials and harmful URLs that disrupt natural users. This spam content is a massive problem in social networks. Spam identification is a vital procedure on social media networking platforms. In this paper, we have proposed a spam detection artificial intelligence technique for Twitter social networks. In this approach, we employed a vector support machine, a neural artificial network, and a random forest technique to build a model. The results indicate that, compared with RF and ANN algorithms, the suggested support vector machine algorithm has the greatest precision, recall, and F-measure. The findings of this paper would be useful in monitoring and tracking social media shared photos for the identification of inappropriate content and forged images and to safeguard social media from digital threats and attacks.
Problem statement: Text documents are the unstructured databases that contain raw data collection. The clustering techniques are used group up the text documents with reference to its similarity. Approach: The feature selection techniques were used to improve the efficiency and accuracy of clustering process. The feature selection was done by eliminate the redundant and irrelevant items from the text document contents. Statistical methods were used in the text clustering and feature selection algorithm. The cube size is very high and accuracy is low in the term based text clustering and feature selection method. The semantic clustering and feature selection method was proposed to improve the clustering and feature selection mechanism with semantic relations of the text documents. The proposed system was designed to identify the semantic relations using the ontology. The ontology was used to represent the term and concept relationship. Results: The synonym, meronym and hypernym relationships were represented in the ontology. The concept weights were estimated with reference to the ontology. The concept weight was used for the clustering process. The system was implemented in two methods. They were term clustering with feature selection and semantic clustering with feature selection. Conclusion: The performance analysis was carried out with the term clustering and semantic clustering methods. The accuracy and efficiency factors were analyzed in the performance analysis
Computational techniques for foreseeing drug-disease associations by means of incorporating gene expression as well as biological network give high intuitions to the composite associations amongst targets, drugs, disease genes in addition to the diseases at a system level. Hepatocellular Carcinoma (HCC) is a malevolent tumor containing a greater rate of sickness as well as mortality. In the present work, an Integrative framework is presented with the aim of resolving this problem, for identifying new Drugs for HCC dependent upon Multi-Source Random Walk (PD-MRW), in which score the complete drugs by means of building the drug-drug similarity network. On the other hand, the collection of clinical phenotypes as well as drug side effects in combination with patient-specific genetic info. As a result, the formation of disease-drug networks that denotes the prescriptions, which are allotted to treat those diseases that are not concentrated by means of PD-MRW model. With the aim of overcoming this issue, this research offers an integrative framework for foreseeing new drugs as well as diseases for HCC dependent upon Multi-Source Simulated Annealing based Random Walk (PDD-MSSARW). Primarily, build a Gene-Gene Weighted Interaction Network (GWIN), dependent upon the gene expression as well as protein interaction network. After that, construct a drug-drug similarity network, dependent upon multi-source random walk in GWIN, disease-drug similarity network with the help of Similarity Weighted Bipartite Graph Network (SWBGN) that is build up in which the nodes are drugs as well as association among one node to another node that explains the disease diagnoses. Lastly, dependent upon the known drugs for HCC, score the entire drugs in the similarity networks. The sturdiness of the likelihoods, their overlap with those stated in Comparative Toxicogenomics Database (CTD) as well as kinds of literature, and their enhanced KEGG pathway illustrate PDD-MSSARW method be capable of efficiently find out novel drug signs.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.