2020
DOI: 10.21307/ijanmc-2020-018
|View full text |Cite
|
Sign up to set email alerts
|

Application Research of Crawler and Data Analysis Based on Python

Abstract: Combined with the actual situation, this paper explores how to develop a crawler method based on the specific framework for the complete interface of steam manufacturers and stores, which should be able to automatically and efficiently crawl the data of specific targets, analyze the dynamic pages, and complete the data cleaning, downloading, saving and other operations, explore the methods of general data analysis, and Analyze the downloaded data, extract useful information from it, analyze and summarize the s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 12 publications
(4 citation statements)
references
References 2 publications
0
4
0
Order By: Relevance
“…In this study, a web crawler computing application was implemented to gather firsthand textual data by accessing the webpage data, through the designed code of computer programming languages R and Python, to obtain useful data from massive data [ 26 , 27 ]. The process was implemented using Python programming, as shown in Figure 1 .…”
Section: Methodsmentioning
confidence: 99%
“…In this study, a web crawler computing application was implemented to gather firsthand textual data by accessing the webpage data, through the designed code of computer programming languages R and Python, to obtain useful data from massive data [ 26 , 27 ]. The process was implemented using Python programming, as shown in Figure 1 .…”
Section: Methodsmentioning
confidence: 99%
“…Using Python software, through the way of crawling, using the URL address of the crawl data, to this URL, and then get the response data from this URL server, in accordance with the requirements of Web crawler standards, in a specific order and principles to get the web page information into the URL parts to be grabbed, the process has been repeated until the information captured to meet the requirements of the system, before stopping the abovementioned action, then the data information obtained by the crawl will be filtered and summarized, while building a search engine [7] , put it into a database or folder, and finally according to the specific query needs [8] , from the database or folder to retrieve the required data information, and the data information through the text of the way presented, and then download, so as to get different shopping data [9] . Then you need to crawl the data to clean, and finally get the project can use the data, such as Table 1, to facilitate the construction of the subsequent database.…”
Section: Crawler-based Data Collectionmentioning
confidence: 99%
“…which significantly accelerate development. By separating concerns with the MTV pattern and providing rich reusable components, Django reduces complexity and speeds up development and maintenance of scalable database-driven web applications [10].…”
Section: Application Architecture Designmentioning
confidence: 99%