(Translated by https://www.hiragana.jp/)
GitHub - Chenmo1212/baidu_qx: Crawl Baidu migration data, divided into 3 types: city, province, country
Skip to content

Crawl Baidu migration data, divided into 3 types: city, province, country

Notifications You must be signed in to change notification settings

Chenmo1212/baidu_qx

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Baidu migration python crawler

Crawl Baidu migration data, divided into 3 types: city, province, country

Description:

Use python to crawl the population flow data in the Baidu migration platform: http://qianxi.baidu.com/, and uncomment the corresponding code according to the required type. The crawled content will be saved as an Excel file in the move_in, move_out folders in the root directory of the code

Terms and Conditions:

  1. Path: Please modify the absolute path before use

  2. Date: Please modify the date in the [migration_all_date] function (note that the left interval includes, and the right interval does not include)

  3. City interval: Since there is no asynchronous operation set during crawling, it will result in a timeout (time_out) if too many cities are crawled at one time. Therefore, it is recommended to crawl between partitions, you can copy multiple source program files, and start crawling multiple times in different intervals at the same time

  4. Run: Run the main.py file directly after making sure the above information is correct

required python packages

  • requests
  • xlwt

crawled data

About

Crawl Baidu migration data, divided into 3 types: city, province, country

Resources

Stars

Watchers

Forks

Packages

No packages published

Languages