搜索资源列表
ZhihuDown
- java写的网络爬虫,可以爬取知乎网站等等网站的文字信息,简单易懂,可以很方便的修改爬取其他网站的关键字段。-java to write the Web crawler can crawl text messages almost known sites, and more websites, easy to understand, you can easily modify key fields crawling other site
pachongtest2
- 运用python爬取知乎日报的内容,对知乎日报网页中的每一个子链接进行爬取,并对内容进行修改,运用re,urllib2,BeautifulSoup模块。-Use python to crawl the contents of daily news, to know every page in the daily sub-links to crawl, and to modify the content, the use of re, u
dgcmscj_sss_v1
- ,采用分布式架构,是一款云端在线智能爬虫,通过使用JS渲染、代理IP、防屏 蔽、验证码识别、数据发布和导出、图表控件等一系列技术实现对全网数据精准迅速采集,无需任何专业知识即可一键爬取微信公众号、知乎、优酷、微博等海量网 站数据,并自动发布到帝国网站。 -That uses a distributed architecture, is an intelligent online cloud crawler, through the
Zhihu-master
- 利用python运用递归对知乎用户信息进行爬取(Using Python to crawl information about known users)
girls_crawler_py36
- 批量爬取知乎图片的小例子,供大家学习交流,欢迎评论(Spyder the picture of zhihu.com)
知乎爬虫任务与示范
- 利用python编写的知乎数据爬取程序,包括提问,精答,精彩评论,以及自动登录知乎网站的相关代码,局哟普一定的参考价值(Using Python to write data crawling program, including questions, answers, wonderful comments, as well as automatic login know the relevant code of the website,
zhihu_spider-master
- 知乎爬虫 此项目的功能是爬取知乎用户信息以及人际拓扑关系,爬虫框架使用scrapy,数据存储使用mongo,下载这些数据感觉也没什么用,就当为大家学习scrapy提供一个例子吧(The function of this project is to crawl and understand user information and interpersonal topological relations. The crawler fr a
Web-crawlers-master
- 这里面包括可以爬取豆瓣、知乎、微博、大麦网的数据的代码(douban,zhihu,weibo,damai)
mtianyanSearch
- 可爬取知乎、拉钩上的信息并存储到elasticsearch数据库,可以供api访问(It can crawl information on knowledge and hook and store it to elasticsearch database, which can be accessed by API.)