-
隔壁子项目 https://github.com/dataabc/weibo-crawler 可以获取用户的信息,但是这个项目必须指定用户ID。但是目前需求是,全网搜索指定关键词的帖子,同时获取到这些用户的信息。
-
在运行后也成功爬取到了数据,但是将相关配置也打开了
ITEM_PIPELINES = {
'weibo.pipelines.DuplicatesPipeline': 300,
'weibo.pipelines.CsvPipeline': 301,
# 'weibo.pipelines.MysqlPipeline': 302,
# 'weibo.pipeli…
-
no html
uid: 6026734981, oid:1005056026734981
p:1, albums:2
p:2, albums:2
Expecting value: line 1 column 1 (char 0)
stop
Traceback (most recent call last):
File "utils", line 1083, in start
Fi…
-
Apache Pinot is a realtime distributed OLAP datastore, which is used at LinkedIn, Uber, Microsoft, Slack and Weibo to deliver scalable real time analytics with low latency. for more details: https://p…
-
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
Progress: 0%| | 0/156 [00:00
-
问题场景:用户在浏览器上访问 https://m.weibo.cn/status/[A-Za-z0-9]{9}
目前的操作:无操作/药方无针对 m.weibo.cn 的相关规则
期望的操作:跳转到 https://weibo.com/{uid}/[A-Za-z0-9]{9}
可能的实现方案:获得当前网址,抓取网页中``````中的用户uid,并将网址字段重新连接为 https://wei…
-
希望支持一下微博视频的下载,比如这种格式的:
https://video.h5.weibo.cn/1034:44bb6f4a7033f03eca1f8cb280f5a0b5/4184920001855081
https://video.h5.weibo.cn/1034:4444720957745002/4444721306607329
https://video.h5.weibo.cn/…
-
你好兄弟,请问这个很多子页面点击报错,查了源码看了是没有文章和评论数据,爬取的代码网站访问报错是咋回事,有爬好的数据么可以分享下么,博主好人qaq!
'https://weibo.com/ajax/feed/hottimeline'比如这个报错:
requests.exceptions.ConnectionError: ('Connection aborted.', ConnectionRes…
-
: bridging header '/Users/machao/Desktop/weibo-swift/weibo-swift/weibo-swift/weibo-Header.h' does not exist
-
本人python小白
这个我已经去注释了
ITEM_PIPELINES = {
'weibo.pipelines.DuplicatesPipeline': 300,
'weibo.pipelines.CsvPipeline': 301,
# 'weibo.pipelines.MysqlPipeline': 302,
# 'weibo.pipelines.…