全网最全的Windows下Anaconda2 / Anaconda3里正确下载安装爬虫框架Scrapy(离线方式和在线方式)(图文详解)
不多说,直接上干货!
参考博客
全网最全的Windows下Anaconda2 / Anaconda3里正确下载安装OpenCV(离线方式和在线方式)(图文详解)
第一步:首先,提示升级下pip
第二步:下载安装wheel
也可以去网站里先下载好,离线安装。也可以如上在线安装。
第三步: 安装lxml
因为,我的是
成功!
验证下
第四步:安装Twisted
PS C:\Anaconda3\Lib\site-packages> pip install Twisted-17.9.-cp35-cp35m-win_amd64.whl
Processing c:\anaconda3\lib\site-packages\twisted-17.9.-cp35-cp35m-win_amd64.whl
Collecting hyperlink>=17.1. (from Twisted==17.9.)
Downloading hyperlink-17.3.-py2.py3-none-any.whl (73kB)
% |████████████████████████████████| 81kB .1kB/s
Collecting incremental>=16.10. (from Twisted==17.9.)
Downloading incremental-17.5.-py2.py3-none-any.whl
Collecting constantly>=15.1 (from Twisted==17.9.)
Downloading constantly-15.1.-py2.py3-none-any.whl
Collecting zope.interface>=4.0. (from Twisted==17.9.)
Downloading zope.interface-4.4.-cp35-cp35m-win_amd64.whl (139kB)
% |████████████████████████████████| 143kB 13kB/s
Collecting Automat>=0.3. (from Twisted==17.9.)
Downloading Automat-0.6.-py2.py3-none-any.whl
Requirement already satisfied: setuptools in c:\anaconda3\lib\site-packages\setuptools-27.2.-py3..egg (from zope.interface>=4.0.->Twisted==17.9.)
Requirement already satisfied: six in c:\anaconda3\lib\site-packages (from Automat>=0.3.->Twisted==17.9.)
Collecting attrs (from Automat>=0.3.->Twisted==17.9.)
Downloading attrs-17.4.-py2.py3-none-any.whl
Installing collected packages: hyperlink, incremental, constantly, zope.interface, attrs, Automat, Twisted
Successfully installed Automat-0.6. Twisted-17.9. attrs-17.4. constantly-15.1. hyperlink-17.3. incremental-17.5. zope.interface-4.4.
PS C:\Anaconda3\Lib\site-packages>
第五步:安装scrapy
PS C:\Anaconda3\Lib\site-packages> pip install scrapy
Collecting scrapy
Downloading Scrapy-1.5.-py2.py3-none-any.whl (251kB)
% |████████████████████████████████| 256kB 82kB/s
Requirement already satisfied: Twisted>=13.1. in c:\anaconda3\lib\site-packages (from scrapy)
Requirement already satisfied: lxml in c:\anaconda3\lib\site-packages (from scrapy)
Collecting parsel>=1.1 (from scrapy)
Downloading parsel-1.3.-py2.py3-none-any.whl
Collecting service-identity (from scrapy)
Downloading service_identity-17.0.-py2.py3-none-any.whl
Requirement already satisfied: pyOpenSSL in c:\anaconda3\lib\site-packages (from scrapy)
Requirement already satisfied: six>=1.5. in c:\anaconda3\lib\site-packages (from scrapy)
Collecting queuelib (from scrapy)
Downloading queuelib-1.4.-py2.py3-none-any.whl
Collecting PyDispatcher>=2.0. (from scrapy)
Downloading PyDispatcher-2.0..tar.gz
Collecting w3lib>=1.17. (from scrapy)
Downloading w3lib-1.18.-py2.py3-none-any.whl
Collecting cssselect>=0.9 (from scrapy)
Downloading cssselect-1.0.-py2.py3-none-any.whl
Requirement already satisfied: incremental>=16.10. in c:\anaconda3\lib\site-packages (from Twisted>=13.1.->scrapy)
Requirement already satisfied: Automat>=0.3. in c:\anaconda3\lib\site-packages (from Twisted>=13.1.->scrapy)
Requirement already satisfied: hyperlink>=17.1. in c:\anaconda3\lib\site-packages (from Twisted>=13.1.->scrapy)
Requirement already satisfied: zope.interface>=4.0. in c:\anaconda3\lib\site-packages (from Twisted>=13.1.->scrapy)
Requirement already satisfied: constantly>=15.1 in c:\anaconda3\lib\site-packages (from Twisted>=13.1.->scrapy)
Requirement already satisfied: attrs in c:\anaconda3\lib\site-packages (from service-identity->scrapy)
Collecting pyasn1-modules (from service-identity->scrapy)
Downloading pyasn1_modules-0.2.-py2.py3-none-any.whl (60kB)
% |████████████████████████████████| 61kB 51kB/s
Requirement already satisfied: pyasn1 in c:\anaconda3\lib\site-packages (from service-identity->scrapy)
Requirement already satisfied: cryptography>=1.3. in c:\anaconda3\lib\site-packages (from pyOpenSSL->scrapy)
Requirement already satisfied: setuptools in c:\anaconda3\lib\site-packages\setuptools-27.2.-py3..egg (from zope.interface>=4.0.->Twisted>=13.1.->scrapy)
Requirement already satisfied: idna>=2.0 in c:\anaconda3\lib\site-packages (from cryptography>=1.3.->pyOpenSSL->scrapy)
Requirement already satisfied: cffi>=1.4. in c:\anaconda3\lib\site-packages (from cryptography>=1.3.->pyOpenSSL->scrapy)
Requirement already satisfied: pycparser in c:\anaconda3\lib\site-packages (from cffi>=1.4.->cryptography>=1.3.->pyOpenSSL->scrapy)
Building wheels for collected packages: PyDispatcher
Running setup.py bdist_wheel for PyDispatcher ... done
Stored in directory: C:\Users\lenovo\AppData\Local\pip\Cache\wheels\\\a1\5857c77600a28813aaf0f66d4e4568f50c9f133277a4122411
Successfully built PyDispatcher
Installing collected packages: w3lib, cssselect, parsel, pyasn1-modules, service-identity, queuelib, PyDispatcher, scrapy
Successfully installed PyDispatcher-2.0. cssselect-1.0. parsel-1.3. pyasn1-modules-0.2. queuelib-1.4. scrapy-1.5. service-identity-17.0. w3lib-1.18.
PS C:\Anaconda3\Lib\site-packages>
验证下
当然,你也可以直接在这里安装
同时,大家可以关注我的个人博客:
http://www.cnblogs.com/zlslch/ 和 http://www.cnblogs.com/lchzls/ http://www.cnblogs.com/sunnyDream/
详情请见:http://www.cnblogs.com/zlslch/p/7473861.html
人生苦短,我愿分享。本公众号将秉持活到老学到老学习无休止的交流分享开源精神,汇聚于互联网和个人学习工作的精华干货知识,一切来于互联网,反馈回互联网。
目前研究领域:大数据、机器学习、深度学习、人工智能、数据挖掘、数据分析。 语言涉及:Java、Scala、Python、Shell、Linux等 。同时还涉及平常所使用的手机、电脑和互联网上的使用技巧、问题和实用软件。 只要你一直关注和呆在群里,每天必须有收获
对应本平台的讨论和答疑QQ群:大数据和人工智能躺过的坑(总群)(161156071)
全网最全的Windows下Anaconda2 / Anaconda3里正确下载安装爬虫框架Scrapy(离线方式和在线方式)(图文详解)的更多相关文章
- 全网最全的Windows下Anaconda2 / Anaconda3里正确下载安装OpenCV(离线方式和在线方式)(图文详解)
不多说,直接上干货! 说明: Anaconda2-5.0.0-Windows-x86_64.exe安装下来,默认的Python2.7 Anaconda3-4.2.0-Windows-x86_64.ex ...
- 全网最全的Windows下Anaconda2 / Anaconda3里正确下载安装用来向微信好友发送消息的itchat库(图文详解)
不多说,直接上干货! Anaconda2 里 PS C:\Anaconda2\Scripts> PS C:\Anaconda2\Scripts> pip.exe install itch ...
- 全网最全的Windows下Anaconda2 / Anaconda3里正确下载安装用来定时任务apscheduler库(图文详解)
不多说,直接上干货! Anaconda2 里 PS C:\Anaconda2\Scripts> PS C:\Anaconda2\Scripts> pip.exe install apsc ...
- 全网最全的Windows下Anaconda2 / Anaconda3里正确下载安装Theano(图文详解)
不多说,直接上干货! Theano的安装教程目前网上一搜很多,前几天折腾了好久,终于安装成功了Anaconda3(Python3)的Theano,嗯~发博客总结并分享下经验教训吧. 渣电脑,显卡用的是 ...
- 全网最全的Windows下Python2 / Python3里正确下载安装用来向微信好友发送消息的itchat库(图文详解)
不多说,直接上干货! 建议,你用Anaconda2或Anaconda3. 见 全网最全的Windows下Anaconda2 / Anaconda3里正确下载安装用来向微信好友发送消息的itchat库( ...
- 全网最全的Windows下Anaconda2 / Anaconda3里Python语言实现定时发送微信消息给好友或群里(图文详解)
不多说,直接上干货! 缘由: (1)最近看到情侣零点送祝福,感觉还是很浪漫的事情,相信有很多人熬夜为了给爱的人送上零点祝福,但是有时等着等着就睡着了或者时间并不是卡的那么准就有点强迫症了,这是也许程序 ...
- 全网最详细的Windows系统里Oracle 11g R2 Database(64bit)的完全卸载(图文详解)
不多说,直接上干货! 前期博客 全网最详细的Windows系统里Oracle 11g R2 Database(64bit)的下载与安装(图文详解) 若你不想用了,则可安全卸载. 完全卸载Oracle ...
- Windows下如何正确下载并安装可视化的Redis数据库管理工具(redis-desktop-manager)(图文详解)
不多说,直接上干货! Redis Desktop Manager是一个可视化的Redis数据库管理工具,使用非常简单. 官网下载:https://redisdesktop.com/down ...
- 全网最详细的用pip安装****模块报错:Could not find a version that satisfies the requirement ****(from version:) No matching distribution found for ****的解决办法(图文详解)
不多说,直接上干货! 问题详情 这个问题,很普遍.如我这里想实现,Windows下Anaconda2 / Anaconda3里正确下载安装用来向微信好友发送消息的itchat库. 见,我撰写的 全网最 ...
随机推荐
- MySQL数据库学习书单/书籍
MySQL数据库学习书单 1.MySQL必知必会(MySQL Crash Course) 豆瓣详情 主要适合前期掌握MySQL,豆瓣评分8.4. 2.高性能MySQL 豆瓣详情 主要适合进阶阅读使 ...
- python编程之如何在Windows上安装python
一.安装python 首先检查你的系统里是否安装了python,开始菜单里点击运行输入cmd打开一个命令窗口,或键盘快捷键windows+R打开,在窗口中输入python并回车,如果出现了Python ...
- Linux下mysql定时自动备份并FTP到远程脚本
1.添加backupmysqleveryday.sh(vi /data/shell/backupmysqleveryday.sh) #!/bin/sh #this shell is user for ...
- PAT乙级1003
1003 我要通过! (20 point(s)) “答案正确”是自动判题系统给出的最令人欢喜的回复.本题属于 PAT 的“答案正确”大派送 —— 只要读入的字符串满足下列条件,系统就输出“答案正确”, ...
- shell遍历文件夹
遍历目录下的所有文件 假如有一个文件夹路径为dir,遍历文件 for file in /path/dir/* do if test -f $file then echo $file arrary=($ ...
- 所有子节点、Procedure、MySQL
在Oracle 中我们知道有一个 Hierarchical Queries 通过CONNECT BY 我们可以方便的查了所有当前节点下的所有子节点.但很遗憾,在MySQL的目前版本中还没有对应的功能. ...
- 利用 awk 统计nginx 中某一个用户的访问次数
线上总是会遇到攻击,所以就需要分析 access.log 看看那些用户的访问次数不正常,针对这些不正常的用户,要做处理,以 access.log为例说明下怎么统计. 通过 access.log 日志来 ...
- jquery validate 校验使用总结
一.jquery.validator表单验证id和name问题 因为后台是struts,表单提交,所有输入框的值保存在name=对象.名字中,而jquery.validator表单验证用的是name, ...
- Gradle项目BUG
Error starting ApplicationContext. To display the auto-configuration report re-run your application ...
- c++ 实现hashmap
由于hashmap不是c++ stl中标准实现,这样在跨平台使用时就可能会出现问题,于是想到自己实现一个hashmap hash算法使用开链法解决hash冲突,主要实现了添加,删除,查找几个方法 头文 ...