Created
November 13, 2016 13:07
-
-
Save sdfsdhgjkbmnmxc/bc013d9602fdbc21542a3cdeaf4024f5 to your computer and use it in GitHub Desktop.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
crawler.py http://cython.readthedocs.io | |
crawler INFO crawler.py:104 load index start | |
crawler INFO crawler.py:116 load index done | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.io | |
crawler DEBUG crawler.py:259 url http://cython.readthedocs.io found | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.io_static/nature.css | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.io_static/nature.css | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.io_static/pygments.css | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.io_static/pygments.css | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.io_static/favicon.ico | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.io_static/favicon.ico | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.io# | |
crawler DEBUG crawler.py:259 url http://cython.readthedocs.io# found | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/quickstart/index.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/quickstart/index.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.io_static', port=80): Max retries exceeded with url: /nature.css (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c05190>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079ef910: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.io_static/nature.css')> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.io_static/readthedocs-data.js | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.io_static/readthedocs-data.js | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.io_static', port=80): Max retries exceeded with url: /favicon.ico (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c05690>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079efa50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.io_static/favicon.ico')> failed with ConnectionError | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /quickstart/index.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c05890>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079eff50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/quickstart/index)> failed with ConnectionError | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.io_static', port=80): Max retries exceeded with url: /pygments.css (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c05410>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079ef9b0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.io_static/pygments.css)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.io_static/readthedocs-dynamic-include.js | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.io_static/readthedocs-dynamic-include.js | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.io#welcome-to-cython-s-documentation | |
crawler DEBUG crawler.py:259 url http://cython.readthedocs.io#welcome-to-cython-s-documentation found | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/quickstart/overview.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/quickstart/overview.html | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/quickstart/install.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/quickstart/install.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.io_static', port=80): Max retries exceeded with url: /readthedocs-data.js (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c05450>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079efaf0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.io_static/readthedocs-)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/quickstart/build.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/quickstart/build.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.io_static', port=80): Max retries exceeded with url: /readthedocs-dynamic-include.js (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c05810>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079ef910: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.io_static/readthedocs-)> failed with ConnectionError | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /quickstart/overview.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c40990>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079eff50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/quickstart/overv)> failed with ConnectionError | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /quickstart/install.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c40950>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079efa50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/quickstart/insta)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/quickstart/cythonize.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/quickstart/cythonize.html | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/index.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/index.html | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/cython_tutorial.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/cython_tutorial.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /quickstart/build.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c050d0>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079ef9b0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/quickstart/build)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/external.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/external.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/index.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c40a10>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079efa50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/index.h)> failed with ConnectionError | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /quickstart/cythonize.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c40990>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079efaf0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/quickstart/cytho)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/clibraries.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/clibraries.html | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/cdef_classes.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/cdef_classes.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/cython_tutorial.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c40fd0>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079eff50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/cython_)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/pxd_files.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/pxd_files.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/external.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c40a10>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079ef910: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/externa)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/caveats.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/caveats.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/clibraries.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace050>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079ef9b0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/clibrar)> failed with ConnectionError | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/cdef_classes.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c40b50>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079efaf0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/cdef_cl)> failed with ConnectionError | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/pxd_files.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace290>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079efa50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/pxd_fil)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/profiling_tutorial.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/profiling_tutorial.html | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/strings.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/strings.html | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/memory_allocation.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/memory_allocation.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/caveats.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c05490>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079eff50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/caveats)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/pure.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/pure.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/profiling_tutorial.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107a16890>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079ef910: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/profili)> failed with ConnectionError | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/memory_allocation.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace2d0>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079efaf0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/memory_)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/numpy.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/numpy.html | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/array.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/array.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/strings.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace850>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079efa50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/strings)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/readings.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/readings.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/pure.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107a16890>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079ef9b0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/pure.ht)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/related_work.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/related_work.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/numpy.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c40ed0>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079eff50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/numpy.h)> failed with ConnectionError | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/array.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace250>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079efaf0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/array.h)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/appendix.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/appendix.html | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/index.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/index.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/readings.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c409d0>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079ef910: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/reading)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/language_basics.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/language_basics.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/related_work.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ac6910>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079efa50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/related)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/extension_types.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/extension_types.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/index.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ac6bd0>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079efaf0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/index.)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/special_methods.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/special_methods.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/appendix.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c050d0>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079ef9b0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/appendi)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/sharing_declarations.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/sharing_declarations.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/language_basics.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ac6850>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079eff50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/langua)> failed with ConnectionError | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/extension_types.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ac6e50>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079ef910: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/extens)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/external_C_code.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/external_C_code.html | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/source_files_and_compilation.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/source_files_and_compilation.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/special_methods.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c05650>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079efa50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/specia)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/early_binding_for_speed.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/early_binding_for_speed.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/sharing_declarations.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ac6710>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079efaf0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/sharin)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/wrapping_CPlusPlus.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/wrapping_CPlusPlus.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/source_files_and_compilation.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c05950>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079ef910: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/source)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/fusedtypes.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/fusedtypes.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/external_C_code.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107acea10>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079ef9b0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/extern)> failed with ConnectionError | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/early_binding_for_speed.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ac6790>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079eff50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/early_)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/pypy.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/pypy.html | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/limitations.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/limitations.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/wrapping_CPlusPlus.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c05750>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079efa50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/wrappi)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/pyrex_differences.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/pyrex_differences.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/fusedtypes.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace290>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079efaf0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/fusedt)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/memoryviews.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/memoryviews.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/pypy.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace1d0>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079ef910: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/pypy.h)> failed with ConnectionError | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/limitations.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c05550>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079eff50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/limita)> failed with ConnectionError | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/pyrex_differences.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace110>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079ef9b0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/pyrex_)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/buffer.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/buffer.html | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/parallelism.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/parallelism.html | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/debugging.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/debugging.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/memoryviews.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace390>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079efa50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/memory)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/numpy_tutorial.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/numpy_tutorial.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/buffer.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace0d0>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079efaf0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/buffer)> failed with ConnectionError | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/debugging.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107acec90>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079eff50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/debugg)> failed with ConnectionError | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/parallelism.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107acea50>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079ef9b0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/parall)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/index.html#indices-and-tables | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/index.html#indices-and-tables | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/reference/index.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/reference/index.html | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/reference/compilation.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/reference/compilation.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/numpy_tutorial.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107b8b550>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079ef910: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/numpy_)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/reference/language_basics.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/reference/language_basics.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /reference/index.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace610>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079ef9b0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/reference/index.)> failed with ConnectionError | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/index.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107b8b5d0>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079efa50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/index.)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/reference/extension_types.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/reference/extension_types.html | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/reference/interfacing_with_other_code.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/reference/interfacing_with_other_code.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /reference/compilation.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107acecd0>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079eff50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/reference/compil)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/reference/special_mention.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/reference/special_mention.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /reference/language_basics.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107b8b690>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079efaf0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/reference/langua)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/reference/limitations.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/reference/limitations.html | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /reference/extension_types.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace490>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079ef910: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/reference/extens)> failed with ConnectionError | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /reference/interfacing_with_other_code.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107acea50>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079efa50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/reference/interf)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/reference/directives.html | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/reference/directives.html | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/reference/index.html#indices-and-tables | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/reference/index.html#indices-and-tables | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /reference/special_mention.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace6d0>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079ef9b0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/reference/specia)> failed with ConnectionError | |
crawler DEBUG crawler.py:247 start http://cython.readthedocs.io_sources/index.txt | |
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.io_sources/index.txt | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /reference/limitations.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107acef50>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079eff50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/reference/limita)> failed with ConnectionError | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /reference/index.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107a16b50>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079efa50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/reference/index.)> failed with ConnectionError | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /reference/directives.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107acee90>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079efaf0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/reference/direct)> failed with ConnectionError | |
Traceback (most recent call last): | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run | |
result = self._run(*self.args, **self.kwargs) | |
File "crawler.py", line 271, in process | |
self._process_(url) | |
File "crawler.py", line 262, in _process_ | |
body = self._fetch_(url) | |
File "crawler.py", line 223, in _fetch_ | |
response = requests.get(url) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get | |
return request('get', url, params=params, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request | |
return session.request(method=method, url=url, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request | |
resp = self.send(prep, **send_kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send | |
r = adapter.send(request, **kwargs) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send | |
raise ConnectionError(e, request=request) | |
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.io_sources', port=80): Max retries exceeded with url: /index.txt (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107b8b550>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',)) | |
<Greenlet at 0x1079ef910: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.io_sources/index.txt')> failed with ConnectionError | |
Traceback (most recent call last): | |
File "crawler.py", line 294, in <module> | |
main() | |
File "crawler.py", line 290, in main | |
crawler.run(result.parallel) | |
File "crawler.py", line 175, in run | |
for (url, children) in self.todo: | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/queue.py", line 345, in next | |
result = self.get() | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/queue.py", line 284, in get | |
return self.__get_or_peek(self._get, block, timeout) | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/queue.py", line 261, in __get_or_peek | |
result = waiter.get() | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/hub.py", line 878, in get | |
return self.hub.switch() | |
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/hub.py", line 609, in switch | |
return greenlet.switch(self) | |
gevent.hub.LoopExit: ('This operation would block forever', <Hub at 0x1075d3eb0 select default pending=0 ref=0 resolver=<gevent.resolver_thread.Resolver at 0x107a16ed0 pool=<ThreadPool at 0x107126990 0/4/10>> threadpool=<ThreadPool at 0x107126990 0/4/10>>) | |
macbook-smee:crawler smee$ |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment