Skip to content

Instantly share code, notes, and snippets.

@sdfsdhgjkbmnmxc
Created November 13, 2016 13:07
Show Gist options
  • Save sdfsdhgjkbmnmxc/bc013d9602fdbc21542a3cdeaf4024f5 to your computer and use it in GitHub Desktop.
Save sdfsdhgjkbmnmxc/bc013d9602fdbc21542a3cdeaf4024f5 to your computer and use it in GitHub Desktop.
crawler.py http://cython.readthedocs.io
crawler INFO crawler.py:104 load index start
crawler INFO crawler.py:116 load index done
crawler DEBUG crawler.py:247 start http://cython.readthedocs.io
crawler DEBUG crawler.py:259 url http://cython.readthedocs.io found
crawler DEBUG crawler.py:247 start http://cython.readthedocs.io_static/nature.css
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.io_static/nature.css
crawler DEBUG crawler.py:247 start http://cython.readthedocs.io_static/pygments.css
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.io_static/pygments.css
crawler DEBUG crawler.py:247 start http://cython.readthedocs.io_static/favicon.ico
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.io_static/favicon.ico
crawler DEBUG crawler.py:247 start http://cython.readthedocs.io#
crawler DEBUG crawler.py:259 url http://cython.readthedocs.io# found
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/quickstart/index.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/quickstart/index.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.io_static', port=80): Max retries exceeded with url: /nature.css (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c05190>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079ef910: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.io_static/nature.css')> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.io_static/readthedocs-data.js
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.io_static/readthedocs-data.js
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.io_static', port=80): Max retries exceeded with url: /favicon.ico (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c05690>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079efa50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.io_static/favicon.ico')> failed with ConnectionError
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /quickstart/index.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c05890>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079eff50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/quickstart/index)> failed with ConnectionError
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.io_static', port=80): Max retries exceeded with url: /pygments.css (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c05410>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079ef9b0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.io_static/pygments.css)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.io_static/readthedocs-dynamic-include.js
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.io_static/readthedocs-dynamic-include.js
crawler DEBUG crawler.py:247 start http://cython.readthedocs.io#welcome-to-cython-s-documentation
crawler DEBUG crawler.py:259 url http://cython.readthedocs.io#welcome-to-cython-s-documentation found
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/quickstart/overview.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/quickstart/overview.html
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/quickstart/install.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/quickstart/install.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.io_static', port=80): Max retries exceeded with url: /readthedocs-data.js (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c05450>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079efaf0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.io_static/readthedocs-)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/quickstart/build.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/quickstart/build.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.io_static', port=80): Max retries exceeded with url: /readthedocs-dynamic-include.js (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c05810>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079ef910: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.io_static/readthedocs-)> failed with ConnectionError
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /quickstart/overview.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c40990>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079eff50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/quickstart/overv)> failed with ConnectionError
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /quickstart/install.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c40950>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079efa50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/quickstart/insta)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/quickstart/cythonize.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/quickstart/cythonize.html
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/index.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/index.html
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/cython_tutorial.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/cython_tutorial.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /quickstart/build.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c050d0>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079ef9b0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/quickstart/build)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/external.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/external.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/index.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c40a10>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079efa50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/index.h)> failed with ConnectionError
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /quickstart/cythonize.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c40990>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079efaf0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/quickstart/cytho)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/clibraries.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/clibraries.html
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/cdef_classes.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/cdef_classes.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/cython_tutorial.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c40fd0>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079eff50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/cython_)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/pxd_files.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/pxd_files.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/external.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c40a10>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079ef910: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/externa)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/caveats.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/caveats.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/clibraries.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace050>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079ef9b0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/clibrar)> failed with ConnectionError
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/cdef_classes.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c40b50>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079efaf0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/cdef_cl)> failed with ConnectionError
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/pxd_files.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace290>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079efa50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/pxd_fil)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/profiling_tutorial.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/profiling_tutorial.html
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/strings.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/strings.html
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/memory_allocation.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/memory_allocation.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/caveats.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c05490>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079eff50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/caveats)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/pure.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/pure.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/profiling_tutorial.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107a16890>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079ef910: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/profili)> failed with ConnectionError
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/memory_allocation.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace2d0>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079efaf0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/memory_)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/numpy.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/numpy.html
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/array.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/array.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/strings.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace850>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079efa50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/strings)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/readings.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/readings.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/pure.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107a16890>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079ef9b0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/pure.ht)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/related_work.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/related_work.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/numpy.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c40ed0>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079eff50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/numpy.h)> failed with ConnectionError
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/array.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace250>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079efaf0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/array.h)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/tutorial/appendix.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/tutorial/appendix.html
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/index.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/index.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/readings.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c409d0>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079ef910: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/reading)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/language_basics.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/language_basics.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/related_work.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ac6910>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079efa50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/related)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/extension_types.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/extension_types.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/index.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ac6bd0>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079efaf0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/index.)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/special_methods.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/special_methods.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /tutorial/appendix.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c050d0>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079ef9b0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/tutorial/appendi)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/sharing_declarations.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/sharing_declarations.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/language_basics.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ac6850>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079eff50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/langua)> failed with ConnectionError
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/extension_types.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ac6e50>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079ef910: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/extens)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/external_C_code.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/external_C_code.html
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/source_files_and_compilation.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/source_files_and_compilation.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/special_methods.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c05650>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079efa50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/specia)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/early_binding_for_speed.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/early_binding_for_speed.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/sharing_declarations.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ac6710>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079efaf0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/sharin)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/wrapping_CPlusPlus.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/wrapping_CPlusPlus.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/source_files_and_compilation.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c05950>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079ef910: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/source)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/fusedtypes.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/fusedtypes.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/external_C_code.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107acea10>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079ef9b0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/extern)> failed with ConnectionError
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/early_binding_for_speed.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ac6790>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079eff50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/early_)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/pypy.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/pypy.html
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/limitations.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/limitations.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/wrapping_CPlusPlus.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c05750>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079efa50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/wrappi)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/pyrex_differences.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/pyrex_differences.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/fusedtypes.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace290>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079efaf0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/fusedt)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/memoryviews.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/memoryviews.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/pypy.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace1d0>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079ef910: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/pypy.h)> failed with ConnectionError
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/limitations.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x108c05550>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079eff50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/limita)> failed with ConnectionError
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/pyrex_differences.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace110>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079ef9b0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/pyrex_)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/buffer.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/buffer.html
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/parallelism.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/parallelism.html
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/debugging.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/debugging.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/memoryviews.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace390>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079efa50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/memory)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/numpy_tutorial.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/numpy_tutorial.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/buffer.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace0d0>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079efaf0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/buffer)> failed with ConnectionError
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/debugging.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107acec90>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079eff50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/debugg)> failed with ConnectionError
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/parallelism.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107acea50>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079ef9b0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/parall)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/userguide/index.html#indices-and-tables
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/userguide/index.html#indices-and-tables
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/reference/index.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/reference/index.html
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/reference/compilation.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/reference/compilation.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/numpy_tutorial.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107b8b550>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079ef910: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/numpy_)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/reference/language_basics.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/reference/language_basics.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /reference/index.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace610>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079ef9b0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/reference/index.)> failed with ConnectionError
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /userguide/index.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107b8b5d0>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079efa50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/userguide/index.)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/reference/extension_types.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/reference/extension_types.html
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/reference/interfacing_with_other_code.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/reference/interfacing_with_other_code.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /reference/compilation.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107acecd0>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079eff50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/reference/compil)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/reference/special_mention.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/reference/special_mention.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /reference/language_basics.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107b8b690>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079efaf0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/reference/langua)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/reference/limitations.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/reference/limitations.html
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /reference/extension_types.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace490>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079ef910: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/reference/extens)> failed with ConnectionError
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /reference/interfacing_with_other_code.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107acea50>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079efa50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/reference/interf)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/reference/directives.html
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/reference/directives.html
crawler DEBUG crawler.py:247 start http://cython.readthedocs.iosrc/reference/index.html#indices-and-tables
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.iosrc/reference/index.html#indices-and-tables
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /reference/special_mention.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107ace6d0>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079ef9b0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/reference/specia)> failed with ConnectionError
crawler DEBUG crawler.py:247 start http://cython.readthedocs.io_sources/index.txt
crawler DEBUG crawler.py:222 url fetch http://cython.readthedocs.io_sources/index.txt
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /reference/limitations.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107acef50>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079eff50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/reference/limita)> failed with ConnectionError
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /reference/index.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107a16b50>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079efa50: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/reference/index.)> failed with ConnectionError
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.iosrc', port=80): Max retries exceeded with url: /reference/directives.html (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107acee90>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079efaf0: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.iosrc/reference/direct)> failed with ConnectionError
Traceback (most recent call last):
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/greenlet.py", line 534, in run
result = self._run(*self.args, **self.kwargs)
File "crawler.py", line 271, in process
self._process_(url)
File "crawler.py", line 262, in _process_
body = self._fetch_(url)
File "crawler.py", line 223, in _fetch_
response = requests.get(url)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/private/tmp/crawler/lib/python2.7/site-packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='cython.readthedocs.io_sources', port=80): Max retries exceeded with url: /index.txt (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x107b8b550>: Failed to establish a new connection: [Errno 8] nodename nor servname provided, or not known',))
<Greenlet at 0x1079ef910: <bound method Crawler.process of <__main__.Crawler object at 0x107a0af10>>(u'http://cython.readthedocs.io_sources/index.txt')> failed with ConnectionError
Traceback (most recent call last):
File "crawler.py", line 294, in <module>
main()
File "crawler.py", line 290, in main
crawler.run(result.parallel)
File "crawler.py", line 175, in run
for (url, children) in self.todo:
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/queue.py", line 345, in next
result = self.get()
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/queue.py", line 284, in get
return self.__get_or_peek(self._get, block, timeout)
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/queue.py", line 261, in __get_or_peek
result = waiter.get()
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/hub.py", line 878, in get
return self.hub.switch()
File "/private/tmp/crawler/lib/python2.7/site-packages/gevent/hub.py", line 609, in switch
return greenlet.switch(self)
gevent.hub.LoopExit: ('This operation would block forever', <Hub at 0x1075d3eb0 select default pending=0 ref=0 resolver=<gevent.resolver_thread.Resolver at 0x107a16ed0 pool=<ThreadPool at 0x107126990 0/4/10>> threadpool=<ThreadPool at 0x107126990 0/4/10>>)
macbook-smee:crawler smee$
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment