Releases: scrapy-plugins/scrapy-zyte-smartproxy
v2.3.5
Ban and throttling responses from Zyte API proxy mode are now handled in line with matching responses from Zyte Smart Proxy Manager.
Full Changelog: https://scrapy-zyte-smartproxy.readthedocs.io/en/latest/news.html
v2.3.4
Use a different stat prefix for Zyte API proxy mode.
Full Changelog: https://scrapy-zyte-smartproxy.readthedocs.io/en/latest/news.html
2.3.3
Fix response handling for Zyte API proxy mode.
Full Changelog: https://scrapy-zyte-smartproxy.readthedocs.io/en/latest/news.html
2.3.2
Detect scenarios where the proxy
Request.meta
key has probably been accidentally copied from an earlier response, warn about it, and fix the value.
The Zyte-Client
header is again sent when using Zyte API proxy mode, now that Zyte API supports it.
Full Changelog: https://scrapy-zyte-smartproxy.readthedocs.io/en/latest/news.html
2.3.1
v2.3.0
v2.2.0
v2.1.0
v2.0.0
Following the upstream rebranding of Crawlera as Zyte Smart Proxy Manager, scrapy-crawlera
has been renamed as scrapy-zyte-smartproxy
, with the following backward-incompatible changes:
- The repository name and Python Package Index (PyPI) name are now
scrapy-zyte-smartproxy
. - Setting prefixes have switched from
CRAWLERA_
toZYTE_SMARTPROXY_
. - Spider attribute prefixes and request meta key prefixes have switched from
crawlera_
tozyte_smartproxy_
. scrapy_crawlera
is nowscrapy_zyte_smartproxy
.CrawleraMiddleware
is nowZyteSmartProxyMiddleware
, and its defaulturl
is nowhttp://proxy.zyte.com:8011
.- Stat prefixes have switched from
crawlera/
tozyte_smartproxy/
. - The online documentation is moving to https://scrapy-zyte-smartproxy.readthedocs.io/
Note: Zyte Smart Proxy Manager headers continue to use the X-Crawlera-
prefix.
- In addition to that, the
X-Crawlera-Client
header is now automatically included in all requests.
v1.7.0
- Added more stats to better understanding the internal states.
- Log warning when using
https://
protocol. - Add default
http://
protocol in case of none provided, and log warning about it. - Fix duplicated request when the response is not from crawlera, this was causing an infinite loop of retries when
dont_filter=True
.