A Web Cache warmer to parse one or many websites based on one or multiple XML sitemaps.
Written in Python 3.7 but works for any website (as long as a XML sitemap is available).
Based on Bitnami's Python Docker image for convenience and small image size ;).
usage: cache-warmer.py [-h] [-t THREADS] [-u URL] [-f FILE] [-v]
Cache crawler based on a sitemap.xml URL (multiple URL are supported) or file
(only one file supported)
optional arguments:
-h, --help show this help message and exit
-t THREADS, --threads THREADS
How many threads to use
-u URL, --url URL The sitemap xml url
-f FILE, --file FILE The sitemap xml file
-v, --verbose Be more verbose
Use the -u
option to pass an URL.
Separate multiple URL with commas: -u url1,url2,url3
docker run kissmy/cache-warmer \
-t 4 -v -u https://example.com/sitemapFR.xml,https://example.com/sitemapEN.xml,https://example.com/sitemapNL.xml
Use the -f
option to pass a local file.
docker run kissmy/cache-warmer -t 1 -v -f /tmp/sitemap.xml
Multiple local files are not supported, only multiple urls as we don't have any use case for multiple files.
Feel free to submit a PR or create an issue if you need this feature.
Check our Docker Hub page for available tags. There's always a latest
tag available that points to the master branch
PR and enhancement ideas welcome.
@superbiche migration to Python 3.7, add multiple URL support, Docker build and current maintainer.
Thanks to Hypernode Support for the initial Gist that made our day.
MIT