You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Replace URLs of saved pages with relative paths of saved pages on the filesystem
So if I understand correctly the HTML extracted by single-file should have all its URLs crawled with the option --crawl-links replaced by the file path on which they are exported.
However, when I try this command I get only the original URLs:
In the first example, there are no inner links. The second example does not work anymore (I'm pretty sure it used to work in the past) because there are no link with a resolved URL starting with "https://www.wikipedia.org/" in the page.
The option
--crawl-replace-urls
indicates:So if I understand correctly the HTML extracted by
single-file
should have all its URLs crawled with the option--crawl-links
replaced by the file path on which they are exported.However, when I try this command I get only the original URLs:
I also tried this command from the README using the option
--crawl-rewrite-rule
but it did not work either:./single-file-x86_64-linux https://www.wikipedia.org --crawl-links=true --crawl-inner-links-only=true --crawl-max-depth=1 --crawl-rewrite-rule="^(.*)\\?.*$ $1"
I was able to make it work on v2.0.0 but not since v2.0.2.
The text was updated successfully, but these errors were encountered: