Open andrewdbate opened 2 years ago
Did you interrupt the command? URLs are replaced when all the pages have been crawled.
No I didn't interrupt the command.
Hi @gildas-lormeau! First, I'd like to appreciate for this amazing extension.
I faced the very same issue @andrewdbate discussed.
I tested https://xmrig.com because of its simple hierarchy.
Following internal links on https://xmrig.com should be considered:
--filename-conflict-action=skip
flag.This is the command I ran:
./single-file --output-directory=saved --filename-template="{url-pathname-flat}.html" --crawl-links=true --crawl-replace-urls=true --filename-conflict-action=skip https://xmrig.com
As the result, following files were created inside saved
directory (as expected):
Everything has been well so far but links inside these files are not changed to relative links on file system.
You may find these files useful:
Thanks
I'm having the same issue, each web page is downloaded successfully, but the links between them link back to the original website:
docker run -v $(pwd):/usr/src/app/out singlefile "https://fiction.live/stories/Fiction-live-Software-Update/S46jksooFQWqqMAsY/home" --dump-content=false --crawl-links=true --crawl-inner-links-only=true --crawl-no-parent=true --crawl-max-depth=1 --crawl-replace-urls=true
When I run this command:
none of the files in the
outdir
directory have URLs of saved pages replaced with relative paths of other saved pages inoutdir
.When I run this command,
_wiki_Thomas_Lipton.html
is downloaded tooutdir
. This is the file of URL from which the crawl started.The Wikipedia page https://en.wikipedia.org/wiki/Thomas_Lipton has a link to https://en.wikipedia.org/wiki/Self-made_man in the first sentence. This page was also downloaded by SingleFile as
_wiki_Self-made_man.html
.I was expecting the
href
tohttps://en.wikipedia.org/wiki/Self-made_man
in_wiki_Thomas_Lipton.html
to be rewritten to_wiki_Self-made_man.html
but it was not. Am I using the CLI options incorrectly?