Last post by The_yoyo - July 10, 2022, 05:44:57 PM
Bump this. I have a very similair issue: The website is fairly basic structure, but slightly complex URI's, like: www.example.com/m/?Cat=Selection&V_Sub=X&Page=1&SortBy=Z and www.example.com/m/?Cat=Selection&V_Sub=X&Page=2&SortBy=Z
I limited the crawl length from root so as to limit the page range (no sure if it worked). Anyway, the files I get are named: Index.htm Index-1.htm Index-2.htm ... Index-45678.htm
While the HREF inside the htm's refer to the URI's without the index, like: C/DownloadedWebsites/example.com/m/?Cat=Selection&V_Sub=X&Page=2&SortBy=Z This means the internal links don't work (404's).
Last post by netzteil - June 02, 2022, 08:46:35 AM
is there any possibility to use the CLI version and specify both, 1) a project file for all rules /settings etc. AND 2) an external argument list file that contains "uri" AND "/o" arguments?
I want to automate loading lots of different URLs, all do different output folders, which I want to define in an argument list file, but still use one project file with its settings to download all of them.
Right now, the documentation says "WCOPY uri | project [OPTION...]". So I should be able to define a project file AND still use the uri parameter to override the project's uri... So this option shouldn't be mutual exclusive...
I have searched the forum and the documentation and not been able to find an answer to this question - apologies if I just haven't found it. Is there any way to resume an interrupted copying process without starting from the beginning again (assuming the process has not been completed once, so it's not like an updated website)? My internet connection is not entirely reliable to run uninterrupted for the several days it appears will be taken to complete website in question.