I've tried many online tools in order to scrape an entire website but none of them allow me to freely download all pages.
Next i found the
webhttrack tool for Windows/Ubuntu/Mac, it works well but needs installation.
The final tested solution is a single line command...
➜ ~ wget \ --recursive \ --no-clobber \ --page-requisites \ --html-extension \ --convert-links \ --restrict-file-names=windows \ --domains target.com \ --no-parent \ https://target.com/shop