I've tried many online tools in order to scrape an entire website but none of them allow me to freely download all pages.

Next i found the webhttrack tool for Windows/Ubuntu/Mac, it works well but needs installation.

The final tested solution is a single line command...

āžœ  ~ wget \                       
     --recursive \
     --no-clobber \
     --page-requisites \
     --html-extension \
     --convert-links \
     --restrict-file-names=windows \
     --domains target.com \
     --no-parent \