Using wget to clone websites

using wget to clone an entire website for offline access and Markdown conversion

wget is a command-line tool for downloading content from the web. It supports HTTP, HTTPS, and FTP protocols and is widely used for recursively downloading entire websites.

On macOS:

brew install wget

To clone a full website:

wget --mirror --convert-links --adjust-extension --page-requisites --no-parent https://example.com

Explanation of key flags:
• --mirror: Enables recursion and time-stamping for mirroring.
• --convert-links: Adjusts links in downloaded files for local viewing.
• --adjust-extension: Saves files with proper extensions.
• --page-requisites: Downloads all necessary assets (images, CSS, etc.).
• --no-parent: Restricts download to the given directory level, avoiding parent links.

The above command creates a local copy of the website at https://example.com that can be viewed offline from your file system.

22 Apr 2025

wget --mirror --convert-links --adjust-extension --page-requisites --no-parent https://knowledge.kaltura.com/help -P /Users/nic/dl/kaltura-knowledge

✅ worked, with 4,490 HTML files

wget --mirror --convert-links --adjust-extension --page-requisites --no-parent https://kaltura.com -P /Users/nic/dl

links

social