0

This website has links to a bunch of sub-pages like /zx-origins/sub-page, and those pages in turn have a link to a .zip file on a sub-domain. My goal is to get those .zip files in one nice command, I thought it would be a lot simpler than it turned out to be.

First of all I have made sure that the links are all server side rendered by using curl and grep, so I know the links exist before Javascript. I also have JS disabled in my browser.

The only inconsistency I see is that the final link to the .zip file has no protocol, just //download.damieng.com...

Is this a bug or a feature in wget not allowing me to follow links without a protocol?

My reasoning so far was that it should be relatively simple to do with wget's -m (mirror) and -A (accept list).

wget -m -e robots=off -D download.damieng.com,damieng.com -A .zip https://damieng.com/typography/zx-origins/

I've tried running it with -d (debug) and it doesn't even list the paths where the .zip file is linked, making me even more confused. I hope some wget ninja can enlighten me.

2
  • If your using wget on a https link you have to authenticate that secure https client. Your command doesn’t do that
    – Ramhound
    Commented Sep 7, 2023 at 1:32
  • @Ramhound what does that mean in practice? Commented Sep 7, 2023 at 8:08

0

You must log in to answer this question.

Browse other questions tagged .