this post was submitted on 30 May 2025
1 points (100.0% liked)

It's A Digital Disease!

23 readers
1 users here now

This is a sub that aims at bringing data hoarders together to share their passion with like minded people.

founded 2 years ago
MODERATORS
 
The original post: /r/datahoarder by /u/Nearby_Relation5021 on 2025-05-29 21:28:37.

I have this simple html site which has indexed many "subdirectories" and files (mostly pdf, some images, zips and txt files)

https://preview.redd.it/c4w01zt1fs3f1.png?width=1053&format=png&auto=webp&s=534b9d83d4a7b0dd693948a5535759922e13f5e3

For some reason it seems that httrack fails quite often to complete many of the downloadable files, giving only partial downloads.

I am certain that some of the files won't download manually via browser but it's just maybe a couple dozen, which is nothing compared to the total (~8k files), the others can be downloaded just fine.

I tried by reducing the number of simultaneous connections and forced them to stay always open, doesn't seem to change anything...

Maybe I should try different software? I don't need to preserve the site, I just need every downloadable file downloaded in its subdirectory.

Any suggestion would be very appreciated

no comments (yet)
sorted by: hot top controversial new old
there doesn't seem to be anything here