this post was submitted on 13 Jun 2024
1 points (100.0% liked)

It's A Digital Disease!

23 readers
1 users here now

This is a sub that aims at bringing data hoarders together to share their passion with like minded people.

founded 2 years ago
MODERATORS
 
The original post: /r/datahoarder by /u/Mode7GFX on 2024-06-13 03:30:49.

I posted this on stack overflow and they called me stupid (I am) and locked the question when someone was in the middle of actually helping me so I'm asking here because it's probably a more welcoming community for this sort of thing.

https://prcm.jp/list/akb48%20%E3%83%97%E3%83%AA

So this (formerly) huge Japanese image sharing site is shutting down in 2 weeks and my sister's begging me to archive at least some of it. I was trying to find a python script to automate it, cuz it will take me forever going one by one on tens of thousands of images, but I can't seem to find one that can archive this sort of website. It only goes by pages of 9 and the thumbnail images are shrunk down a ton so you have to click on the image twice to open the full size. Luckily it seems like every thumbnail image shares a name with the source image, but with an added suffix of the preview size, so I imagine it would be possible to have the script delete the underscore and re-add the .jpeg extension. As for going through pages, I'm not so sure on, but if I can input a list of URLs and just batch copy the original URL adding every page number (I know how to do this without a script), I could just use that.

I was only going to download AKB48 related images, because most of these images aren't available anywhere else online, and soon won't be anywhere period. They're all old and thus fairly small, so I'm not too worried about it taking a long time to download.

If anyone can direct me to a script that can do this or has some other mass image downloading method you're aware of, please let me know. Thanks!

no comments (yet)
sorted by: hot top controversial new old
there doesn't seem to be anything here