I posted this on stack overflow and they called me stupid (I am) and locked the question when someone was in the middle of actually helping me so I'm asking here because it's probably a more welcoming community for this sort of thing.
https://prcm.jp/list/akb48%20%E3%83%97%E3%83%AA
So this (formerly) huge Japanese image sharing site is shutting down in 2 weeks and my sister's begging me to archive at least some of it. I was trying to find a python script to automate it, cuz it will take me forever going one by one on tens of thousands of images, but I can't seem to find one that can archive this sort of website. It only goes by pages of 9 and the thumbnail images are shrunk down a ton so you have to click on the image twice to open the full size. Luckily it seems like every thumbnail image shares a name with the source image, but with an added suffix of the preview size, so I imagine it would be possible to have the script delete the underscore and re-add the .jpeg extension. As for going through pages, I'm not so sure on, but if I can input a list of URLs and just batch copy the original URL adding every page number (I know how to do this without a script), I could just use that.
I was only going to download AKB48 related images, because most of these images aren't available anywhere else online, and soon won't be anywhere period. They're all old and thus fairly small, so I'm not too worried about it taking a long time to download.
If anyone can direct me to a script that can do this or has some other mass image downloading method you're aware of, please let me know. Thanks!