It's A Digital Disease!

23 readers
1 users here now

This is a sub that aims at bringing data hoarders together to share their passion with like minded people.

founded 2 years ago
MODERATORS
3376
 
 
The original post: /r/datahoarder by /u/s4lt3d_h4sh on 2025-02-17 13:01:55.

Hey fellow datahoarders!

I recently picked up a Sans Digital TR8U-B (the 8-bay USB DAS), but it’s not being recognized on any PC I connect it to—Windows, Linux, nothing sees it at all. I’ve tried different cables, ports, machines, you name it. It’s looking like the internal controller board is probably dead (picture of the board for reference)

Rather than scrapping the enclosure, I’m thinking of buying an M.2 USB-C enclosure (like this https://www.aliexpress.com/item/1005006917952495.html) and pairing it with an M.2-to-SATA adapter to hook up the original backplane. I’m basically aiming to bypass the dead board and repurpose the chassis for SATA drives.

Has anyone here done a similar retrofit or have any suggestions on the best approach (or different hardware options) to get this working as a standalone DAS again? Any tips or pitfalls I should watch out for?

Thanks in advance for any advice

3377
 
 
The original post: /r/datahoarder by /u/cyrilmezza on 2025-02-17 09:52:18.

I've been the unofficial photographer for my family for years (film then digital, so I'm old...), have tons of files and keep them mostly to myself (I would regularly transfer some to my parents' computer while they were alive and had desktops), but occasionally I leave some on a shared drive. Now, with my sisters and their children, everyone has a laptop or a smartphone. I'm not the only one who can take decent pictures, and beside the occasional one sent via Whatsapp and Co., there's not much sharing from their side.

My idea is to create a shared repository, where anyone could easily drop their favorite pictures of themselves or siblings, so that we could use as slideshows (alexa), prints and whatnot. No need for something 100% automated, I can deal occasionally with updating the folders for the Alexas and Google homes, the main thing is a repository accessible from inside the LAN, and from the palm of their hands (mostly Android) wherever they are (some have left the nest, and not everyone lives under the same roof)

I use Syncthing on my own phone with great success, but I would not suggest that to my nieces and nephews; ideally they would share one or more pictures of their choosing right from their phone (or laptop), with as little friction as possible. They are not dumb or tech illiterate, I'd just like to make it simple.

What tools do I already have ? 2 Synology NASes (2 homes with Fiber internet, that I manage myself, including firewalls) with TBs of free space, no yet open to the outside world, OneDrive account with 1TB, Amazon Photos (Prime), and free Google accounts.

I'm looking for a self hosted, or at least self managed (cloud/hybrid) solution, but I don't mind paying a little something (perpetual rather than subscription to a new service) for a commercial solution if there's a great App for that out there...

3378
 
 
The original post: /r/datahoarder by /u/Organic-Tie-4265 on 2025-02-17 09:22:13.

I wanted to share the extremely beginner steps I’ve taken to start collecting for anyone else that’s brand new. This is specifically oriented towards people interested in physical copies of data.

The title is not an exaggeration. I hate technology and I went 5 years without ever touching my computer. When I started looking into this two weeks ago I literally had to search things like “server,” “operating system,” and “gigabyte.” Due to recent events I’ve been drawn to archiving and locally storing data (this will help me be even more offline lol). I also don’t have a lot of money so, thanks to Reddit, I worked out the cheapest way to begin a tiny, clumsy collection and I hope this might help someone else.

My primary goal was to come up with a way to reliably store physical copies of data outside of a cloud service. That means choosing two places to store the important stuff. The hard drives will serve as primary storage, because they can concentrate a lot more data, and the discs will be a fail-safe for backing up anything I don’t want to lose. I opted for two different formats of data storage just because I don’t know how affordable hard drives will be in the future and, while they’re inconvenient, blu-ray discs are reliable and (ime) generally cheaper per gigabyte. I went on Ebay and searched for open box/refurbished to save money.

Total: $222

  • BD-R Blu-ray discs (25 for $30): I was told to specifically get BD-R and not BD+R because those are lower quality or something idk. I got the Verbatim brand.
  • Blu-ray burner ($106): For transferring data onto the discs. I got the Buffalo brand.
  • Hard drive enclosure ($21): For transferring data onto hard drives. I got the Orico brand.
  • 1 TB Hard drive ($40): I was told that HDD/SSD isn’t too important to differentiate however you want to make sure whatever you’re buying is compatible with the enclosure. I got a 3.5” Seagate Firecuda.
  • (optional) CD cases (40 for $25): Keep those puppies safe and dust-free!

If you have more money to burn and/or comfort with technology there are absolutely better and easier ways to go about collecting (see: this whole subreddit). For instance, there are easy ways to backup hard drives onto other hard drives (something something RAID idk don’t ask me). As I mentioned, price is a major inhibiter for me when it comes to hard drives so this may be a more affordable set-up for anyone that has patience to manually burn discs. It’s a very basic method for someone like me who’s intimidated by like, the task of converting a file lol and so far it’s working out great for my simple needs!

3379
 
 
The original post: /r/datahoarder by /u/Aggressive_Limit_657 on 2025-02-17 07:20:45.

I am trying to mirror CNET website upto depth 1 using this command - httrack "https://www.cnet.com/" -s0 -r1

But the issue is I cannot get all the images and js for that site. Any solution to this issue?

3380
 
 
The original post: /r/datahoarder by /u/wolf2002x on 2025-02-17 14:37:30.

Original Title: For the 3-2-1 rule, do I have to apply that same rule again for the back up drives or is that doing too much? But then I think like in the year 2030, those back up drives will probably stop working aswel, so I'm not sure, I just don't want all my data to be lost. thank you :)

3381
 
 
The original post: /r/datahoarder by /u/antidumb on 2025-02-17 14:34:48.

Hey, all. I'm missing something here and I'm not sure what it is. I have a script that should be working to run 10 docker instances on Windows and populating the info necessary. Something's going wrong somewhere and I can't figure out what I'm missing. It's not populating the downloader name, project, download number, and number of rsync thread. I'm not even sure this is the correct sub, but I hope it is. Script is as follows:

@echo off

setlocal enabledelayedexpansion

REM Run Watchtower to automatically update containers

docker run --detach ^

--name watchtower ^

--restart=on-failure ^

--volume /var/run/docker.sock:/var/run/docker.sock ^

containrrr/watchtower --label-enable --cleanup --interval 3600

REM Run 10 instances of ArchiveTeam Warrior with custom settings

set NUM_INSTANCES=10

set CONTAINER_NAME="archiveteam_warrior"

set IMAGE="atdr.meo.ws/archiveteam/warrior-dockerfile"

set PROJECT_NAME=auto

set DOWNLOADER=antidumb

set RSYNC_THREADS=20

set DOWNLOADS=6

REM Ensure Docker is running

docker info >nul 2>&1

if %errorlevel% neq 0 (

echo Docker is not running. Please start Docker and try again.

exit /b 1

)

REM Pull the latest ArchiveTeam Warrior image

echo Pulling the latest ArchiveTeam Warrior Docker image...

docker pull !IMAGE! || (

echo Failed to pull Docker image. Exiting.

exit /b 1

)

echo Docker image pulled successfully.

REM Create and start 10 Docker containers with unique ports and names

for /L %%i in (1,1,!NUM_INSTANCES!) do (

REM Set variables specific to each instance

set /a PORT=8000+%%i

set CONTAINER=!CONTAINER_NAME!_%%i

REM Echo the container start command for debugging

echo Starting container: !CONTAINER! on port !PORT!

echo docker run -d --name !CONTAINER! -p !PORT!:8001 !IMAGE! --project "!PROJECT_NAME!" --downloader "!DOWNLOADER!" --concurrent "!DOWNLOADS!" --rsync-threads "!RSYNC_THREADS!"

REM Run the container

docker run -d ^

--name !CONTAINER! ^

--restart unless-stopped ^

-p !PORT!:8001 ^

!IMAGE! ^

--project "!PROJECT_NAME!" ^

--downloader "!DOWNLOADER!" ^

--concurrent "!DOWNLOADS!" ^

--rsync-threads "!RSYNC_THREADS!"

REM Check if the container started successfully

if !errorlevel! equ 0 (

echo !CONTAINER! started on port !PORT!.

) else (

echo Failed to start !CONTAINER!.

REM Check the logs to understand why the container didn't start

docker logs !CONTAINER!

)

)

echo All !NUM_INSTANCES! ArchiveTeam Warrior instances have been started.

I've tried the variables mentioned above both with and without quotation marks, neither works as expected.

I have a similar bash script for my Macs and Linux systems, but the Windows one is messing me up.

Thanks!

3382
 
 
The original post: /r/datahoarder by /u/dirtybluper on 2025-02-17 14:15:40.

Hi everyone,

I run a small architectural visualization business, and in the years since, I have collected a lot of 3D models and textures, totaling around 13TB across different disks. Now, I am running out of space, and I want to set up a good storage system that can guarantee I can get my data back if something happens. Although I can redownload the assets from the publishers if the worst happens, I have spent countless hours organizing these models, and doing that again is unthinkable for me. I am kinda on a budget but I do want to set a good foundation.

I am the only person using this storage, and I am almost always in one location and on one computer, so I don't really need it to be accessible from a network for now. Also, I only write data to disk once and the read rate is around twice a week at the maximum, and performance is not important.

So here is my plan: I was thinking of getting 3 X18 14TB drives with a DAS enclosure and maybe setting up the DAS's RAID(5) functionality. I also want to get a Backblaze personal or Crashplane professional subscription to have a second version online in case of fire, flood, or theft. And if I need to expand or need network availability later, I will migrate my system to an Unraid NAS.

What do you think?

Any feedback is appreciated.

3383
 
 
The original post: /r/datahoarder by /u/Marchellok on 2025-02-17 13:32:56.

Hello i want to watch alien romulus. It came in 2.4:1 and IMAX (1.9:1 or sth) so while i got 16:9 the imax is much closer and almost all of pixels shoulkd be used that way.

Are IMAX versions of movie rips available? Is it made that way that i get more screen used , like 3840x2160 with maybe just tiny little bars on top and bottom or sides? Im really strolling in the mist here since im new in that area so i could use any help. Also let me know if i am not specific enough since english isnt my mother language

3384
 
 
The original post: /r/datahoarder by /u/KaleidoscopeReady161 on 2025-02-17 12:29:27.

This post and many other sources mention if an SSD is unpowered, the data may in some cases last mere months before bits start to flip: https://www.reddit.com/r/DataHoarder/comments/ba8o0b/will_ssd_lose_data_if_left_unpowered_for_extended/ I'm interested in this use case for an off-site backup that is unpowered.

Now that left me wondering, to do a best effort approach of mostly preventing that outside of the usual other risks like controller failure and so on, what is a likely sufficient strategy? Is it enough to simply plug it into a machine and mount it for a few minutes, twice a year? Would that "restore" whatever gradually less powered bit state all of the cells on there might have had? Or do I need to rewrite the data that is already on there, or do something else specific?

Summed up, I don't understand enough about SSDs to understand what's likely a somewhat sufficient strategy here. And I was curious if somebody here might know how this works.

3385
 
 
The original post: /r/datahoarder by /u/dohat34 on 2025-02-17 12:23:36.

I bought this startech X000W69PS1 and surprise md to see it didn’t include a power cable. The manual states no power cable is needed, and it simply uses power from the USB 3.0 port. I am actually only going to connect to a USB 2.0 port and I’m going to connect both spinning hard drives as well as SSD. Does anybody see this as a potential problem and should I rather get an adapter with power? I just have to secure wipe clean 10 hard drives and get rid of them.

3386
 
 
The original post: /r/datahoarder by /u/new2bay on 2025-02-17 11:53:59.
3387
 
 
The original post: /r/datahoarder by /u/ll777 on 2025-02-17 11:24:55.

Thus guaranteeing any article you read at any point is available offline.

Thoughts ?

3388
 
 
The original post: /r/datahoarder by /u/Pheonix412 on 2025-02-17 09:05:19.

Hi all, Just finally got most of the parts I need for my Unraid NAS and I was hoping for a bit of help with the last part, the storage. I am in Australia and am looking to buy 4x16tb hard drives and found some Refurbished WD Ultrastar DC HC550's for ~$320ish AUD but I also saw they had some lightly used Seagate EXOS X18 for $370 AUD. I am hoping you can give me a hand with the decision as I was thinking about getting 2 of each but was wanting to know if the EXOS X18's would need to be modded to stop the 3.3V reset as I have not found anything talking about it and I did not see anything on seagate's spec sheet. So yeah

  1. Should I get 2 of each to save some money and also get a few drives that would last longer?

2.1 Does the Seagate EXOS X18 16TB ST16000NM000J have the power disable (PWDIS) feature?

2.2 If it does have this feature how should I deal with it?

  1. Has anyone used Neology before? if so how was the experience?

Thank you for your time.

WD HC550 - https://neology.com.au/products/u-wuh721816ale6l4?variant=47026506760514

EXOS X18 - https://neology.com.au/products/an-st16000nm000j?variant=50224330735938

3389
 
 
The original post: /r/datahoarder by /u/theoldgaming on 2025-02-17 08:54:06.

Pretty much title.

I would be interested in particular in HDDs, SSDs, CDs, SDs, MicroSDs, CFexpress cards and whatever else there is.

Not fast, just really and i mean REALLY reliable.

3390
 
 
The original post: /r/datahoarder by /u/SomeRetard- on 2025-02-17 08:27:27.

A large tracker featuring mostly Japanese content is going to be shut down. As a result many torrents of niche content and original TV broadcasts will disappear within 28 or so days. Free invites will be provided to anyone who wants to help archive this tracker and download anything they want. Please hurry.

If anyone requires an invite, they will have to have an email. A burner is fine as long as it receives emails.

3391
 
 
The original post: /r/datahoarder by /u/DJ_Minikelk on 2025-02-17 08:17:23.

Hi there, I have a Synology rs814+ NAS and I need to upgrade all drives in the system because they are getting old. I have tried to find out what restrictions there are on the NAS and i have found a datasheet specifying that the max. internal capacity is 20TB spread out over 4 slots. however it also states: "The actual capacity will differ according to volume types". We are using Synologys hybrid raid and since the storage capacity is limited by the smallest drive, does this mean that I can install above 5TB pr. slot or am i still limited by the total volume of the installed drives

Im very new to this and therefore dont know what im doing, sorry.

3392
 
 
The original post: /r/datahoarder by /u/LocalIdiot227 on 2025-02-17 07:12:12.

So I was downloading files to an external drive via web browser. Suddenly the web browser stopped responding and my PC froze. So I had to perform a forced shutdown. After restarting, my external drive no longer showed in "This PC". So I check Disk Management and the drive is there but is not listed as "Not Initialized" and its space is "Unallocated". So before I panic, I'd like to ask about what people here would recommend for rather repairing the drive if possible and saving the data or salvaging what data I can off of it.

3393
 
 
The original post: /r/datahoarder by /u/PuzzleheadedFly3699 on 2025-02-17 03:21:15.

Disclaimer: I am new to large scale data hoarding, and so I have come to this internet cave for help from you wizards.

I am starting a project for my PhD where I need to store upwards of 40 TB worth of data for training a predictive model. It should ideally be expandable. What would you all suggest for this sort of setup? I am needing to build a new desktop setup for this project anyways, so would internal or external storage be best? SSD or HDD? What brands would be best and what models of products?

3394
 
 
The original post: /r/datahoarder by /u/ericlindellnyc on 2025-02-17 05:29:07.

I am using MacOS, so that means BSD linux. The problem is I pipe results of find into rmlint, and the filtering criterion is ignored. find . -type f -iname '.png' | rmlint -xbgev This command will pipe all files in current directory into rmlint -- both PNGs and non-PNGs. If I pipe the selected files to ls, I get the same thing -- PNGs and non-PNGs. When I use exec find . -type f -iname '.png' -exec echo {} ; This works to echo only PNGs, filtering out non-PNGs. But if I pipe the results of exec, I get the same problem -- both PNGs and non-PNGs. find . -type f -iname '*.png' -exec echo {} ; | ls This is hard to believe, but that's what happened. Anybody have suggestions? I am deduplicating millions of files on a 14TB drive. Using MacOS Monterey on a 2015 iMac. Thanks in advance PS I just realized by ubuntu is doing the same thing -- failing to filter by given criteria

3395
 
 
The original post: /r/datahoarder by /u/Ok-Respond-600 on 2025-02-17 05:16:53.

My Dad has been a professional photographer for over 50 y but has kind of lost track of archiving in the post CD era. He has hundreds of thousands of photos on loads of HDDs that I want to put into a database

Which software can I use to catalogue them all into a searchable database where there is previews but the original files stay where they are. Something that can be accessed online from any device.

Thanks for any help

3396
 
 
The original post: /r/datahoarder by /u/KJSS3 on 2025-02-17 05:09:12.
3397
 
 
The original post: /r/datahoarder by /u/cyrilio on 2025-02-17 04:54:06.

TL;DR reddit still hosts tons of images people have uploaded to subreddits that are now banned.

While I'm not a massive hoarder of data I do have a decent collection of books and research papers on my PC (22k+ and rising). Really love the data hoarding mentality and with my new PC upgrade I'm definitely ensuring plenty of additional storage space.

I moderate a couple dozen subreddits. Sadly a few have been banned throughout the years. One I really like is r/drugstashes. Interestingly enough while the sub itself is inaccessible. At least a large chunk of all images people uploaded to that subreddit are still hosted by reddit and accessible without any shenanigans or hard workaround. You can use the Reddit Archive or PushShift search sites) as if there's no ban at all. Images hosted elsewhere are obviously still acessible. Unless OP deleted their account.

See this for example: https://i.redd.it/qzdhq20k2tg31.jpg

Using Reddit Archive you can find the original image post at the 9th place or so. [NSFW warning: discussion about and images of drugs are visible, there is no nudity, gore, or violence visible]


Immediatelijk after the ban happened I already scraped every image with the help of a friend. I'm sitting on 7k + images, about 3 GB. Haven't thought of a final solution for permanent static 'museum' site that's accessible to anyone. Perhaps there will be some exceptions. Like adding a YES/NO pop-up asking to verify 18+ age before being able to see the images.

Is this common knowledge? Do you mod a banned subreddit and want to save any data/images that where uploaded and can't be reached through normal ways? Now's your change to at least recover some of it. Until reddit admins decide to close the loophole (for advertising reasons probably).


Hope many can benefit from this. Would love to see how you guys wull use this super sloppy reddit 'fix'.

3398
 
 
The original post: /r/datahoarder by /u/coolestbean4ever on 2025-02-17 03:49:27.

Hey fellow data hoarders,

I’m looking for the best SSD options for long-term data storage and live access. I have 10TB+ of data total (8K/12K videos, high-res photos, documents, archives), I need reliable, high-speed SSDs for:

  • Live access to large files without slowdowns
  • Longevity & durability (minimal risk of data degradation)
  • High capacity
  • Fast speeds for editing & playback
  • Possibility of 6 months cold storage

Right now, I’m considering:

  • SanDisk Extreme Pro – good speeds, but is it reliable long term?
3399
 
 
The original post: /r/datahoarder by /u/Ok_Tip3706 on 2025-02-17 03:46:11.

file share* not server

The only things I find online are just samba shares and ancient programs that only work well (or at all) for windows.

Looking for linux to linux software. Samba constantly tells me my permissions are wrong when they are not. I have to keep restarting my file manager to get it to see it has permissions.

3400
 
 
The original post: /r/datahoarder by /u/Vatican87 on 2025-02-17 03:41:58.

I haven't been keeping up for a while, but I currently have 4 EXO's that I purchased from Serverpartdeals around 5 years ago and they were 18TB each for <$600 total I believe. I just got a new NAS 6-bay from UGREEN and I'm looking to fill this baby up with 20TB or higher, but the prices have gone up drastically? What are you all doing for some good recertified drives? These will be mainly used in raid config.

view more: ‹ prev next ›