It's A Digital Disease!

23 readers
1 users here now

This is a sub that aims at bringing data hoarders together to share their passion with like minded people.

founded 2 years ago
MODERATORS
1076
 
 
The original post: /r/datahoarder by /u/msgenhances on 2025-06-07 16:26:56.

hello. My Softraid Raid 5 setup suddenly stopped giving me access to the data and now it's showing up as it's missing a Disk.

  1. observed degraded performance and couldn't write on certain folder structure
  2. Reboot
  3. Power cycle of the enclosure
  4. Validation with repair
  5. Windows Disk error check
  6. Reseated the Disk 2
  7. Changed to a new drive for Disk 2

Did everything above but all I see is drive letter and error saying its not accessible.

Under all drive I see all the drive

Is it worth trying to swap to a new enclosure? possible enclosure issue?

1077
 
 
The original post: /r/datahoarder by /u/kettu92 on 2025-06-07 16:17:53.
1078
 
 
The original post: /r/datahoarder by /u/thomedes on 2025-06-07 16:15:04.

Been using SyncThing and love it.

Up to now I've only used for "small" work. Some dozens of GB and a maximum a 100K files.

Now I'm doubting on wether to trust it for keeping replicas of may main disc, a few TB and file count of a million, maybe two.

Have you used it for something similar? What is your experience?

And the big question: What about security? Would you trust all your files to it?

1079
 
 
The original post: /r/datahoarder by /u/BobDaSloth180 on 2025-06-07 15:53:19.

I was seeing a lot of a brand called UnionSine. Is this brand trusted?

1080
 
 
The original post: /r/datahoarder by /u/voidsyourwarranties on 2025-06-07 15:44:27.

Looking for an inepensive high-capacity nvme, but not sure if non-branded used drives like this are worth the low cost.

1081
 
 
The original post: /r/datahoarder by /u/axebulb_Alex on 2025-06-07 15:18:49.

Last week, I bought a 2TB HDD from ebay so I could back up a load of my old photos. When I plugged it in, I found it had loads of Nintendo games on it. I was going to wipe the drive but what wondering what (legally) can I do with them? Should I message the seller? I don't own any Nintendo consoles myself.

https://preview.redd.it/61eampe5vi5f1.jpg?width=1280&format=pjpg&auto=webp&s=51face2e4f5670bde549ad50d6ba57b7b8c3fd72

1082
 
 
The original post: /r/datahoarder by /u/jugendabest on 2025-06-07 14:45:50.

Hello everyone,

I don't know if it is the right place for such question, but let's go.

I started to do backups of my important files recently and I currently doing it naively.

What I do is that I copy (using cp command) my home folder and other important personnal folder on a HDD drive on my computer, also on an external drive and twice a year I copy that external drive on a home server. For now it works, but with time, the transfer and the copy will start to take more time.

But is it the correct way ? I mean is the "blind" copy/paste a correct way to keep folders/files ? Is there a best and faster way to do it ?

For information, I don't need to do snapshots of my system, just keep my important config files and personal folders safe.

Thanks all !

1083
 
 
The original post: /r/datahoarder by /u/MarinatedPickachu on 2025-06-07 13:52:12.

I have found myself pondering this topic more than once so I wonder if others have tools that served them well.

In the current case I'm using an exFAT formatted external drive. ExFAT because I need to use it between windows and MacOS (and occasionally Linux) for reading and writing so there doesn't seem to be a good alternative to that.

exFAT is certainly not the most resilient filesystem so I wonder if there are things I can use on top to improve

  1. the detection of data corruption
  2. the prevention of data corruption
  3. the recovering from data corruption

?

For 1 actually a local git repository where every file is an LFS file would be quite well suited as it maintains a merkle tree of file and repository hashes (repositories just being long filenames), so the silent corruption or disappearance of some data could be detected, but git can become cumbersome if used for this purpose and it would also mean having every file stored on disk twice without really making good use of that redundancy.

Are you using any tools to increase the resilience of your data (outside of backups) independent of what the filesystem provides already?

1084
 
 
The original post: /r/datahoarder by /u/Short-Guide4913 on 2025-06-07 07:16:39.

Cant use ytdlp or anything like that for the next 10 ish days and i kind of need this now there was another few posts like this but none of the links from there worked

1085
 
 
The original post: /r/datahoarder by /u/orderdisord on 2025-06-07 06:35:02.

Hi all, I need some recommendations for some sort of optical printers that could safely print onto my discs. I don't wanna just write on them with sharpie when I'm trying to record and burn my grandmother's old VHS tapes onto them before the tapes give out, and I know sharpie could affect the shelf life (Plus i want an excuse to design nice labels for my discs!) any recommendations would be great, especially if they might be easy to get second hand!

1086
 
 
The original post: /r/datahoarder by /u/Equivalent_Host3709 on 2025-06-07 06:19:27.

A have a huge repository of downloaded porn I am looking to clean-up/downsize; specifically, I have a lot of semi-duplicate videos, where I was able to find one source with a really high quality but low bitrate/FPS, others with lower FPS but higher bitrate, low quality but very high bitrate or FPS, etc. etc.

Obviously, I want to keep the versions that can give the best viewing experience and pleasure (no motion blur, skin detail, etc.). I am wondering how the three metrics affect video quality, which is the most important to keep high, and which I should prioritize when deleting duplicates (i.e., should I delete the one with lower FPS, or lower bitrate? Always prefer 1080p to higher FPS? 2160p but low FPS or 720p but high FPS?)...

Some other dilemmas I'm having: 24FPS vs. 30FPS, if/how high bitrate compensates for lower FPS, data rate vs total bitrate, bitrate vs video quality...

I'm a newbie to datahoarding, so try not to get too technical, but feel free to give me more considerations I should take into account.

Edit: also curious about what AV1 and h264 mean, and which file I should download from my source, if I have the option for one or the other.

1087
 
 
The original post: /r/datahoarder by /u/Jman5150mib on 2025-06-07 05:56:00.

I have synology nas and have used shucked drives. Was wonderimg expected lifespans. They are all wd, some are 14tb, 18tb, 20tb, 22 tb and looking into maybe gettimg some 24tb.

Are any of these sizes in a nas like ds1520 or ds1522 have different exoected lifespans. I heard 10tb, 18tb , 20 tb and 24tb are likely to last longer than 14tb, 22 tb but was given no evidence. I was told avearge lifesoand was 3-5 but the longer ones are more like 5. Is all if that bull and they are all likely 3-5 or are some really expected to expire sooner?

Aboit to buy another 5 drives for a dx517 and cocnerned about longevity.

That being said any evidence that some nas or extenders help shorted or lengthen drive life?

Thanks!

1088
 
 
The original post: /r/datahoarder by /u/Dev_was_here on 2025-06-07 02:11:53.

I heard earlier firmware caused the SSD to prematurely die

1089
 
 
The original post: /r/datahoarder by /u/True_Pirate on 2025-06-06 23:08:20.

I have ~20tb of data currently and it is growing. I don’t trust or use cloud storage and am curious what you guys think about it. Here is what I do. I have my primary data hard drives connected to my pc. I keep a full offsite backup at a relatives house scattered across some older drives.

As a redundancy on really important data I have about 250 blank blu rays burned with irreplaceable/harder to replace stuff. These are not M discs and may not be in great shape in a decade but in a worst case scenario they make me feel better.

To keep it all straight I have a bunch of excel spreadsheets that I can reference to see what is stored where. What do you guys think?

1090
 
 
The original post: /r/datahoarder by /u/Friendly_Guard694 on 2025-06-06 22:21:13.

I keep buying these because they're cheap and easily portable with a laptop. I'm very minimalist, my life fits in a suitcase. Is there something bigger and better but also portable?

1091
 
 
The original post: /r/datahoarder by /u/ryerhino on 2025-06-06 20:59:08.

Got tired of duplicate versions eating up space, so I made PlexDeDupe.

It scans your library, shows all duplicates, lets you pick which to keep (largest for quality or smallest for space), then removes the extras. Files go to Recycle Bin, not permanently deleted. Freed up .5TB on my first run.

Free & open source: https://github.com/SabrosoCuy/PlexDeDupe

Requirements:Python 3.6 or higher, Plex Media Server, Plex authentication token (Instructions provided in GUI), PlexAPI Python library (pip install plexapi).

I have not tried this with remote drives as mine are all local but it should work.

PS: I used Claude Opus 4 to help write this.

https://preview.redd.it/r0bpxqb1cd5f1.png?width=890&format=png&auto=webp&s=e9790c9616684b4a6053042654b95d8d66b87c18

1092
 
 
The original post: /r/datahoarder by /u/mikepm07 on 2025-06-06 20:33:39.

Hey, I started a new job recently that has nearly 600TB of video footage, with about 80% of it sitting on hard drives that are over 10 years old and that isn't kept in an alternate location.

It sounds like some of these drives haven't been turned on and verified in three years.

My new boss just requested we come up with some proposals on how we could safely update our storage and protect from hard drive failure.

We have a DAM (Digital Asset Management Tool) that keeps a lot of the footage we need regularly accessible, but I know he won't want to delete any of the 600TB of footage.

What's our best option here?

My thought is just to buy new hard drives and make it a policy to verify each drive once a year. In addition to that, we need to clone the contents of each drive to a backup and keep it at a separate location as a safety precaution.

I think that will be cheaper than a server or NAS type system?

Would love any thoughts from people who operate in this field more than I.

Thank you

1093
 
 
The original post: /r/datahoarder by /u/JamesRitchey on 2025-06-06 20:16:48.

This tutorial is for comparing the contents of 2 folders to confirm they contain the same files, when the filenames, or folder structure are different. This is accomplished by hashing the contents.

Steps:

  • Download Ritchey Hash Directory i2 v2. It's an opensource PHP function I made for hashing directories by treating all the files as part of the input to be hashed.
git clone https://github.com/jamesdanielmarrsritchey/ritchey_hash_directory_i2.git

  • Make a PHP script which uses this function to hash both directories' files, and compare the checksums. To do this, paste the following into "ritchey_hash_directory_i2/custom_script.php" (the file doesn't exist, so you'll need to create it).
<?php
$location = realpath(dirname(__FILE__));

$dir1 = "{$location}/temporary/Example 1"; // Change this!
$dir2 = "{$location}/temporary/Example 1"; // Change this!
$algo = 'sha3-256'; // Optionally, change this. Only select algorithms are supported by the hashing function. For most users 'sha3-256' or 'sha256' should be fine.

require_once $location . '/ritchey_hash_directory_i2_v2.php';
$checksum1 = ritchey_hash_directory_i2_v2($dir1, $algo, FALSE, NULL, TRUE);
$checksum2 = ritchey_hash_directory_i2_v2($dir2, $algo, FALSE, NULL, TRUE);
if (is_string($checksum1) === TRUE && is_string($checksum2) === TRUE){
if ($checksum1 === $checksum2){
echo "Checksums match." . PHP_EOL;
} else {
echo "Checksums differ." . PHP_EOL;
}
} else {
echo "ERROR" . PHP_EOL;
}
?>

(You might need to clean-up the formatting if it doesn't paste nicely)

  • Edit the custom PHP script to have your values for the directories to hash, and the algorithm to use. To do this, change the values of $dir1, $dir2, and $algo.

  • Make any other desired changes (if any) to your script. For example, maybe you want it to display the checksums?

  • Run the script.

cd ritchey_hash_directory_i2 && php custom_script.php && cd -

  • Examine the result. You should get a return that is either "Checksums match." or "Checksums differ.".

Note:

  • The hashing function relies on checksums to decide the order of files for the input when hashing. The order of files for the input impacts the checksum produced. This means collisions between checksums could cause incorrect results, by disrupting the order of the input, so it's advisable to use a strong hashing algorithm, to avoid collisions.

--

There's obviously other ways to do this sort of thing, so please share other programs, scripts you've made, etc. Help save the next person some work :)

EDIT: fixed post formatting

1094
 
 
The original post: /r/datahoarder by /u/D3VEstator on 2025-06-06 19:07:28.

I have bunch of dvds and im debating on if i should rip them because of quality?

The bluerays i rip, but im not sure about dvds in today day in age?

Thoughts

[EDITED]: Thanks for everyone who commented, i will continue to look at these. I will continue my ripping process of tv shows and movies that i know i will watch many times over

1095
 
 
The original post: /r/datahoarder by /u/Arcueid-no-Mikoto on 2025-06-06 18:57:53.

Yesterday I tried to download all their manga DB by simply using HHTrack with "https://www.mangaupdates.com/series" URL, as all manga are within. Before I went to bed it had scanned 70k+ links and the folder was 9GB size. It had a ton of the manga pages downloaded, but when I woke up, it said "Task Finished" and most files were deleted from the folder, and now it's 2GB size with most folders empty.

Any idea why would it delete what it downloaded?

Also, I'm new with HHTrack or downloading sites at all. Any reliable way to download their full Manga DB? I'd love if I could use their advanced search offline.

Conveniently, both all the manga and advanced search are behind the /series, so downloading this url succefully should make it work right?

This is the advanced search URL:

https://www.mangaupdates.com/series/advanced-search

And this any random manga:

https://www.mangaupdates.com/series/ygablqw/tsugumomo

How would you go about this? Should I keep using HHtrack or is there a more suitable program? I'd love to know if there's any configuration option I'm missing and should add for this task.

Thanks!

1096
 
 
The original post: /r/datahoarder by /u/TheRealFutaFutaTrump on 2025-06-06 18:17:50.

I have a couple Sony Hi8 tapes (no camcorder) and some mini VHS. I know they make those boxes you can plug into a VCR for the VHS tapes, then you plug into a capture device and pray it works.

Is there a solution that does all of it? Both types of tape plus the capture? I have a pretty awesome computer, but nothing but the tapes (no VCR, no camcorders at all.) I could probably hire a service for as much to do it but I would prefer to screen the tapes myself.

1097
 
 
The original post: /r/datahoarder by /u/sprfreek on 2025-06-06 18:00:00.
1098
 
 
The original post: /r/datahoarder by /u/nando1969 on 2025-06-06 17:49:10.
1099
 
 
The original post: /r/datahoarder by /u/Difficult-Stuff-9800 on 2025-06-06 17:35:15.

I’ve been using GoodSync to back up my files to the cloud, and I’ve enabled encryption for both file content and names. It’s great for security, but I’m worried about what happens if GoodSync terminates its service one day. How would I decrypt my files without their software?

I noticed that if I only encrypt the file content (not the names), I can decrypt those files using 7zip. I’m concerned that GoodSync could change their encryption method in the future, leaving my data inaccessible.

Does Syncbackpro provide all features of goodsync? I noticed that its doc mentioned can be decrypted by zip.

1100
 
 
The original post: /r/datahoarder by /u/wade-wei on 2025-06-06 17:17:10.

We have plans to decommission a Dell SCv2080 storage with 8T SAS drives. I am thinking of putting a few of those drives on my own for-fun server, but I heard these Compellent drives may have different firmware. Can I use them directly on Dell R730/740 servers or do I need to reformat from (528b to 512b?), or do I have to flash some sort of normal firmware onto it?

view more: ‹ prev next ›