It's A Digital Disease!

18 readers
3 users here now

This is a sub that aims at bringing data hoarders together to share their passion with like minded people.

founded 2 years ago
MODERATORS
151
 
 
The original post: /r/datahoarder by /u/luigisbiggreenpipe on 2025-02-16 04:17:20.

I’d personally like to start digitally archiving important educational texts for preservation and personal reference for the future, but I don’t really know where to start and what the important texts would be. I’m talking all aspects of science from elementary to graduate PhD level, including medical reference texts, all subjects of history and literature, the works.

Where can we get information on what these texts are and where can I start to get digital copies of them to archive and, in the future, if necessary, distribute.

For now, I want to just start with step one in figuring out what these texts would be, then figuring out where to get digital copies or how to digitize them myself.

It would be nice to have a collaborative effort on this, but if I could be pointed in the right direction to start on my own, I’d appreciate it.

If anyone is interested in collaborating, we could start a discord on this subject.

152
 
 
The original post: /r/datahoarder by /u/Hunkamunkawoogywoo on 2025-02-16 04:13:06.

Obviously I'm trying to download every video from a channel. 1.4k videos, so one at a timing it isn't going to cut it

153
 
 
The original post: /r/datahoarder by /u/CliffJump_BlackRocks on 2025-02-16 03:15:56.
154
 
 
The original post: /r/datahoarder by /u/AutomaticDriver5882 on 2025-02-16 01:42:10.

Is there any archiving service that stores transcripts of interviews speeches etc. A lot of things are said these days. A lot of people and politicians seem to have amnesia so I’d like to use it as a research platform.

155
 
 
The original post: /r/datahoarder by /u/That-Interaction-45 on 2025-02-16 01:02:36.

Hey team, I picked up a 20 TB used drive from goharderive, but was surprised when I looked up badblocks to see it's Linux only.

Is just a full format via windows 10 built in tool enough? Would you recommend a different tool?

Thanks!

156
 
 
The original post: /r/datahoarder by /u/d2racing911 on 2025-02-16 00:22:15.

Hi everyone, I would like to know if that version is still safe to use on Windows 11 24H2 ?

I'm against subscription and I don't plan to pay for version X, since I have 2 PCS.

I'm not in a good financial situation right now but I still want to backup my stuff at least for cheap.

I'm also checking AOMEI Backupper, the free version.

Thanks for you inputs/comments.

157
 
 
The original post: /r/datahoarder by /u/qqwertyy on 2025-02-15 22:04:41.

My music library is a bit of a mess; I have a tonne of music on a (512 GB) SD card in my MP3 player. Some is in the cloud (rclone, mountable, so can be explored by file explorer and other tools), some is there with a slightly different naming syntax, some isn't there at all.

Finding dupes is easy. But I'd like to find a fairly straightforward way to locate folders/files that are on my SD card that aren't in the cloud under any name.

Why not just upload everything and windows explorer/ teracopy etc will check if I want to overwrite existing folders? Because on the SD, a file may be be:

'RHCP/2016 - The Getaway/01 The Getaway.mp3',

and on the cloud:

'RHCP - The Getaway/01 - The Getaway.mp3'.

So I won't be prompted , I'll end up wasting bandwidth on uploading duplicate data (my collection is enormous), and then have to clean it all with Dupeguru anyway...

Anyone have a tool for this usage case? Cheers guys n gals

158
 
 
The original post: /r/datahoarder by /u/KJSS3 on 2025-02-15 21:35:26.

https://www.bestbuy.com/site/seagate-expansion-20tb-external-usb-3-0-desktop-hard-drive-with-rescue-data-recovery-services-black/6609643.p?skuId=6609643#tabbed-customerreviews

Only 8.5 hours left. Is that a good deal? Or wait till black Friday or prime days or some other sale?

159
 
 
The original post: /r/datahoarder by /u/jonylentz on 2025-02-15 20:22:44.

So I was making an SSD backup using freefilesync and before I started to sync (copy) to the other drive I noticed that my 500gb SSD was showing in the file list as having 1.15TB of files. This is strange as I do not have compression enabled on this drive... I used another program called TeraCopy to copy files over and this one correctly copied ~500gb of files to the backup folder... to check if all files were copied I used Freefilesync again and clicked compare, strangely this time it is showing as ~830GB of files missing in the backup folder.... (I have double checked the paths and they are correct) What is wrong? It does not make sense that a 500gb drive has 1.15TB of data if this data is not compressed, should I trust that TeraCopy did in fact copied all files? Or should I go with the Freefilesync?

160
 
 
The original post: /r/datahoarder by /u/HalFWit on 2025-02-15 20:17:17.

I have 30+ years of images that I need to organize and store. Date, subject, etc... Is there a tool to simplify this?

161
 
 
The original post: /r/datahoarder by /u/BuyHighValueWomanNow on 2025-02-15 20:07:40.
162
 
 
The original post: /r/datahoarder by /u/TheUnofficialGamer on 2025-02-15 19:36:44.

Too many bad decisions, I know, but any help is appreciated!

163
 
 
The original post: /r/datahoarder by /u/ComeHomeTrueLove on 2025-02-15 19:22:59.

Is there any new fansly Downloader that works?

The last one I know of is one by prof79. Is there a updated one? Or does this still work? I had issues last I tried.

164
 
 
The original post: /r/datahoarder by /u/PratikBhowmik on 2025-02-15 15:44:44.

Is a 5TB WD passport Hard Disk a good deal for 70 Bucks ? Its completely new and sealed. Seller is trusted.

165
 
 
The original post: /r/datahoarder by /u/singingpraise on 2025-02-15 14:44:00.

Hi all,

Is there a way to download individual files? I'm on Leakedzone. Even if I enter the URL of one particular video, it ends up downloading the whole page

Thanks

166
 
 
The original post: /r/datahoarder by /u/These_Growth9876 on 2025-02-15 14:15:15.

I think the question is clear enough. I have noticed that when on linux I have to mount a drive to use it where as on windows the drives are always available. I want to know when the drives are unmount in linux are they in switched off state, and does it help improve the life of the drives, compared to windows?

167
 
 
The original post: /r/datahoarder by /u/FitAd4613 on 2025-02-15 04:20:51.

I had this batch purchased year ago , now it's gonna end soon but haven't watch half the vids. So how can I download the video's.

168
 
 
The original post: /r/datahoarder by /u/NajdorfGrunfeld on 2025-02-15 07:47:27.

I was trying to download all the pdfs from this collection at once: https://archive.org/details/pub_mathematics-magazine?tab=collection

Couldn't find anything useful on the web other than a chrome extension that seems to have expired. I'd appreciate any help.

169
 
 
The original post: /r/datahoarder by /u/comradesean on 2025-02-15 17:11:20.

I'm working on restoring an item unlocker for a video game that relied on an API from 2018, which is no longer active. This process included an HTTP request to a news article that no longer exists and wasn't archived. The good news is that I can take an article from that time period and modify it (which I've already done). However, the JavaScript is broken, and after spending the last week debugging minified and obfuscated JavaScript, I've made no progress.

I'm not familiar enough with HTTrack or other methods for capturing web pages, and no matter what I try, it always seems to break the JavaScript in some way. If anyone has any tips or tricks for capturing a single page with all the necessary scripts intact, your help would be greatly appreciated.

The page I've been trying to use as my base

https://web.archive.org/web/20180414013843fw_/https://blog.twitch.tv/overwatch-league-all-access-pass-on-twitch-8cbf3e23df0a?gi=4debdce8040a

and the httrack that fails me

httrack "https://web.archive.org/web/20180414013843fw_/https://blog.twitch.tv/overwatch-league-all-access-pass-on-twitch-8cbf3e23df0a?gi=4debdce8040a" -O "E:\My Web Sites" "+*.*" "+*.css" "+*.js" "+*.html" "+*.png" "+*.jpg" "+*.gif" "+*.woff" "+*.woff2" "+*.ttf" "+*.svg" --mirror --keep-alive -r99 --max-rate=1000000 --assume "https://web.archive.org/web/20180414013843fw_/https://blog.twitch.tv/overwatch-league-all-access-pass-on-twitch-8cbf3e23df0a?gi=4debdce8040a" --robots=0 --referer "https://web.archive.org/" -%Pt

If anyone has any questions about the whole thing feel free to ask. The rest of the application is essentially done as I've mapped out the memory addresses in the application and recreated an barebones and stubby api just to emulate the process. This is the last piece needed.

170
 
 
The original post: /r/datahoarder by /u/g-e-walker on 2025-02-15 17:00:41.

youtube-dl-react-viewer is a web app for yt-dlp that supports viewing and downloading videos. youtube-dl-react-viewer is 100% free and open-source.

Live Demo | Screenshots | GitHub Repo

Major Changes

  • Added Return YouTube Dislike integration
  • SponsorBlock now uses K Anonymity by default
  • Made improvements to the Dockerfile to allow for easier configuration and future updates
  • Verifying file hashes, importing videos, and repairing videos can now be canceled/stopped early

The full changelog can be found on the releases page

171
 
 
The original post: /r/datahoarder by /u/koiyama on 2025-02-15 14:12:34.

Hi,

I'm currently ripping all my 4K UHDs for easy access through a media player (and maybe an upcoming MadVR HTPC). I currently have two copies of the files:

  • one copy on two HDDs in RAID 1

  • ony copy on a third HDD as cold storage

This means I'm dedicating HDDs in a 3 to 1 ratio.

As I really do not care about uptime, I'm about to split the RAID 1 into two separate single HDDs, making it 3 copies on 3 HDDs.

One copy will be on the HDD plugged into the NAS. This will also be the one that is being accessed.

My question is whether the other two separate single HDDs should both be cold storage or should I have one of them plugged into the NAS basically on "idle".

Best regards

172
 
 
The original post: /r/datahoarder by /u/joker_17SajaD on 2025-02-15 14:00:13.

I saw a script from 3 years ago that did something similar, but it no longer worked. So, I made my own version that downloads audiobook chapters from TokyoBook.

Check it out

If you have any suggestions or improvements, feel free to comment!

173
 
 
The original post: /r/datahoarder by /u/rarelyactivethrow on 2025-02-15 11:18:35.

I've been using a PlusTek large-format scanner and while it's nice to scan 2 pages at a time, I'm at the point where I need to get something I can put a ton of pages in and let it do its thing. I'm trying to keep it under $500. The big snag I hit is that the one that sounded perfect says it only scans documents up to 14" long.

I don't need anything super-duper fancy or fast. Preferably a duplex scanner. Just something reliable that can do a minimum of 300 dpi. Ideally, it would also be capable of making 2 images per "page" to save me from having to manually go in and split every single image.

I saw a couple when I searched the older posts, but the ones that interested me aren't sold new anymore and I have bad luck with used electronics.

174
 
 
The original post: /r/datahoarder by /u/IllPlankton27 on 2025-02-14 18:30:45.

Hi,

1)As part of my backup strategy, I'm using Total Commander to sync my local drive (D:\) with an external drive (E:\Backup\H), where I keep both current and deleted files. I know that Total Commander doesn't verify integrity during copying, and I want to ensure that every file in H has been copied correctly.

Since H contains both my current files and older, deleted ones, I can't just compare it to D:\ using simple checksums. Would this rclone command help?

rclone check "D:\" "E:\Backup\H" --ignore-extra --checksum

  1. My understanding is that this will only verify files that still exist in D:\, but it won’t tell me if older files in H have gone corrupt. Is there a way to verify everything, including old files?

Any advice on a better approach? Thanks!

175
 
 
The original post: /r/datahoarder by /u/KitoLoid on 2025-02-15 10:47:59.

I had Arch Linux installed on my V-Gen Platinum SATA SSD, but I decided to wipe it and use the SSD as a normal storage drive. However, no matter what I do, the partitions keep coming back. Every tool I’ve tried says "success," but when I check again, the partitions are still there. I'm using the SSD as an external boot btw.

My Setup:

  • SSD: V-Gen Platinum (SATA)
  • OS: Dual-booting Windows & Linux Mint
  • Secure Boot: Off
  • BIOS/UEFI Mode: UEFI

What I've Tried So Far (No Success):

Windows Methods:

  • Disk Management – Tried formatting & deleting → No effect
  • Diskpart (Admin CMD)
    • clean / clean all → Permission denied
    • delete partition override → Says success, but still there
    • attributes disk clear readonly → No change

Linux Methods (Arch Live USB):

  • wipefs --all /dev/sdX → Says success, partition still there
  • dd if=/dev/zero of=/dev/sdX bs=1M count=100 → No change
  • blkdiscard /dev/sdX → No effect
  • sgdisk --zap-all /dev/sdX → Says success, but partition is still there

Other Debugging:

  • Checked for GRUB (ls /boot/efi/EFI/) → No GRUB
  • Checked if disk is read-only (lsblk -o name,ro) → Shows R:0 (not read-only)
  • Checked with Windows Disk Management → Partition is fully visible and accessible, but cannot be removed

What Could Be Happening?

  • SSD is locked in firmware write-protection?
  • Corrupt NAND flash making it read-only?
  • Windows keeping some kind of boot protection active?

I haven’t tried using an SSD manufacturer tool yet, but V-Gen doesn’t seem to have one.

What’s Left to Try?

  • Parted Magic Secure Erase?
  • Trying a different OS/Tool?
  • SSD might be dying and needs replacement?

If anyone has ideas or has faced something similar, I’d really appreciate the help.

view more: ‹ prev next ›