I have a 2 bay NAS, and I was planning on using 2x 18tb HDDs in raid 1. I was planning on purchasing 3 of these drives so when one fails I have the replacement. (I am aware that you should purchase at different times to reduce risk of them all failing at the same time)

Then I setup restic.

It makes backups so easy that I am wondering if I should even bother with raid.

Currently I have ~1TB of backups, and with restics snapshots, it won’t grow to be that big anyways.

Either way, I will be storing the backups in aws S3. So is it still worth it to use raid? (I also will be storing backups at my parents)

  • Admiral Patrick@dubvee.org
    link
    fedilink
    English
    arrow-up
    31
    ·
    edit-2
    25 days ago

    I always do some level of RAID. If for no other reason, I’m not out of commission if a disk fails. When you’re working with multi TB, restoring from a backup can take a while. If rapid recovery from a disk failure is not a high priority for you, then you could probably do without RAID.

    Either way, make sure you test your backups occasionally.

    Another way to put it: With RAID, a disk failure is like your Check Engine light coming on. You can still drive, but you should address the problem as soon as you can. Without RAID, it’s like your engine has seized up and you have to tow it for repair and are without your car until it’s fixed.

    • Dust0741@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      7
      ·
      25 days ago

      Hmm that’s a good point.

      Aws also can cost a good chunk if you restore un-optimally

  • BakedCatboy@lemmy.ml
    link
    fedilink
    English
    arrow-up
    11
    ·
    25 days ago

    Keep in mind that if you set up raid using zfs or btrfs (idk how it works with other systems but that’s what I’ve used) then you also get scrubs which detect and fix bit rot and unrecoverable read errors. Without that or a similar system, those errors will go undetected and your backup system will backup those corrupted files as well.

    Personally one of the main reasons I used zfs and now btrfs with redundancy is to protect irreplaceable files (family memories and stuff) from those kinds of errors, as I used to just keep stuff on a hard drive until I discovered loads of my irreplaceable vacation photos to be corrupted, including the backups which backed up the corruption.

    If your files can be reacquired, then I don’t think it’s a big deal. But if they aren’t, then I think having scrubs or integrity checks with redundancy so that issues can be repaired, as well as backups with snapshots to prevent errors or mistakes from messing up your backups, is a necessity. But it just depends on how much you value your files.

  • RAID 1 is mirroring. If you accidentally delete a file, or it becomes corrupt (for reasons other than drive failure), RAID 1 will faithfully replicate that delete/corruption to both drives. RAID 1 only protects you from drive failure.

    Implement backups before RAID. If you have an extra drive, use it for backups first.

    There is only one case when it’s smart to use RAID on a machine with no backups, and that’s RAID 0 on a read-only server where the data is being replicated in from somewhere else. All other RAID levels only protect against drive failure, and not against the far more common causes of data loss: user- or application-caused data corruption.

    • whodatdair@lemmy.blahaj.zone
      link
      fedilink
      English
      arrow-up
      3
      ·
      edit-2
      24 days ago

      I know it’s not totally relevant but I once convinced a company to run their log aggregators with 75 servers and 15 disks in raid0 each.

      We relied on the app layer to make sure there was at least 3 copies of the data and if a node’s array shat the bed the rest of the cluster would heal and replicate what was lost. Once the DC people swapped the disk we had automation to rebuild the disks and add the host back into the cluster.

      It was glorious - 75 servers each splitting the read/write operations 1/75th and then each server splitting that further between 15 disks. Each query had the potential to have ~1100 disks respond in concert, each with a tiny slice of the data you asked for. It was SO fast.

  • Mikelius@lemmy.ml
    link
    fedilink
    English
    arrow-up
    5
    ·
    24 days ago

    Raid 1 has saved my server a couple of times over from disaster. I make weekly cold backups, but I didn’t have to worry about it when my alert came in notifying me which drive went dead - just swap, rebuild, move along. So yeah I’d say it’s definitely worth it. Just don’t treat raid as a backup solution - and yes, continue to use an external cold storage backup solution as you mentioned. Fires, exploding power supplies, ransomware, etc don’t care if you’re using raid or not.

  • MangoPenguin@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    4
    ·
    24 days ago

    RAID means that if a drive fails you don’t have some downtime while your backups restore. It depends on how you feel about waiting for that.

  • RedEye FlightControl@lemmy.world
    link
    fedilink
    English
    arrow-up
    4
    ·
    25 days ago

    Yes yes yes yes yes

    Raid1 that thing and sleep easier. Good on you for having a cold spare, and knowing to buy your drives at different locations/times to get different batches. Your head is in the right place! No reason to leave that data unprotected if you have the underlying tech and hardware.

  • sugar_in_your_tea@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    2
    ·
    24 days ago

    I absolutely would, for a few reasons:

    • restoring from backup is a last resort and involves downtime; swapping a disk is comparatively easier and less disruptive
    • it’s possible your backup solution fails, so having some redundancy is always good
    • read performance - not a major factor, but saturating a gigabit link is always nice
  • kn33@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    ·
    edit-2
    25 days ago

    It’s up to you. Things to consider:

    • Size of data
    • Recovery speed (Internet speed)
    • Recovery time objective
    • Recovery point objective (If you’re backing up once per day, is it okay to lose 23 hours of data when a disk fails?)

    If your recovery objectives can be met with the anticipated data size and recovery speed, then you could do RAID 0 instead of RAID 1 to get higher speeds and capacity. Just know that if you do that, you better be on top of your backups because they will be needed eventually.

  • Moonrise2473@feddit.it
    link
    fedilink
    English
    arrow-up
    1
    ·
    24 days ago

    i was also thinking like this, then i had to restore everything from a backup when the ssd suddenly died. I wasted so much time setting everything back as before

  • Shimitar@feddit.it
    link
    fedilink
    English
    arrow-up
    1
    ·
    20 days ago

    Yes, I would still do raid. Because a disk fail will not cause a blackout. Much better than have your server offline waiting to replace disk and restore backup.

    And no way you can backup 18tb in 1tb, restic or no restic.