0

OpenZFS Native Encryption Use Has New(ish) Data Corruption Bug - Slashdot

 7 months ago
source link: https://hardware.slashdot.org/story/24/02/17/0613215/openzfs-native-encryption-use-has-newish-data-corruption-bug
Go to the source link to view the article. You can view the picture content, updated content and better typesetting reading experience. If the link is broken, please click the button below to view the snapshot at that time.
neoserver,ios ssh client

OpenZFS Native Encryption Use Has New(ish) Data Corruption Bug

Sign up for the Slashdot newsletter! OR check out the new Slashdot job board to browse remote jobs or jobs in your areaDo you develop on GitHub? You can keep using GitHub but automatically sync your GitHub releases to SourceForge quickly and easily with this tool so your projects have a backup location, and get your project in front of SourceForge's nearly 20 million monthly users. It takes less than a minute. Get new users downloading your project releases today!
×

OpenZFS Native Encryption Use Has New(ish) Data Corruption Bug (phoronix.com) 13

Posted by EditorDavid

on Saturday February 17, 2024 @03:34PM from the Zettabyte-File-System dept.

Some ZFS news from Phoronix this week. "At the end of last year OpenZFS 2.2.2 was released to fix a rare but nasty data corruption issue, but it turns out there are other data corruption bug(s) still lurking in the OpenZFS file-system codebase."

A Phoronix reader wrote in today about an OpenZFS data corruption bug when employing native encryption and making use of send/recv support. Making use of zfs send on an encrypted dataset can cause one or more snapshots to report errors. OpenZFS data corruption issues in this area have apparently been known for years. Since May 2021 there's been this open issue around ZFS corruption related to snapshots on post-2.0 OpenZFS. That issue remains open. A new ticket has been opened for OpenZFS as well in proposing to add warnings against using ZFS native encryption and the send/receive support in production environments. jd (Slashdot reader #1,658) spotted the news — and adds a positive note. "Bugs, old and new, are being catalogued and addressed much more quickly now that core development is done under Linux, even though it is not mainstreamed in the kernel."

I read through the entire thread over there and all people keep talking about is filesystems that implement redundancy. It's like md doesn't exist. And md is tasty and delicious. I really do not get it.

  • Re:

    For much of my life, I've avoided md because I didn't know how to use it and I was afraid that if something went wrong while removing a failed drive or rebuilding the new drive, I could create a mess that results in losing data on all of my drives rather than just the failed drive. After learning how simple md is in Linux, I was just about ready to start using it. However at the same time I've been looking into Proxmox with Ceph and I don't know if it makes sense to use Ceph and md together - it seems lik
    • Re:

      Hah, I saw what you did there.

      My current media server volume was created in early 2013 and has migrated between Gentoo and Ubuntu, and RAID5 and RAID6 with no issues. Did the one at a time increase in drive sizes twice so far. Reinstall Linux, mount md device, really that simple. I recommend it heartily. I use it elsewhere but that's my crown jewel. I'd used hardware RAID controllers before that for close on to 20 years and it's really just so much better in Linux.

  • Re:

    md is great for sure, but requires the drive firmware to accurately report errors. ZFS can do parity checking on reads and writes allowing you to know if your data is stored and read correctly and recover in the right configuration.

    • Re:

      I suppose the ability to overcome bad hardware is pretty impressive. Of course trusting data to hardware that doesn't report errors correctly - that's another matter.

      • You do this all the time.

        ZFS was designed by engineers that knew _exactly_ what assurances the RAS features of Sun hardware provided, on every cache, on every bus, every chip, and that was quite a lot more than what Intel servers had, or in a lot of ways have today probably.

        You use ECC and RAID because of hardware that doesn't report errors correctly, and those don't cover all the gaps.

  • Re:

    When you break down very complex problems of data reliability into a single word "redundancy" I'm sure you would be confused as to why people don't just use MD. It's like saying solving the wave equation is the same as adding 2+2 together because both are just "math" and why do people even bother learning math beyond grade 2.

    ZFS has an insanely rich feature set, "redundancy" is just one tiny part of it, and if that were all people cared about, they would just use md.

  • Re:

    I used to use md for setting up software RAID 5. Decided to try OpenZFS with my new NAS. I definitely regret it... it's one of those cases of people reinventing the wheel to replace it with something far more complicated with features nobody really needs. It's working, but I'm really not confident in it. I should really reimage it before it gets too full...

  • Re:

    I don't use ZFS for redundancy, although it does have that. I use it for its subvolume and snapshot capabilities, and also its resistance to bitrot and silent file corruption. I have used BtrFS in the past, which also has those same capabilities. I never could get BtrFS tuned very well, and no one could tell me specifically what to look at. Performance was totally abysmal on spinning rust after a while. Very high CPU load numbers from waiting on the disk all the time. I've never had that issue with ZFS o


About Joyk


Aggregate valuable and interesting links.
Joyk means Joy of geeK