Moey posted:How would send/receive help me in this current situation though? Without bringing in some temporary storage as a dumping ground for a two-transfer migration. If you’re smart about it, you enable zpool checkpoint until you’re satisfied that everything made it over - that way, you can revert administrative changes like dataset removal. Just don’t forget to turn it off again. BlankSystemDaemon fucked around with this message at 11:37 on Mar 25, 2024 |
|
# ¿ Mar 25, 2024 11:30 |
|
|
# ¿ May 16, 2024 23:03 |
Moey posted:Neato. Gracias. I still periodically do it if it's been a while since I've done some administrative task and want to make sure I'm doing it right. This, of course, goes hand in hand with using the -n flag at least once before running it without, on any administrative command.
|
|
# ¿ Mar 25, 2024 15:11 |
Computer viking posted:Something about playing with a large stack of real disks (before putting them to their final use) feels good, though. I can't explain why.
|
|
# ¿ Mar 26, 2024 01:40 |
Computer viking posted:Sure, but they don't make fun disk access noises. BlankSystemDaemon fucked around with this message at 15:20 on Mar 26, 2024 |
|
# ¿ Mar 26, 2024 15:18 |
Wibla posted:I thought that was BSD's job Talorat posted:Tell me more about that second option. What’s a vdev? Will this allow the single pool to have a single mount point? If you add a vdev to an existing pool, you expand the pool, and data will be distributed across then span such that they ahould end up being approximately equally full. See zfsconcepts(7). EDIT: Looking at it, I think this article from Klara explains it best. BlankSystemDaemon fucked around with this message at 08:25 on Mar 28, 2024 |
|
# ¿ Mar 28, 2024 08:17 |
Just an observation, but if you aren't willing to switch to ZFS, it probably means you either don't have backups, or don't trust them - so you might also wanna address that. If you do trust your backups, you can remove one of the disks from the existing mirror, put a ZFS pool on it, move data over to it, and then the second disk and use zpool-attach(8) to add the second disk to turn the vdev into a mirror. If you're using something based on FreeBSD, you can also use gnop(8) and zpool-replace(8).
|
|
# ¿ Apr 13, 2024 09:53 |
Moey posted:Or they are using a different mature solution. Not everyone is in the same ZFS cult as you. What I should've written is "if you aren't able to switch to ZFS by degrading your existing mirror" - because that's a fairly trivial operation if you have backups.
|
|
# ¿ Apr 14, 2024 11:57 |
ZFS works best with whole disks without any partitioning, if you don’t need to plug the disks into a system likely to try and “initialize” a disk that appears empty, or don’t need things like boot records or swap partitions.
|
|
# ¿ Apr 20, 2024 01:52 |
mekyabetsu posted:Yup, this is what I did when I created the pool. I ran “zpool create” with 2 drives that were unpartitioned, and that was the result. If that works for ZFS, it’s fine with me. I just wasn’t sure why it chose those particular partition types. I know ZFS was originally a Sun Solaris thing, so it’s probably related to that. You need to do this, because Linux is the one Unix-like that doesn't understand that it shouldn't reassign drives between reboots (the reasons why it does this has to do with its floppy disk handling) - so there's a small risk that you'll trigger a resilver; typically this isn't a problem, but does degrade the array, meaning that a URE could cause dataloss. On my fileserver, the 24/7 online pool is is a raidz2 of 3x6TB+1x8TB internal disks in raidz2 totalling ~20TB, and the offline onsite backup pool is just shy of 200TB in total made up of 15x2TB raidz3 vdevs each in their own SAS2 enclosure. The internal drives are what the system boots to and they're where all the 24/7 storage lives, so they have partitioning both for the EFI System Partition, a swap partition on the 8TB, and the rest is used for root-on-ZFS The external drives are all completely unpartitioned, because this lets me simply run sesutil locate to turn on a LED to make it easy to identify the disk that needs replacing, and then I just go pull the disk and insert a new one - this is the advantage of unpartitioned disks, because zfs automatically starts replacing the disk on its own, and if all devices in a vdev have been replaced with something bigger, the vdev grows automatically too (this is accomplished using the autoreplace and autoexpand properties documented in zpoolprops(7)). EDIT: I still need to figure out if it's possible to automatically turn on the fault LED in FreeBSD. Trouble is, every failure of spinning rust I've had has been the kind of error that's hard to know about without ZFS (and about half have been impossible to figure out by using S.M.A.R.T alone), so I'm not sure I'd even have benefited. BlankSystemDaemon fucked around with this message at 13:20 on Apr 20, 2024 |
|
# ¿ Apr 20, 2024 13:10 |
Computer viking posted:Depends on if you use it as your everything-server, I guess - with enough VMs or containers it could make sense. A FreeBSD machine can sit with a load average of 2x the number of cores it has, and still manage just fine so long as the priorities of synchronous, interactive and asynchronous tasks aren’t hosed up. BlankSystemDaemon fucked around with this message at 13:46 on Apr 22, 2024 |
|
# ¿ Apr 22, 2024 13:41 |
movax posted:But, question then -- for my metadata / special vdev, currently I have 2x 7.68TB SATA SSDs in a mirror. If I have 4 ports available, should I do a RAID-Z1 on that? Or, if I want to maintain the double-drive failure tolerance (less worried about this on a SSD to be fair), is my only choice a 4-drive RAID-Z2? Striped mirrors doesn't make sense because the wrong 2 drives there nuke the vdev whereas RAID-Z2 is any two. It's not NVMe but it's 6Gbps SATA on the PCH controller for a server mostly doing Linux ISOs, so I think it's fine. You can N-way mirroring, though - so for example, instead of mirroring across 2 disks, you can mirror across more than two; that way, you can lose two disks and still not lose the pool. Also, it's worth mentioning that SAS can be daisy-chained up to 5 SAS enclosures per external SAS port - but unfortunately this is only available for real SAS enclosures. For DIY, the best you can do is find a way to supply power to a SAS expander in the UNAS case, then buy SAS expanders to reduce the number of cables going from/to each device. Minor nit: The ZFS Intent Log is part of the on-disk specification - when you add a slog device, you're really adding a separate log device to be used instead of the ZIL. BlankSystemDaemon fucked around with this message at 09:18 on Apr 25, 2024 |
|
# ¿ Apr 25, 2024 09:11 |
I just use plain FreeBSD with zfs, bhyve, and jails - it works great, because it isn't limited to what the appliance makers has decided the appliance should be capable of, and equally importantly what they've decided they won't support.
|
|
# ¿ Apr 26, 2024 15:00 |
movax posted:Thanks -- I may look at adding just one drive then so the redundancy is matched, cheaper that way as well vs buying two more SATA SSDs. So it really shouldn't be an issue, so long as you avoid some of the EMC SANs of the past, where they'd actually lock out those features until the customer paid for them. EDIT: The biggest real difference with SAS is that you get access to the SCSI READ DEFECT DATA command, which is what we all wish S.M.A.R.T was. BlankSystemDaemon fucked around with this message at 23:19 on Apr 26, 2024 |
|
# ¿ Apr 26, 2024 23:13 |
Hardware for a NAS is easy to get right. Buy a low-power Intel i3 or AMD APU with proper ECC support, throw in some memory, and make sure you have enough SATA and PCIe ports/slots to fill the bays. Software-wise, there's an unenumerable amount of ways to gently caress up.
|
|
# ¿ Apr 29, 2024 19:12 |
Korean Boomhauer posted:Dumb question but is there anything that truenas scale does zfs-wise that i couldn't replicate with a cronjob if i were to just use zfs on proxmox and have a nfs/smb container somewhere? Like what all would I have to setup to make sure my data doesn't get exploded aside from a weekly scrub? I'm pretty sure I can view drive health right in proxmox as well. Gonna Send It posted:I like Sanoid/Syncoid for my simple home ZFS backup uses. I have it setup to pull from the primary server to the backup server via ssh juuuuust in case the primary is somehow compromised.
|
|
# ¿ May 3, 2024 18:55 |
IOwnCalculus posted:These are two very different failure modes. Total and unrecoverable loss of three drives in the original vdev or four in the new vdev will cause you to lose the whole array, yes. It should, of course, never be run without using the -n flag first, because it's one of those destructive administrative tasks that not even zpool checkpoint can save against (despite being made to guard against destructive administrative tasks), and may also require the use of the -X flag. Basically, it's the thing you try right before you start restoring from the backups that you've hopefully automated.
|
|
# ¿ May 4, 2024 22:14 |
Pablo Bluth posted:Best case for 1Gbe is 125MB/s, which would be close to 12 hours (2.5GbE would get that down to under 5). Use iperf3 to get a measure of just network transfer speed without potential disk limits at either end. Jumboframes can help a bit, but needs some amount of hardware support. EDIT: This also assumes a high goodput ratio. BlankSystemDaemon fucked around with this message at 00:18 on May 5, 2024 |
|
# ¿ May 5, 2024 00:15 |
PitViper posted:After dealing with a clusterfuck of a drive swap (one disk failing smart and doing a replacement, then two more disks in the second z1 pool giving read errors and being faulted during the resilver) i'm left with a list of about 1400-1500 file errors from status -v. Similarly, textfiles aren't really susceptible to a bitflip, because you can open the file, and probably figure out which value was switched (especially if it's a config file that has a linter/checker). The worst files to have bitflips happen to is anything that's in a binary format, without any kind of built-in error checking - things like the Windows Registry, logs from systemd, save files for most video games, programs and libraries, and stuff like that. OpenZFS 2.2 has also implemented corrective receive whereby, if you have a snapshot lying about that contains a good copy of the data that's been flipped, you can use zfs recv to fix it. There's also an issue open to try and fix if you've only got very very big snapshots, but there's no implementation for that yet. IOwnCalculus posted:Right? A regular RAID would've gone completely unrecoverable very early on in the process, here you're dealing with bitrot that's probably nigh-undetectable because it's such a small amount of data in a video file. It's still not backup, though. BlankSystemDaemon fucked around with this message at 00:04 on May 8, 2024 |
|
# ¿ May 8, 2024 00:01 |
I can't remember if I've posted this before, but I'm just gonna do it again if so: The READ column identifies number of failed READ commands issued by ZFS to the kernels disk driver subsystem. The WRITE column is the same but for WRITE commands. The CKSUM column identifies number of times that the drive returned-from-READ-command something other than what ZFS knows should be there, based on the checksum. BlankSystemDaemon fucked around with this message at 00:42 on May 8, 2024 |
|
# ¿ May 8, 2024 00:40 |
Harik posted:took nearly a year because my dog got sick and wiped out my toy fund (he's fine now, good pupper) Also, cute pupper. Please pet him from me Harik posted:I wasn't even aware there was u.3 and lol it exists because the sas lines weren't shared with pcie on u.2. why do we have to keep dragging sata/sas into every interface? So of course now's the time for the hyperscalers to move to E1.L or E3.S, or even using NVMe-over-PCIe for spinning rust, because it simplifies the design of the rack servers. For ARC, the only thing that really matters is the hit/miss ratio. If you're below 90% and haven't maxed your memory, download more RAM. If you've maxed your memory, you can look into L2ARC. Just remember that L2ARC isn't MFU+MRU like ARC (it's a simple LRU-evict cache), and that every LBA on your L2ARC device will take up 70 bytes of memory that could otherwise be used by the ARC (meaning you can OOM your system if you add one that's too big). Combat Pretzel posted:ARC total accesses? I presume so.
|
|
# ¿ May 10, 2024 07:42 |
Combat Pretzel posted:FFS, you keep claiming this. It's 70 bytes per ZFS data block. The problem is, records in ZFS are variable size - and there's no real way to get the distribution across an entire pool. You should still max out your memory before using L2ARC, though.
|
|
# ¿ May 10, 2024 21:47 |
Combat Pretzel posted:A PR was just opened on the OpenZFS Github that'll enable a failsafe mode on metadata special devices, meaning all writes get passed through to the pool, too, so that in case that your special vdev craps out (for reasons, or simply lacking redundancy), your pool will be safe anyway. Guess I'll be using a spare NVMe slot for metadata, as soon this is deemed stable.
|
|
# ¿ May 13, 2024 19:12 |
Generic Monk posted:Lmao I added a SATA ssd to my pool for this only to realise i hadn’t read the small print that you can’t remove it and it’s now a single point of failure for the whole pool. Who knows this might hit truenas before either the drive dies or I nuke and rebuild the pool ZFS doesn't give a gently caress about what driver you're using, nor what the disk is. Kibner posted:It was tunable but could result in system stability issues, iirc. I found this after a brief search: https://github.com/openzfs/zfs/commit/518b4876022eee58b14903da09b99c01b8caa754 That'll be a pretty big advantage, if it turns out to be possible - something that I'm not quite sure of. I think it's possible on FreeBSD and Illumos, but it doesn't seem likely to be possible on Linux. Harik posted:we were talking about this: I'm interested to learn how Generic Monk managed to learn about the 'special' vdev, without learning that it should always have its own redundancy via mirroring (or even n-way mirroring), though. All the documentation I know of makes a big deal out of making it very explicit. BlankSystemDaemon fucked around with this message at 12:16 on May 15, 2024 |
|
# ¿ May 15, 2024 12:03 |
Whoever made that video needs to delete it, because it's almost deliberate misinformation to claim it's a "cache" and that you can use a single disk.
|
|
# ¿ May 16, 2024 11:35 |
|
|
# ¿ May 16, 2024 23:03 |
Combat Pretzel posted:4:18 "it should have some level of redundancy". Wendel is a good enough educator that he should've known better.
|
|
# ¿ May 16, 2024 22:14 |