Replies: 4 comments 16 replies
-
At this time ZFS still has no NUMA awareness outside of what underlying OS can do. So unless you are going to manually manage some CPU affinity for all relevant individual threads to one domain, I would make topology symmetrical. But then your application or networking may not get balanced, etc. so it all may end up out of reasonable control. |
Beta Was this translation helpful? Give feedback.
-
To begin with, for 24x Nvme (u.2/u.3/e1) SSDs, no one that wants performance would even thinking of using ZFS. EDIT: Just to avoid further stupid discussions: Theory how it could work: But thats only memory, there are so many other factors, even checksumming will bottleneck. I have 2x such Genoa Systems, with maxed out memory channels and tryed ZFS with 8x Micron 7450 Max in Raid10 on each server. Lets hope things changes, because if not ZFS will get only a filesystem for HDD's and Homelabbers, maybe Bcachefs might be a solution, but first benchmarks shows that its even slower as zfs and there seems to be still bugs. Don't want to attack anyone who loves zfs, i love it myself, its just the reality because we move more and more to faster/more affordable nvme's. (Data center U.3 ssds aren't that expensive anymore) |
Beta Was this translation helpful? Give feedback.
-
But tbh, if your Workload is such a big static Files, why do you need 24x NVME's? But your workload doesnt need performance at all, wouldn't it be a better solution to save the money and use ZFS with HDD's + a really fast mirror a bigger special vdev (like 3,8TB) in case you need still something that needs a lot of iops or a VM or something? If i would be you, i would buy instead some big HDD's, like 12x22TB use them in Raid10, or a stripe of 2x Raid-Z2 with a mirror of special vdev like i suggested above, just to make sure that all sorts of Workloads are fast. That gives you 130TB of usable Space, which is plenty and you can still do 1,2GB/s write and 2,4GB/s speed with Raid10. Don't missinterprete my writing, but i readed you initial Request as, 24xNVME and ZFS, which leaded me to all this. Cheers |
Beta Was this translation helpful? Give feedback.
-
Apologies for the ignorance, but I thought ZFS allowed expansion of the pool by adding additional vdev's. So if we started with a 6x30TB RAID-Z1, we could (down the road) add another 6x30TB RAID-Z1 to the existing pool without requiring rebuilds or data migrations. |
Beta Was this translation helpful? Give feedback.
-
Let's assume I have the following hardware configuration.
A 2U server with 24x 2.5" U.2 NVMe slots.
Let's assume the PCIe lane configuration is such that the first 12 slots of the server map directly to CPU socket 1, and the remaining 12 slots map directly to CPU socket 2.
https://infohub.delltechnologies.com/en-US/l/nvme-and-i-o-topologies-for-next-generation-dell-poweredge-servers-1/poweredge-r7525-populated-with-24-nvme-drives-4/
Let's assume my objective is to extract the maximum performance from the hardware. Is it better to have all drives <12 mapping to one CPU or the other? Or would there be performance benefits (or other considerations) splitting the drives between different CPU sockets?
We can hopefully ignore any mirror vs RAID-Z discussions because the crux of the question surrounds optimal drive configuration at the hardware layer.
Beta Was this translation helpful? Give feedback.
All reactions