BLUF: For now, I’m keeping both the mirrored NVMe SLOG and the L2ARC because they align with my Time Machine sync write needs and frequently-read large file access. I’ll monitor real-world results, and if the data shows they’re not helping, I’m open to removing them.
Thanks for all the follow-ups — appreciate the mix of perspectives here. I’ll try to address a few of the points raised.
On SLOG and async vs. sync writes (@Sara)
I agree: async writes never touch the SLOG. My reason for setting sync=always on the Time Machine dataset is that macOS uses sync writes over SMB for many operations, and for backups, I prefer the durability guarantee even at the cost of some throughput. Without a SLOG, sync writes go to spinning rust in the RAIDZ2 vdev, the slow path. With a fast mirrored NVMe SLOG, I’m aiming to shift that bottleneck closer to the network link speed. I’m not expecting async performance here, just better sync performance where it matters.
On SLOG mirroring (@Sara)
I understand the argument for a single SLOG device, but for me the mirrored SLOG is about peace of mind. If the pool crashes and the single SLOG also fails in that window, I’d rather not lose in-flight sync writes. It’s low probability, but my NVMe slots are already populated, so I’m fine paying the small “opportunity cost” in capacity.
On L2ARC with a special vdev (@Sara & @Constantin)
Point taken that with a fast mirrored special vdev holding metadata and small files, L2ARC may offer minimal benefit for metadata access. However, in my own usage I’ve seen the L2ARC hold ~700 GB even with the special vdev active, which tells me it’s caching more than just metadata — likely frequently accessed larger files that don’t live on the special vdev. I plan to keep it for now, monitor ARC/L2ARC hit rates, and remove it if the benefit turns out to be negligible.
On special vdev redundancy (@Constantin)
I did look at my current metadata usage, and the 2-way mirror gives me headroom for growth. I appreciate the 3-way mirror suggestion; if I see sustained growth or increased risk tolerance needs, I’ll revisit.
On SLOG impact for Time Machine (@Stux & @swc-phil)
@Stux — thanks for confirming what I’ve read and seen elsewhere: in the right conditions, the jump from HDD-bound sync writes to NVMe-accelerated sync writes can be dramatic.
@swc-phil — I think the big difference in results comes down to hardware and layout. If the base pool can already do decent sync speeds, the uplift will be smaller. In my case, the pool is 10 × 18 TB HDDs in RAIDZ2, so raw sync write speed without a SLOG is in the ~10 MB/s range. I’m expecting improvement into the hundreds of MB/s with the SLOG, but I’ll measure and report back so it’s data, not just theory.
Thanks again for the constructive debate — even where we disagree, it’s helped me sanity-check the design.