X10SDV-6C Build w/ Much Bifurcation/Adapting

Hello, I have been running truneas on a old gaming machine.

I found an X10SDV-6C-TLN4F for 150$ on craigslist and the guy let me test it for aobut 20 minutes and it all checked out…

First, I ordered 4x32GB modules (likely overkill but whatever) of recommended RAM from supermicro website. they are 2933 mhz but my bios shows them as 2133MT/s and I have fiddled with the settings to the point I have to reset cmos… anyone have any info about that?

Next, I have a CoolJag BufB on the way… its the same thing except full copper. I hope that in addition to cleaning and applying PTM7958 (no 7950) and the new full copper heatsink that I get better temps… I also have a ultra low noise fan I got for it. It is a silverstone FTF5010 and it has under 20 dbA.

The chassis I have is 2U PlinkUSA IPC-2026-BK i got for a deal open box on ebay. I thought it was super cool at first but I increasingly dont like it… maybe if I find a solution to where I can mount 3x 3.5" HDD with anti vibration or noise dampening I will be happier… the chassis came with 4 80mm fans that were not so quiet but I installed 4X NF8-ULN and they actually had every so slightly more cfm (20.8) but instead of over 35 dB they are under 15. It is also what I installed in the evercool mount. The noise is a issue due to my living situation, otherwise, I wouldnt care about nosie in a closet. So dont hate on me for the choice but let me know if I need extra cooling anywhere and thank you.

I just wanted to share my current state of affairs because I think it is kind of cool to use a double reverse pcie extension riser, and a x16 to x8x8 adapter so I can use 2 PCIE cards.

So easily enough set the bios to x8 x8 and it works without a hitch. Next I have a M.2 to Pcie 3.0 x 4 adapter/riser/extension whatever… and I plan to use a 280gb 905P.

I am still working out what to do about the drive mounts because the stock ones you can see below the motherboard is one 3x 3.5" stack where the drives basically touch each other… and then in between my psu and the 2x5.25" bay which I got the evercool adapter for. I figure I put 3x HDD in front of that fan and then the other 3x HDD in front pof the fan south of the motherboard and then I can put 6x SATA SSD in between the psu and the evercool adapter.

I was going to grab 6x 1TB Samsung 870 Evo on ebay listed as new for like70 dollars… from china… .is this legit?
I want DRAM sata ssd… there doesnt seem to be a big selection… and i read enough about errors with the MX500 that I would rather avoid…

Oh, and in regards to the 9400-16i… i just happened to have it. so yeah, no hate please thanks and if anything Ill instead a u.2 or nvme in between my pcie cards so there.

comments suggestions feedback etc…

1 Like

I think that board only supports 2133 and 64GB UDIMM or 128 RDIMM. Looks like a Rev 1 board.

Man how was I supposed to know that I looked up the board I got very familiar with manual and I was picking ram by copy and pasting the exact model of recommended ones from supermicro website for the board. Lame as hell

The specs are pretty clear. 2133/1866/1600MHz ECC DDR4 ECC RDIMM and ECC/Non-ECC UDIMM. My X10-SDV is using 64gb of 2133 ECC UDIMM. The tested memory list isn’t a recommended list, it just says it works. Higher speed supported memory will usually work at a lower supported speed.

I assume we are talking about speed–you have RDIMM at 128?

We HATE you. Period. :wink:

D-1500 run at 2400 MHz (1541) or 2133 (most others). Don’t even think about any form of overclocking on a server.

If you want to play… but a 40 mm fan on the heatsink is all that’s necessary to keep a D-1500 in check.

Not sure what you mean here, and what you want to achieve.
How many drives? 3.5" or 2.5"? Why an Optane 900p? Why a HBA?

Fractal Design Node 304 is a proven solution to hold and cool 6 HDDs in a big and reasonably quiet shoebox. Perfect fit with a mini-ITX X10SDV.

3 Likes

Idk I was a bit high and had incredible day in regards to both a stock investment killing it and a raise at work so I thought ima go for it.

There is support ram listed at 2993mhz…rdimm for 128gb… so I am to understand this is for a newer revision of the board than what I have?? where can i find info on the difference betwen revisions and where does it say what my revision is?

Next, your right why an HBA?

I have decided I jumped the gun on multiple things - first of all . I could run 6x 14TB HDD on the sata and return the satadoms that ar expensive…this leaves a m.2 or pcie 3.0x4 and 3.0 x 16…
One Raidz2 would be sufficient for me for quite a while. I guess the HBA would be for having a SSD pool for apps on tnscale and also possibly a slog or maybe a l2arc (likely not) but a slog might be nice… especially with raidz2… and I only have 6 sata ports.

so if I use the nvme for a boot drive… then use the 6 sata ports for the 6x 14TB 3.5" hdd… a HBA, 9400-16i would allow for maybe an additional mirrored HDD pool, a sata ssd pool, and possibly a slog… as well as option to use a u2 or nvme…

thoughts considering this revelation?

Not, it has been tested… to run at 2133 MHz in this board. There’s no nower revision of Xeon D-1500 which would run RAM at 2666 MHz or higher.

SLOG is only for sync writes. While a Data Centre SATA SSD with PLP is an option, NVMe are even better. The x16 PCIe slot can bifurcate x4x4x4x4: With a simple passive riser, this is your way to add an app pool, SLOG and/or L2ARC to an X10SDV.

should most people be disabling sync and not worrying about it?
is a slog not a way of having faster write ? why not enable it and have a slog

ah so just a 3.0 x 16 to 4x nvme non switch chip, which will have full support without being janky like the plx chip ones on ali express…
i like this logic thank you-- also with this said can i get suggestions for nvmes high end to budget baseline acceptible range

i think my plan is to also get a lenovo tiny with a pcie card and have 10gb card in there, direct conenct to truenas and stetup isci storage and

probably use alot less power with only 6 hdd on sata , nvme boot , 3.0 x 16 to 4x nvme card and nothing else…

&& thank you everyone whoe has chimed in

To repeat and clarify:

  • SLOG is only ever used for sync writes
  • write speeds: async >> sync with SLOG >> sync without SLOG

Sync writes are only used in very specific applications like block storage for VMs or databases. Specifically SMB sharing does not use sync writes.

So you probably will not profit from an SLOG device. It simply won’t be used at all, whether you add it to the pool or not. The write cache of ZFS is in memory.

HTH
Patrick

3 Likes