Physical drives missing from lsblk

Today I installed scale on bare metal after proxmox wouldn’t meet my needs regarding the drives.

I have an x21 that has 18 working m2 drives attached, all 18 of them appeared in proxmox, as shown in the image below,
image

I installed scale on the same hardware that proxmox was on however, I am only seeing 15 drives?

The drives are 3.6T, I used lsblk -a, as I understand that if a drive has a partition then it wont appear usable to scale.

I setup a pool using the 15 drives that were presenting as I need to move data from the 7.3t drives.

I am trying to work out why the other drives have disappeared? looking for any tips on how to find them

Your question, while fascinating, is does not give enough data to formulate an answer. Please read Joes Rules (linked below in my signature) to post the data we need. I suspect the hardware is the issue off the top of my head.

When you list the hardware, also list “exactly” how all drives (including the 7.3TB formatted drives) are physically connected to the system. Example: drive sda is connected to SATA port 1, nvm0 (serial number) is connected to motherboard M.2 slot 1. I really do think it is hardware however let’s see what you have going on and see if we can fix it.

That is a lot of NVMe drives, I like it. If you can include a few photos of the system (insides), that would be nice for me to see, however it is not required.

Thanks for the reply


As requested pics of the insides, so the x21 which has the m2 drives is a single pci card, that hosts 21 nvme drives

The x21 is connected to the top PCIe 4.0 x16 slots - all drives simply plug into this card
The SAS2308 is connected to Middle 4.0 x16 slots - sas cables them connect to the 5 x 8tb drives

CPU = Intel(R) Xeon(R) W-1350
RAM = 64 gig
Nextwork = Intel x520 - 10-Gigabit SFI/SFP+
Mobo = Supermicro X12SAE-5
Storage 1 = LSI SAS2308
Storage 2 = Apex storage x21

The

Below is the output of lspci

Now looking at the lspci
image
these appear to be the x21 ports and can see all 21 appearing

These are the 15 drives that are appearing

The two samsung devices are the boot pool


I will remove all the drives and test them off the board to see which drives are not appearing and see if they appear within windows or if they damaged.

Anything else you can suggest?

I would not do that. There is no reason for these drives to fail, especially so many of them.

What I would do is to examine the M.2 card, and record which slot each NVMe is plugged into, by serial number. Why do I ask that? To see if there is some commonality.

Let me digest this data, I have something else in mind.

You have 14 drives on the PCIe card and the 2 boot-pool drives are on the motherboard, correct?

I’m curious if 5 of the drives are on the daughter board.

Check that the secondary card of the Apex X21 has not come loose.

Unrelated to this issue, where would you plug the NIC, and why use HBA rather than SATA ports from the motherboard?

Correct, the two boot drives are nvmes directly attached to the motherboard, the x21 has all 21 slots filled.

The nic is plugged into the very bottom of the motherboard it’s a pci 1x slot, working fine, I don’t get full 10g throughput, the hba is old gear that the x21 is replacing and will be removed once I get the x21 full working along with its attached drives.

Right, I went through each drive matched the S/N and ones there were missing in scale I tested in windows, each one worked fine

top left, bottom left and far right do not appear in scale


top left does not appear

far right does not appear

bottom does not appear

The developer has also confirmed that pcm is the x21
image
meaning scale is seeing all the ports, but not the drives attached

Fun fact

lspci is seeing 19 drives now, I am not seeing them appear in scale, so where are those other 4 drives?

Those drives appeared after I noticed I had the x21 in the second pci slot not the first.

Reading the motherboard manual, as sometime there can be issue with m2 slots on the motherboard and PCIe lanes, nothing is mentioned of shared lanes

reading the system block diagram this also appears to be true as well


if I understand it correctly the two x16 slots in the top left don’t share with the m2 slots

so getting back to this, I reinstalled scale this morning, after removing the onbaord nvme drives and the old lsi card plus the old hdds as well.

During the install of scale, I could see a further four nvmes I could select to install scale on, I installed scale on two sata ssds I had installed, however once in scale, no new drives were present.

Looking at dmesg, I can see that four drives simply just don’t bother appearing.

root@truenas[~]# dmesg | grep nvme
[    1.874809] nvme nvme0: pci function 0000:03:00.0
[    1.874812] nvme nvme1: pci function 0000:04:00.0
[    1.874890] nvme nvme3: pci function 0000:08:00.0
[    1.875114] nvme nvme4: pci function 0000:09:00.0
[    1.875116] nvme nvme12: pci function 0000:10:00.0
[    1.875139] nvme nvme5: pci function 0000:07:00.0
[    1.875140] nvme nvme2: pci function 0000:05:00.0
[    1.875146] nvme nvme10: pci function 0000:0e:00.0
[    1.875259] nvme nvme9: pci function 0000:0f:00.0
[    1.875262] nvme nvme11: pci function 0000:11:00.0
[    1.875958] nvme nvme7: pci function 0000:06:00.0
[    1.876466] nvme nvme17: pci function 0000:16:00.0
[    1.876477] nvme nvme8: pci function 0000:0d:00.0
[    1.876496] nvme nvme16: pci function 0000:13:00.0
[    1.876498] nvme nvme6: pci function 0000:0a:00.0
[    1.876499] nvme nvme14: pci function 0000:12:00.0
[    1.876499] nvme nvme15: pci function 0000:15:00.0
[    1.877241] nvme nvme13: pci function 0000:14:00.0
[    1.877360] nvme nvme18: pci function 0000:17:00.0
[    1.912123] nvme nvme1: missing or invalid SUBNQN field.
[    1.943460] nvme nvme7: missing or invalid SUBNQN field.
[    1.943730] nvme nvme17: missing or invalid SUBNQN field.
[    1.946828] nvme nvme9: allocated 64 MiB host memory buffer.
[    1.953676] nvme nvme2: allocated 64 MiB host memory buffer.
[    1.955912] nvme nvme6: missing or invalid SUBNQN field.
[    1.957254] nvme nvme3: missing or invalid SUBNQN field.
[    1.963520] nvme nvme12: allocated 64 MiB host memory buffer.
[    1.963721] nvme nvme5: allocated 64 MiB host memory buffer.
[    1.964479] nvme nvme10: allocated 64 MiB host memory buffer.
[    1.965117] nvme nvme4: allocated 64 MiB host memory buffer.
[    1.965245] nvme nvme1: allocated 32 MiB host memory buffer.
[    1.969048] nvme nvme7: allocated 32 MiB host memory buffer.
[    1.969839] nvme nvme17: allocated 32 MiB host memory buffer.
[    1.973634] nvme nvme8: allocated 64 MiB host memory buffer.
[    1.974242] nvme nvme16: allocated 64 MiB host memory buffer.
[    1.981490] nvme nvme6: allocated 32 MiB host memory buffer.
[    1.982362] nvme nvme3: allocated 32 MiB host memory buffer.
[    1.982977] nvme nvme14: allocated 64 MiB host memory buffer.
[    1.985179] nvme nvme15: allocated 64 MiB host memory buffer.
[    2.022729] nvme nvme6: 8/0/0 default/read/poll queues
[    2.024516] nvme nvme17: 8/0/0 default/read/poll queues
[    2.025743] nvme nvme7: 8/0/0 default/read/poll queues
[    2.026269]  nvme6n1: p1
[    2.028039]  nvme17n1: p1
[    2.029278]  nvme7n1: p1
[    2.031156] nvme nvme5: 12/0/0 default/read/poll queues
[    2.033037] nvme nvme3: 8/0/0 default/read/poll queues
[    2.034415] nvme nvme10: 12/0/0 default/read/poll queues
[    2.035168] nvme nvme3: Ignoring bogus Namespace Identifiers
[    2.035311] nvme nvme1: 8/0/0 default/read/poll queues
[    2.036541]  nvme3n1: p1
[    2.038180] nvme nvme1: Ignoring bogus Namespace Identifiers
[    2.038216] nvme nvme9: 12/0/0 default/read/poll queues
[    2.040008]  nvme1n1: p1
[    2.040476]  nvme5n1: p1
[    2.042527] nvme nvme4: 12/0/0 default/read/poll queues
[    2.042671] nvme nvme2: 12/0/0 default/read/poll queues
[    2.045991]  nvme10n1: p1
[    2.047549]  nvme9n1: p1
[    2.049676] nvme nvme16: 12/0/0 default/read/poll queues
[    2.049992] nvme nvme12: 12/0/0 default/read/poll queues
[    2.050257] nvme nvme14: 12/0/0 default/read/poll queues
[    2.052047]  nvme4n1: p1
[    2.052054]  nvme2n1: p1
[    2.052250] nvme nvme15: 12/0/0 default/read/poll queues
[    2.053828] nvme nvme8: 12/0/0 default/read/poll queues
[    2.061272]  nvme16n1: p1
[    2.061652]  nvme12n1: p1
[    2.061853]  nvme14n1: p1
[    2.063987]  nvme15n1: p1
[    2.065560]  nvme8n1: p1
[   31.899620] nvme nvme0: Device not ready; aborting reset, CSTS=0x1
[   31.913566] nvme nvme11: Device not ready; aborting reset, CSTS=0x1
[   31.959619] nvme nvme18: Device not ready; aborting reset, CSTS=0x1
[   31.959619] nvme nvme13: Device not ready; aborting reset, CSTS=0x1