Recently I bought a couple of Dual Socket LGA2011 boards. I was really enamored to see 48 cores available with hyperthreading.
Once that newness wore off I started benchmarking my NVMe with
So I bought a few Single Socket LGA2011 boards to see if the PCIe bus is taking a hit, by spreading IO over 2 CPU's.
Where does NUMA play into this? Is it not very efficient?
The reason I ask is I see CPU benchmarked where a Single CPU shows ~15K but a Dual CPU run only shows ~20K [1]
So the second CPU is only contributing <50%.
Is this just a benchmarking phenomenon or is this really how additional CPU's contribute to load sharing?
Is this just a load related issue? Benchmarking is synthetic. The extra cores are very useful when compiling software.
I wonder how much hit Quad LGA2011 takes.
[1]
Once that newness wore off I started benchmarking my NVMe with
diskinfo -t
and noticed the speeds seemed low.So I bought a few Single Socket LGA2011 boards to see if the PCIe bus is taking a hit, by spreading IO over 2 CPU's.
Where does NUMA play into this? Is it not very efficient?
The reason I ask is I see CPU benchmarked where a Single CPU shows ~15K but a Dual CPU run only shows ~20K [1]
So the second CPU is only contributing <50%.
Is this just a benchmarking phenomenon or is this really how additional CPU's contribute to load sharing?
Is this just a load related issue? Benchmarking is synthetic. The extra cores are very useful when compiling software.
I wonder how much hit Quad LGA2011 takes.
[1]