Unfortunately the box everything is being hosted on does not have a non SSD drive - boot drive is NVME (for Proxmox), but it is not exposed to VMs. Everything is on the ZFS pool.
When I say the drives are fast - I mean it - it's fast - I had to increase to 50 gigs because it was finishing too fast lol
Code: Select all
[email protected]:/home/user $ dd if=/dev/zero of=~/test.log bs=1024k count=50000 status=progress oflag=direct
3098542080 bytes (3.1 GB, 2.9 GiB) copied, 2 s, 1.5 GB/s
6128926720 bytes (6.1 GB, 5.7 GiB) copied, 4 s, 1.5 GB/s
9187622912 bytes (9.2 GB, 8.6 GiB) copied, 6 s, 1.5 GB/s
12233736192 bytes (12 GB, 11 GiB) copied, 8 s, 1.5 GB/s
15272509440 bytes (15 GB, 14 GiB) copied, 10 s, 1.5 GB/s
18355322880 bytes (18 GB, 17 GiB) copied, 12 s, 1.5 GB/s
19863175168 bytes (20 GB, 18 GiB) copied, 13 s, 1.5 GB/s
23037214720 bytes (23 GB, 21 GiB) copied, 15 s, 1.5 GB/s
24611127296 bytes (25 GB, 23 GiB) copied, 16 s, 1.5 GB/s
27704426496 bytes (28 GB, 26 GiB) copied, 18 s, 1.5 GB/s
29256318976 bytes (29 GB, 27 GiB) copied, 19 s, 1.5 GB/s
30762074112 bytes (31 GB, 29 GiB) copied, 20 s, 1.5 GB/s
32330743808 bytes (32 GB, 30 GiB) copied, 21 s, 1.5 GB/s
33921433600 bytes (34 GB, 32 GiB) copied, 22 s, 1.5 GB/s
35446063104 bytes (35 GB, 33 GiB) copied, 23 s, 1.5 GB/s
38507905024 bytes (39 GB, 36 GiB) copied, 25 s, 1.5 GB/s
40066088960 bytes (40 GB, 37 GiB) copied, 26 s, 1.5 GB/s
41653633024 bytes (42 GB, 39 GiB) copied, 27 s, 1.5 GB/s
43199234048 bytes (43 GB, 40 GiB) copied, 28 s, 1.5 GB/s
46262124544 bytes (46 GB, 43 GiB) copied, 30 s, 1.5 GB/s
47822405632 bytes (48 GB, 45 GiB) copied, 31 s, 1.5 GB/s
49347035136 bytes (49 GB, 46 GiB) copied, 32 s, 1.5 GB/s
50896830464 bytes (51 GB, 47 GiB) copied, 33 s, 1.5 GB/s
50000+0 records in
50000+0 records out
52428800000 bytes (52 GB, 49 GiB) copied, 33.9484 s, 1.5 GB/s
And then within the docker container:
Code: Select all
[email protected]:/# dd if=/dev/zero of=/test bs=1024k count=10000 status=progress oflag=direct
2685403136 bytes (2.7 GB, 2.5 GiB) copied, 2 s, 1.3 GB/s
4049600512 bytes (4.0 GB, 3.8 GiB) copied, 3 s, 1.3 GB/s
5384437760 bytes (5.4 GB, 5.0 GiB) copied, 4 s, 1.3 GB/s
6708789248 bytes (6.7 GB, 6.2 GiB) copied, 5 s, 1.3 GB/s
8039432192 bytes (8.0 GB, 7.5 GiB) copied, 6 s, 1.3 GB/s
9289334784 bytes (9.3 GB, 8.7 GiB) copied, 7 s, 1.3 GB/s
And on the proxmox host itself on the zfs pool (g'damn it's 2x as fast lol):
Code: Select all
[email protected]:/ZFS-RAID10# dd if=/dev/zero of=./test bs=1024k count=50000 status=progress oflag=direct
2713714688 bytes (2.7 GB, 2.5 GiB) copied, 1 s, 2.7 GB/s
5492441088 bytes (5.5 GB, 5.1 GiB) copied, 2 s, 2.7 GB/s
8229224448 bytes (8.2 GB, 7.7 GiB) copied, 3 s, 2.7 GB/s
11011096576 bytes (11 GB, 10 GiB) copied, 4 s, 2.8 GB/s
13840154624 bytes (14 GB, 13 GiB) copied, 5 s, 2.8 GB/s
16658726912 bytes (17 GB, 16 GiB) copied, 6 s, 2.8 GB/s
19421724672 bytes (19 GB, 18 GiB) copied, 7 s, 2.8 GB/s
22183673856 bytes (22 GB, 21 GiB) copied, 8 s, 2.8 GB/s
24935137280 bytes (25 GB, 23 GiB) copied, 9 s, 2.8 GB/s
27698135040 bytes (28 GB, 26 GiB) copied, 10 s, 2.8 GB/s
30547116032 bytes (31 GB, 28 GiB) copied, 11 s, 2.8 GB/s
33376174080 bytes (33 GB, 31 GiB) copied, 12 s, 2.8 GB/s
36197892096 bytes (36 GB, 34 GiB) copied, 13 s, 2.8 GB/s
39044775936 bytes (39 GB, 36 GiB) copied, 14 s, 2.8 GB/s
41879076864 bytes (42 GB, 39 GiB) copied, 15 s, 2.8 GB/s
44668289024 bytes (45 GB, 42 GiB) copied, 16 s, 2.8 GB/s
47420801024 bytes (47 GB, 44 GiB) copied, 17 s, 2.8 GB/s
50217353216 bytes (50 GB, 47 GiB) copied, 18 s, 2.8 GB/s
So back to "performance" bottle necks - outside of docker, I get about 170k on pystone with the VM having 6 cores assigned to it; on "bare" metal; (i.e. on the PVE host itself), I get about 175k - so about a 3% loss for the overhead of running in a VM; not bad. However, not sure if it's a "good" or "bad" score for a Xeon D-1528. For comparison - I ran it on an AWS box with a E5-2686 with 8 provisioned cores and I got 190k - so I assume 170k for a D-1528 is decent.
I did as you suggested - did the 10 gig download test via the wrench and, at about the 6.5 gig mark, it slowed down to about 4.5 mb/s - so my previous 15 gig threshold was incorrect as it's actually 10 (I guess I wasn't paying attention). While the download was progressing, load was at about 0.3, and used cache would climb to about 30-35 and drop down to 10 and sort of yo-yo there... not sure if that's expected or...? In any case- during the slowdown I checked ARC stats and if there was any other IO going on - no red flags there at all. After it was done - I clicked on the wrench icon and did not see "download speed limited" message
So then I decided to try sab on the bare VM (not docker) and see what happens... and it too was capped at 50 mb/s on a 10 gig download test (direct unpack is not enabled). However unlike docker, however, it did not slow down to a crawl at the 6.5 gig mark. On bare metal, pystone test within sab was at about 150k and but internet bandwidth comes up as only 21 mb./s... disk write is 570 mb/s to the incomplete folder and 700 mb/s to the complete folder. On bare VM, article cache is set to 1G, and I've played with connection counts - 5, 10, 25, 50... speed is still the same.
Any thoughts?
Thanks.