Message boards : Graphics cards (GPUs) : needed pci-e bandwidth? mining rig idea
Author | Message |
---|---|
does gpugrid really need the full 16x or 8x bandwidth of pci-e 3? | |
ID: 51740 | Rating: 0 | rate: / Reply Quote | |
You can certainly try it to see what GPU usage you get with SWAN_SYNC on Linux. Without SWAN_SYNC I notice about 30% PCIe usage on an 8x link, but with SWAN_SYNC on Linux I notice only about 2% usage. | |
ID: 51741 | Rating: 0 | rate: / Reply Quote | |
You can certainly try it to see what GPU usage you get with SWAN_SYNC on Linux. Without SWAN_SYNC I notice about 30% PCIe usage on an 8x link, but with SWAN_SYNC on Linux I notice only about 2% usage.it is my purpose to "mine" for gpugrid -:) no, just kidding....i do not want to mine any crypto at all. my question was: setting up gpu machine for only 1 purpose = gpugrid, (in a way as miners do). a lot of gpu cards in 1 case, connected to 1x lane pcie-3. is 1 lane of pcie-3 enough to use gpugrid? or i need at least 8 lanes or 16 lanes pcie-3?? if my question or idea not clear enough, plz ask me | |
ID: 51742 | Rating: 0 | rate: / Reply Quote | |
He kind of did answer it. 30% usage at 8x. Theoretically one could run 3 GPUs across an 8x link but I'd wager there would be some performance losses at 90% utilization. 1x would not be enough in that case. | |
ID: 51743 | Rating: 0 | rate: / Reply Quote | |
it is my purpose to "mine" for gpugrid -:) I apologize if my answer was not clear enough. Basically at 8x I use 30% of the 8x bandwidth. But with SWAN_SYNC on Linux I only use 2% of the bandwidth of 8x. So I would imagine you could see less than 80% usage on 1x pcie. I have not tested this myself but I would imagine it would still work just fine with minimal loss. Keep in mind you need 1 CPU thread per GPU unlike mining which is almost 0% reliant on CPU threads. | |
ID: 51744 | Rating: 0 | rate: / Reply Quote | |
He kind of did answer it. 30% usage at 8x. Theoretically one could run 3 GPUs across an 8x link but I'd wager there would be some performance losses at 90% utilization. 1x would not be enough in that case. this whole mining sh*t is making me crazy.....I am not talking about mining....I am talking about building a system like miners do. putting a lot of nvidia gtx-cards in 1 case, connecting them to pcie-3 x1 lane. and then only running boinc for gpugrid, using windows 10 with swansync enabled. no mining program at all should gpugrid run OK on pcie-3 x1 lane? | |
ID: 51745 | Rating: 0 | rate: / Reply Quote | |
i am sorry. i didn't see your reply on my smaal screen smartphone.it is my purpose to "mine" for gpugrid -:) now i see it. 1 cpu thread per 1 pgu. my cpu is threadripper 1950x, 16 core with STM (or SMT) 32 viewable core's. is this enough for 4 gpu's? and...if i understand your answer rigth, then i cann't punt 4 gpu's in pcie-3 x1 lane because of that 30% of x8 lanes. am i right? once again, i am sorry for confusion | |
ID: 51746 | Rating: 0 | rate: / Reply Quote | |
30% on an 8x pcie is without SWAN_SYNC. With SWAN_SYNC on Linux I get 2% on 8x. As long as you use SWAN_SYNC it might be theoretically possible. | |
ID: 51747 | Rating: 0 | rate: / Reply Quote | |
He kind of did answer it. 30% usage at 8x. Theoretically one could run 3 GPUs across an 8x link but I'd wager there would be some performance losses at 90% utilization. 1x would not be enough in that case. I was comparing how mining is different from BOINC crunching and in particular GPUGrid. How mining can get away with using just a 1x slow and how GPUGrid cannot. There ya got it. It won't work. You want a Black and white answer without any understanding, here it is. No Just test it. It's not that hard. | |
ID: 51748 | Rating: 0 | rate: / Reply Quote | |
thanksHe kind of did answer it. 30% usage at 8x. Theoretically one could run 3 GPUs across an 8x link but I'd wager there would be some performance losses at 90% utilization. 1x would not be enough in that case. | |
ID: 51749 | Rating: 0 | rate: / Reply Quote | |
In the past I have used a usb cable connected riser to a 1x pci-e slot using a GTX750ti for a few months on GPUgrid. (I no longer use it) | |
ID: 51750 | Rating: 0 | rate: / Reply Quote | |
The people in my team at Seti who are using mining hardware to support multiple high gpu count hosts (7-12 cards) find the risers and particularly the USB cables have to be high quality and shielded to crunch without constant issues of cards dropping offline. | |
ID: 51751 | Rating: 0 | rate: / Reply Quote | |
The people in my team at Seti who are using mining hardware to support multiple high gpu count hosts (7-12 cards) find the risers and particularly the USB cables have to be high quality and shielded to crunch without constant issues of cards dropping offline. thanks for your reply. to try I firstly need to buy gtx1070 or gtx1080 cards, used ones. the price here for used cards are around 250-300 euro;s. spending 1000 euro for 4 cards to try something is odd for me. that is why asked here before I spend 1000 euro. and I would by 4 gpu's just only for gpugrid, no other intended use. so now I know it will not work because of limitation of speed of x1 lane. thanks all who replied | |
ID: 51752 | Rating: 0 | rate: / Reply Quote | |
The people in my team at Seti who are using mining hardware to support multiple high gpu count hosts (7-12 cards) find the risers and particularly the USB cables have to be high quality and shielded to crunch without constant issues of cards dropping offline. I think your motherboard already has four PCIe3.0x16 slots (probably they will run at x8 if all of them occupied). You should look for x16 PCIe3.0 risers and use them. It is recommended to resolve cooling issues. Or you can build a water cooled rig with 4 cards. The heat output will be around 1.2kW, so it's not recommended to put 4 air cooled cards close to each other. As for the original question: In my experience the performance loss (caused by the lack of PCIe bandwidth) is depends on the workunit (some will suffer more, some will suffer less/none). To achieve optimal performance high end cards need at least PCIe3.0x8. If you are not bothered by the performace loss, perhaps they will work even at PCIe3.0x1. I'm not sure because I've never put more than 4 GPUs in a single host. I build single GPU hosts lately, because I can spread them across our flat (I use them as space heaters). | |
ID: 51753 | Rating: 0 | rate: / Reply Quote | |
Yes, any HEDT motherboard should be able to support 4 gpus natively. I have an Intel X99 board that supports 4 gpus at PCIe 3.0 X16 speeds and one X399 board that supports 2 gpus at PCIe 3.0 X8 speeds along with 2 gpus at PCIe 3.0 X16 speeds. As long as the gpus are no wider than two slots, even air cooled cards fit. Water cooling or hybrid cooled cards keep the cards cool so they clock well with the best performance. | |
ID: 51754 | Rating: 0 | rate: / Reply Quote | |
Hybrid or custom cooling loops is best options for multi card systems. Prevent throttling of the cards if the are all in 1 box. However, if you go the route of hanging them from a support beam above the Mobo, then you could probably get away with air cooled as long as you have proper ventilation. | |
ID: 51756 | Rating: 0 | rate: / Reply Quote | |
my idea is to build a setup with mining rig frames, like this one | |
ID: 51757 | Rating: 0 | rate: / Reply Quote | |
This software is fundamentally different than mining software and requires more resources. You will need at least 1 CPU core per GPU and I highly doubt PCIe 1x is enough bandwidth to supply a fast GPU like a 1070 (PCIe 16x riser on a PCIe 1x slot is the same thing as a 1x riser on a 1x slot.) You will need minimum 4x and recommended 8x PCIe per GPU. | |
ID: 51759 | Rating: 0 | rate: / Reply Quote | |
Yes, any HEDT motherboard should be able to support 4 gpus natively. I have an Intel X99 board that supports 4 gpus at PCIe 3.0 X16 speeds and one X399 board that supports 2 gpus at PCIe 3.0 X8 speeds along with 2 gpus at PCIe 3.0 X16 speeds. As long as the gpus are no wider than two slots, even air cooled cards fit. Water cooling or hybrid cooled cards keep the cards cool so they clock well with the best performance. intel x99 boards have mostly 16x/8x/4x modus. probably you are talking about asus ws-serie mainboards when you say 4 gpu's natively x16. but do you realize about the presence of PLX pcie-switches on those ws-boards? the net result is much lesser bandwidth take a look at block diagram https://www.overclock.net/content/type/61/id/2674289/width/350/height/700/flags/LL my idea was not makng hedt-system with threadripper x399-chipset. way to expensive. but....thanks for your reply. i found someone near to me where i can test my idea with asus b250 mining motherboard, intel g4400 cpu, 8gb ram, 2x gtx1070. i will give update about my progress. | |
ID: 51760 | Rating: 0 | rate: / Reply Quote | |
It's not less, it's better utilization of available lanes. For this example I am only talking about Intel chips . There are only as many lanes as the CPU chip has. If you get a low end CPU with 24 lanes, then that is all you get. If you get a high end CPU then you might have 40 or 44. For the ASUS X99e-ws ($$$$) the PCIe are x16/x16/x16/x16 because of the PLX chip. As long as you don't add other things that take up any of the lanes (m2, etc) that the GPUs are using then you can get close to x16 for the GPUs. The PLX chips have their own lanes as well that allow you to attach other items of your computer (Lan, USB etc). Here's a link to post where someone is attempting to describe what is occurring. He's quoting an article we both read about this a long time ago but can't find right now. https://www.overclock.net/forum/6-intel-motherboards/1618042-what-multiplexing-how-does-plx-chips-work.html ____________ | |
ID: 51761 | Rating: 0 | rate: / Reply Quote | |
so, you recommend to get a motherboard with plx-swtich? is this better than b250 mining bord with x1 lanes? do i understand you correct? a motherboard with plx will crunch faster than x1-lanes? | |
ID: 51762 | Rating: 0 | rate: / Reply Quote | |
I'm saying, PCIe speed does make a difference. You are better off with higher than PCIe x1. What level of PCIe is up to you. X16 will process the data faster than x8 and so forth. | |
ID: 51763 | Rating: 0 | rate: / Reply Quote | |
my idea is to build a setup with mining rig frames, like this one That frame should work. I believe that the ASUS B250 Mining Expert motherboard is the one TBar is using with 12 gpus and a i7-6700. https://www.newegg.com/Product/Product.aspx?Item=9SIA96K7TC8911&Description=B250%20MINING%20EXPERT&cm_re=B250_MINING_EXPERT-_-13-119-028-_-Product His host is here. https://setiathome.berkeley.edu/show_host_detail.php?hostid=6813106 One of his stderr.txt outputs is here showing the 12 gpus. https://setiathome.berkeley.edu/result.php?resultid=7649459019 | |
ID: 51764 | Rating: 0 | rate: / Reply Quote | |
so, you recommend to get a motherboard with plx-swtich?Yes. is this better than b250 mining bord with x1 lanes?Yes. do i understand you correct? a motherboard with plx will crunch faster than x1-lanes?Yes. | |
ID: 51766 | Rating: 0 | rate: / Reply Quote | |
edit...deleted | |
ID: 51767 | Rating: 0 | rate: / Reply Quote | |
ok, ppl. today i have been testing asus b250 mining board. | |
ID: 51768 | Rating: 0 | rate: / Reply Quote | |
I would go with either of the Asrock boards on that site. The Taichi has four X16 PCIe slots so you could fit four gpu cards. The Pro Gaming 6 satisfies your minimum 3 gpu requirement. | |
ID: 51769 | Rating: 0 | rate: / Reply Quote | |
ok, little update on my build project. | |
ID: 51941 | Rating: 0 | rate: / Reply Quote | |
ok, little update on my build project. Hello erik, Nice build! If you ever have any cooling problems make sure to install 120mm fans blowing the heat over and away from the GPUs. Is the 85% GPU load with windows SWAN_SYNC? If so, I believe you will be in the 95%+ range with SWAN_SYNC on Linux once their application is fixed soon. You mention building another system. As far as I am aware the only motherboards with more than 4 full size PCIe slots are workstation or server boards. They typically use PLX chips to act as a PCIe lane 'switch' to the CPU. The old fashioned way to achieve high GPU counts is PCIe x1 but I don't think we've had anyone test the GPU utilization with the GPUGRID app with SWAN_SYNC under Linux. | |
ID: 51943 | Rating: 0 | rate: / Reply Quote | |
forgotten to say: under windows 10 is Swan_sync enabled (set to 1). but having 80% gpu load, I don't mind that. I think the reason for such a low gpu load is my cpu. because my cpu is partially broken. (but this is another issue, asked rma at AMD) | |
ID: 51944 | Rating: 0 | rate: / Reply Quote | |
It's hard to see but is there support under those GPUs or are they just hanging by the PCI bracket? I have this one that 'supports' 6x GPUs but you can easily drill more holes for the PCI brackets. Not even $30 and it comes with fan mounts. | |
ID: 51945 | Rating: 0 | rate: / Reply Quote | |
It's hard to see but is there support under those GPUs or are they just hanging by the PCI bracket?see first picture. those 2 white parallel aluminium bars are the support for riser-slots. the cards are not hanging "in the air". they sit on riser-slot, and riser-slot sits on those parallel bars. all white bars are my own adjustments. I have this one that 'supports' 6x GPUs but you can easily drill more holes for the PCI brackets. Not even $30 and it comes with fan mounts.my cards are itx-format. so putting a fan at back-side of the cards will not help that much to cool the gpu's. if I should hang fans on the backside, then there is about 7-8cm space between the fan and backside of gpu. so I don't expect much cooling effect. I could put fans on the frontside of the gpu's where hdmi cables are connecting. but as of now, I don't see any advantage of this. gpu-temperature is now around 50 degree Celsius. No need to limit yourself to an ATX board for either type of case. Those open air mining rigs are just made out of aluminum t-slot pieces. https://8020.net/shopi am in doubt between those 2 boards, Asus X99-E WS (socket 2011-3) or Asus P9X79-E WS (socket 2011) | |
ID: 51946 | Rating: 0 | rate: / Reply Quote | |
Ah missed the 1st link. 2nd one with fewer images are darker. | |
ID: 51949 | Rating: 0 | rate: / Reply Quote | |
https://edgeup.asus.com/2018/asus-h370-mining-master-20-gpus-one-motherboard-pcie-over-usb/could this board working with gpugrid? because data connection is usb 3.1 gen 1 (5gbps speed). is this enough data speed (usb 3.1) for gpugrid? | |
ID: 51951 | Rating: 0 | rate: / Reply Quote | |
now for testing....connected 1 gpu to pcie-x1 v2 slot, gpu load at 68-70%, bus interface load 65-67%. That is not as bad as I thought, though you did mention you lowered the power limit of the cards a lot. What GPU clockspeed are they running at? I would imagine with a faster card you would get less GPU utilization. Also keep in mind this is Windows SWAN_SYNC and not Linux SWAN_SYNC so I think there is still much performance to be gained with PCIe x1 You might also be able to maximize what you can out of the limited PCIe x1 bandwidth. If you lower the power limit enough, which in turn lowers clockspeed, you could potentially maximize GPU utilization making it not only more efficient from the power limit but also more efficient with the higher utilization. | |
ID: 51952 | Rating: 0 | rate: / Reply Quote | |
now for testing....connected 1 gpu to pcie-x1 v2 slot, gpu load at 68-70%, bus interface load 65-67%. all my cards are the same, msi gtx 1070 8gb itx. all of them are power limited to 50%. one of them is connected to pcie-x1-lane v2.0. this card has: gpu load around 70%, bus interface load 67%, clock speed around 1650-1680 MHz, under power limit 50% all other cards with same power limit 50% are running around 1530-1570MHz, gpu load around 85%. all have the same temperature...around 50-52 degree celsius | |
ID: 51953 | Rating: 0 | rate: / Reply Quote | |
I too have a mining-esque system with hopefully 5-6 gpus. The case is already designed for open air 6 GPUs. My current and hopefully only problem is the cards I currently have require two 6 and 8 pin 12v plugs and I've run out of cables from my power supply. I should have done a bit more research before buying! | |
ID: 51954 | Rating: 0 | rate: / Reply Quote | |
I too have a mining-esque system with hopefully 5-6 gpus. The case is already designed for open air 6 GPUs. My current and hopefully only problem is the cards I currently have require two 6 and 8 pin 12v plugs and I've run out of cables from my power supply. I should have done a bit more research before buying! for psu you cxan use atx-psu with enough power, starting at 1200 watt. preferably full modular, so dont want to have molex or sata power connectors. or use hp server psu with special modules for 12x 6pin pcie-power-connector.https://tweakers.net/aanbod/1983990/mining-starters-kit-benodigdheden.html this site is in dutch. but you can check the pictures about ho server psu. for your build...what kind of motherboard are you using? and cpu? | |
ID: 51955 | Rating: 0 | rate: / Reply Quote | |
I have a 1200 watt AX1200 fully modular from Corsair paired with an AB350 Pro4 from ASRock and a r7 1700. I would not recommend this motherboard as it has caused me great agony with GPU detection as for at least a year after its release it did not have a BIOS that allowed for what I was trying to do. It has 6x PCIe 1x slots as I don't have a need for more than 6 GPUs. I personally would recommend literally any other board. I think a mining specific board would work the best as that is what you will be doing with it. It probably has other mining specific features built in. I have the r7 1700 at full load with World Community Grid and Rosetta@home while also having multiple GPUs at high load. As long as you don't overwhelm the CPU with too much CPU work, everything should run at peak efficiency and speed. | |
ID: 51956 | Rating: 0 | rate: / Reply Quote | |
does gpugrid really need the full 16x or 8x bandwidth of pci-e 3 No - though more lanes will perform better on GPUGRID WUs that are bandwidth heavy. In my GPUGRID experience performance loss varies 33-50% running PCIe 2.0 x1 with GTX 750 / GTX 970 / GTX 1060 / GTX 1070 or any card. My z87 MB with 5 GPUs PCIe 2.0 x1 bus interface load 82% on any card. PCIe 3.0 x4 70-75% bus usage. PCIe 3.0 x8 50-60% bus usage. GTX 970 PCIe2.0 x1 has 55% the performance of a GTX970 on PCIe3.0 x4. GTX 1060 and 1070 PCIe2.0 x1 has 66% the performance compared to 1060 / 1070 PCIe3.0 x4. Turing GPU on PCIe2.0 x1 I suspect will run ACEMD at 70-75% performance of PCIe3.0 x4. On a z87 MSI XPOWER motherboard I have RTX 2070 (PCIe3.0 x8) / RTX 2060 (PCIe3.0 x4) / RTX 2080 (PCIe3.0 x4) with GTX 1080 / 1070 (PCIe2.0 x1) running (integer) Genefer n=20 Primegrid app. x1 PCIe bus shows a 7-10% performance loss compared to PCIe3.0 x4 or 8-13% loss vs. PCIe 3.0 x8. | |
ID: 51958 | Rating: 0 | rate: / Reply Quote | |
i am considering this setup: | |
ID: 51959 | Rating: 0 | rate: / Reply Quote | |
That CPU only has 16 lanes on the CPU so it would be pretty pointless to have a full PCIe connector for each GPU. It would only make sense if there were enough PCIe lanes to go around | |
ID: 51960 | Rating: 0 | rate: / Reply Quote | |
so, your better choice is using mainboards with PLX-switches? | |
ID: 51961 | Rating: 0 | rate: / Reply Quote | |
That CPU only has 16 lanes on the CPU so it would be pretty pointless to have a full PCIe connector for each GPU. It would only make sense if there were enough PCIe lanes to go around The purpose of the full PCIe connector is to give the highest possible mechanical stabilty for all the cards as these x16 connectors have latches on the "inner" end. This latch on the "inner" edge of the PCIe connector is available only on the x16 slot. This provides compatibility with shorter cards, while shorter open end PCIe slots can accomodate longer cards while providing fewer lanes, and no latch on the end. Take a look at the last picture: https://www.fasttech.com/product/9661024-authentic-onda-b250-d8p-d4-btc-mining-motherboard Only the first (the closest to the CPU) PCIe slot has 16 lanes, the others have only 1 lane. According to the Intel specification the lanes of the PCIe controller integrated into the CPU can't be used as 16 separate x1 lanes. See the expansion options: https://ark.intel.com/content/www/us/en/ark/products/191047/intel-core-i7-9850h-processor-12m-cache-up-to-4-60-ghz.html https://ark.intel.com/content/www/us/en/ark/products/135457/intel-pentium-gold-g5620-processor-4m-cache-4-00-ghz.html It goes like: PCI Express Configurations: Up to 1x16, 2x8, 1x8+2x4 for every socket 115X CPU. (or less for low-end Celerons and Pentiums)That is there could be at most 3 GPUs connected to the CPU: 1 on 8 lanes, the 2 other in 4 lanes. However the "South Bridge" chip provides further PCIe (2.0) lanes: (these lanes have higher latency than the the lanes built into the CPU) https://ark.intel.com/content/www/us/en/ark/products/98086/intel-b250-chipset.html PCI Express Revision : 3.0
PCI Express Configurations: x1, x2, x4
Max # of PCI Express Lanes: 12 Perhaps that's the tick: the 1st PCIe is connected with all the 16 lanes to the CPU, the other 11 is connected to the south bridge (1 lane each). There's no unnecessary peripherals (PS/2 keyboard, serial and parallel ports, additional USB ports, sound controller), only one PCIe gigabit network interface controller (occupying the 12th lane). | |
ID: 51962 | Rating: 0 | rate: / Reply Quote | |
ok, yet another update, using x1-slot (on asrock x399 taichi mainboard with threadripper 1950x) | |
ID: 51965 | Rating: 0 | rate: / Reply Quote | |
Turing RTX 2080 PCIe 2.0 x1 performance 50% of PCIe 3.0 x4 on Primegrid n=20 Genefer. | |
ID: 52124 | Rating: 0 | rate: / Reply Quote | |
Wow I think you got the deal of the century, I bought a new EVGA 2080ti from Microcenter for $1015 and I thought that was a pretty good deal. The difference is, my max boost speed seems to be 1800 using 275 watts according to GPU-Z. Pretty amazing you can get 1970Mhz with only 280 watts. | |
ID: 52125 | Rating: 0 | rate: / Reply Quote | |
i will not use pcie-x1 slot. | |
ID: 52126 | Rating: 0 | rate: / Reply Quote | |
Thanks - its an amazing price. Someone returned an efficient chip. Even with the Turing refresh on the horizon soon and price drops - I couldn't pass a 2080ti at such cost. I always check newegg and Microcenter for open box deals. The 2080ti was returned the night before being purchased on sale for 979usd. I saw 2080ti deal online early this morning then walked in when the store opened. | |
ID: 52127 | Rating: 0 | rate: / Reply Quote | |
Message boards : Graphics cards (GPUs) : needed pci-e bandwidth? mining rig idea