Like you said! IF anyone can offer a recommendation for a optimized system for crunching I am all ears!
Like you I am into a single GPU system. I guess the discussion after that is which CPU offers the most bang for the buck?
I am super impressed with the i5-4590 from a gamers perspective. YES you can pay more for the overclocked i5 BUT in my experience it doesn't deliver the difference in performance. PLUS I don't overclock, and that is a major area of discussion for gaming.
SO is it the i3 or i5 for a single GPU? Which GPU offers the most bang for the buck? FOR a gamer, it comes down to which Monitor you are willing to buy? IF you can only afford to game at 1080P then my configuration of i5 4590 and nvidia 960 is hard to beat cost and power wise.
OH YEAH! Woo Hoo we hit 40k on the daily crunching for the first time!!! And that is with the system only running about 18 hours a day or less. I turn off crunching when I get on the computer to play.
AGAIN A Big Thank You For All Your Help
Mike Brown
I won't pretend to know optimum, but for people interested in economy in both purchase and operating cost, I think that a quite inexpensive Haswell dual-core CPU hosting one or two GTX 750Ti cards is something worth setting as a comparison reference to try to decide why a specific configuration under consideration is better for a particular user's interests.
Of course, neither that CPU nor that GPU will give a romping stomping system competing up near the top of the RAC rankings, but if one is careful on spending on the other components--or better yet is in a position to re-use existing case/power supply/disk hardware, then the purchase price and operating cost for such a choice can be rather far below most of the leading configurations.
My biggest concern with actively suggesting either of these parts (as distinct from proposing them as a reference comparison) is that they are at the point of likely being superseded. In the normal Intel progression, most likely some Skylake offering may soon offer somewhat better speed/power/price tradeoff than Haswell, or quite well may already. The Nvidia sequence has taken an oddly long pause at the 28 nm TSMC process "node", so the very next offering, slated to be Pascal series parts on the TSMC 14 nm finFET "node" may offer a considerable improvement in performance at a given power consumption point.
Skylake is here and now, though somewhat early days. Pascal is still in preshipment hints, winks, nods, and sightings of alleged prototype sample chips in India customs paperwork. I've seen no credible claim as to whether the first Pascal part offered will be the "big chip" probably meant to displace GTX970/980 or a "small chip" meant to display the GTX750/750Ti. In the previous generation there was considerable delay between these, even though the supplier presumably had plenty of fab capacity. This time TSMC finFET capacity and yields may both influence this choice. Still, some Pascal flavor will surely ship in 2016, or Nvidia has had a big setback.
In counting the cost, one must not neglect the power supply, though sizing it properly is a bit vexing. An advantage of a single 750Ti system on a dual-core Haswell is that very likely any supply you might already have may be sufficient. Even a dual 750 system does not need a very capable supply at all (and if I were selecting a motherboard for such a system today, I'd try to find a economical one which had two side PCIe slots, just to protect the possibility of going double later).
I used to be somewhat skeptical of multiple GPU systems, believing from some data on other people's systems that the extra GPUs were so slowed below maximum output by sharing host support as to deplete the economies considerably. But the current Einstein CUDA55 Parkes PMPS application seems to "share" very well indeed (at least if you don't load up the CPUs with BOINC CPU tasks--I run zero), and from both a purchase economy and an operating cost point of view, sharing the host cost over two cards is a very big advantage.
Of course, this may all change with another application release, and all tradeoffs may be different for work from another project, gaming, or just a different application on this project.
WOW your dual 750ti setup is putting out serious RAC! My system does not run 24/7, probably at best 18 hours on a good day. I don't track it. I just hit 40k for the first time with the changes that Gary has helped me with. IT seems GPU's are the bomb and IF you can put together a dual gpu it is way more powerful for crunching. My board only has one slot, boo Hoo!
I might have to see IF i can pick up a dual slot intel 1150 board. Then I could use my cpu & ram from this board. Maybe later.
WOW your dual 750ti setup is putting out serious RAC!
A word of caution in reviewing multiple-GPU system performance:
You cannot rely on the identification of the GPUs shown on the computer details page to be accurate, as it just lists ONE of the types present, and the total number. For CUDA systems, if there are GPUs at different capability levels, it will show the type of the highest capability--which means the most modern, NOT the fastest.
In my particular case I own three systems, each of which has two GPUs installed, but only one of them actually has two GPUs of the same type--the one shown has having two GTX750 Ti units really does.
However the one shown as being two GTX970 is actually one 970 plus one 750 Ti, and the one shown as being two GTX750 is actually one GTX750 plus one GTX660.
Still, it is true that my dual 750Ti system has substantial output, and that one actually has the components shown on the web page. I have the GPU units overclocked in both core and memory clock quite close to their error limit, the host runs Einstein exclusively 24/7, and has almost no interactive use.
That was my design decision 3 years ago. I did say that I would most likely go for two GPUs now.
Back then, the GPU apps here were very hungry for PCIe bandwidth. Two or more GPUs would be sharing that bandwidth so the potential was there to have an adverse effect on the crunch times for each GPU. There have been changes to the apps that reduce the bandwidth requirements. Its not a problem any more.
There have been improvements in power consumption for both CPUs and GPUs. You can't really rely on what worked well three years ago. I would have no qualms in agreeing with a design just like archae86 mentioned - a Pentium dual core with two 750Tis.
The motherboard you choose is probably a more important decision than the CPU. With the latest apps now much less reliant on high PCIe bandwidth, you don't need the very expensive boards that allowed the full complement of PCIe lanes (x16) on each slot. I'm sure x8/x8 would give you full performance on both cards. I've noticed some quite cheap boards that are shown as x16/x4 in the specs. I don't know if they could be configured as x8/x8 when two similar cards are installed. Maybe the x4 slot might still be OK for the performance of the second card. These are the sorts of questions you need to address. Maybe archae86 might comment on the boards he uses.
In the future, particularly at Einstein, the importance of CPUs is going to continue to diminish. Eventually, I would expect most (if not all) science runs here to have a high performance GPU app available. It might take a year or so but it should figure in your planning for a crunching box. If you're serious about crunching in the future, save on the CPU cost so that you can spend on a GPU upgrade down the track a bit. That's where you will get most benefit from your spend.
A Pentium dual core is perfectly adequate to service two 750Tis each crunching at least 2x. I would expect that you could crunch CPU tasks on at least one of the cores, maybe even both with little effect on GPU crunch time. It's actually quite interesting to experiment and find out. You would probably get the best credit output to power consumption ratio by disabling CPU crunching completely. So, it really depends on what is most important to you - saving electricity or helping to find a new Gamma-ray Pulsar ;-).
(Pssst - don't look now but I think you've really caught the crunching bug!!) :-).
(Pssst - don't look now but I think you've really caught the crunching bug!!) :-).
Looks like it! It slightly depends whether he's caught the generic 'crunching' bug, or the specific 'einsteinian' variant of it.
My personal preference for the last few years has been - for projects where there is a choice, like here - to crunch exclusively on GPUs. And then, to attach to projects which only have CPU applications, and let the CPUs work there. For preference, I run projects which I observe to place a relatively light load on the CPU - perhaps predominantly integer maths, rather than heavily optimised FPU code. I recently gave quite a big share to FightNeglectedDieseases@Home - I liked the style of Ben, the self-described 'lonely sysadmin' on the front page, and was pleased to help where I could. At the moment, that project is in hiatus, while the previous science run is analysed and Ben writes up his PhD thesis. I hope they come back, but I'm sure there are other similar projects out there.
It's probably in the context of that sort of cross-project working that I suggested four full cores, rather than dual-core with HT. I don't think there's a big difference in either purchase price or electricity consumption, these days.
The motherboard you choose is probably a more important decision than the CPU. With the latest apps now much less reliant on high PCIe bandwidth, you don't need the very expensive boards that allowed the full complement of PCIe lanes (x16) on each slot. I'm sure x8/x8 would give you full performance on both cards. I've noticed some quite cheap boards that are shown as x16/x4 in the specs. I don't know if they could be configured as x8/x8 when two similar cards are installed. Maybe the x4 slot might still be OK for the performance of the second card. These are the sorts of questions you need to address. Maybe archae86 might comment on the boards he uses.
My most recent build in early 2014 used an ASROCK Z87 EXTREME3 motherboard to host a dual-core Haswell. At the time of purchase the price was $115 US, so not a true economy board but not remotely high-end. The full-length PCIe slots on that one configure as x8/x8 with one GTX970 and one GTX750Ti plugged in. I no longer recall what considerations drove me to that board rather than one of the sub $100 boards at the time.
I recently reduced all three of my main crunching machines from running just one CPU job to running zero, expecting to get a power efficiency of computing gain at some modest loss in total Einstein production.
I did indeed get power efficiency improvement from each system, but on the dual core Haswell system in question I was quite surprised to see sufficient output increase from the GPUs to actually get a noticeable net Einstein production increase.
One possible interpretation of that result is that my dual-core Haswell on that motherboard is not quite fully keeping up with the needs of the combination of a 970 and a 750Ti, though I think it is doing pretty well.
Experimentation is key, as there are enough variations in system configurations and applications to make predicting your results from others' experience (or even your own experience) surprisingly iffy.
There are some very low-priced Skylake CPUs sold under the Pentium brand these days. I recently reviewed a list of the capabilities disabled by the manufacturer on these chips as compared to i3/i5/i7 chips sold from the same die, and got the tentative answer that for my purposes they would probably be entirely adequate. "Pentium", "Core", i3, i5, i7, and even Xeon are just brands, and as with, say, Ford/Mercury/Lincoln, the names are not a durably reliable indicator of the detailed characteristics of the products sold under that name in any given year. Anyone telling you what a Skylake Pentium is like based on their memory of ancient products sold under that name is speaking through their hat.
Other interesting thought, you have a two 750 and a two 750ti setup and they are showing about the same RAC per day. 750 has 1 gig of ram and 512 cores the 750ti has 2 gigs of ram and 640 cores??? What gives with that? The ti should be significantly faster?
750ti cards are down to 109 bucks, 1150 z97 mb 100 bucks, 69 buck pentium 3258 cpu, 40 for 8 gigs of ram, 50 for a hard drive, 50 buck 500 watt power supply.
I think IF I am going to invest in more hardware, it sounds like going to the above setup with a linux OS might be the cheapest. Plus I like the idea of a system that runs 24/7 doing crunching only.
Good Morning Gary, Like
)
Good Morning Gary,
Like you said! IF anyone can offer a recommendation for a optimized system for crunching I am all ears!
Like you I am into a single GPU system. I guess the discussion after that is which CPU offers the most bang for the buck?
I am super impressed with the i5-4590 from a gamers perspective. YES you can pay more for the overclocked i5 BUT in my experience it doesn't deliver the difference in performance. PLUS I don't overclock, and that is a major area of discussion for gaming.
SO is it the i3 or i5 for a single GPU? Which GPU offers the most bang for the buck? FOR a gamer, it comes down to which Monitor you are willing to buy? IF you can only afford to game at 1080P then my configuration of i5 4590 and nvidia 960 is hard to beat cost and power wise.
OH YEAH! Woo Hoo we hit 40k on the daily crunching for the first time!!! And that is with the system only running about 18 hours a day or less. I turn off crunching when I get on the computer to play.
AGAIN A Big Thank You For All Your Help
Mike Brown
I think the Core-i ranges
)
I think the Core-i ranges come in a number of core-thread combinations. For crunching, you probably want 4 full cores without hyperthreading.
I won't pretend to know
)
I won't pretend to know optimum, but for people interested in economy in both purchase and operating cost, I think that a quite inexpensive Haswell dual-core CPU hosting one or two GTX 750Ti cards is something worth setting as a comparison reference to try to decide why a specific configuration under consideration is better for a particular user's interests.
Of course, neither that CPU nor that GPU will give a romping stomping system competing up near the top of the RAC rankings, but if one is careful on spending on the other components--or better yet is in a position to re-use existing case/power supply/disk hardware, then the purchase price and operating cost for such a choice can be rather far below most of the leading configurations.
My biggest concern with actively suggesting either of these parts (as distinct from proposing them as a reference comparison) is that they are at the point of likely being superseded. In the normal Intel progression, most likely some Skylake offering may soon offer somewhat better speed/power/price tradeoff than Haswell, or quite well may already. The Nvidia sequence has taken an oddly long pause at the 28 nm TSMC process "node", so the very next offering, slated to be Pascal series parts on the TSMC 14 nm finFET "node" may offer a considerable improvement in performance at a given power consumption point.
Skylake is here and now, though somewhat early days. Pascal is still in preshipment hints, winks, nods, and sightings of alleged prototype sample chips in India customs paperwork. I've seen no credible claim as to whether the first Pascal part offered will be the "big chip" probably meant to displace GTX970/980 or a "small chip" meant to display the GTX750/750Ti. In the previous generation there was considerable delay between these, even though the supplier presumably had plenty of fab capacity. This time TSMC finFET capacity and yields may both influence this choice. Still, some Pascal flavor will surely ship in 2016, or Nvidia has had a big setback.
In counting the cost, one must not neglect the power supply, though sizing it properly is a bit vexing. An advantage of a single 750Ti system on a dual-core Haswell is that very likely any supply you might already have may be sufficient. Even a dual 750 system does not need a very capable supply at all (and if I were selecting a motherboard for such a system today, I'd try to find a economical one which had two side PCIe slots, just to protect the possibility of going double later).
I used to be somewhat skeptical of multiple GPU systems, believing from some data on other people's systems that the extra GPUs were so slowed below maximum output by sharing host support as to deplete the economies considerably. But the current Einstein CUDA55 Parkes PMPS application seems to "share" very well indeed (at least if you don't load up the CPUs with BOINC CPU tasks--I run zero), and from both a purchase economy and an operating cost point of view, sharing the host cost over two cards is a very big advantage.
Of course, this may all change with another application release, and all tradeoffs may be different for work from another project, gaming, or just a different application on this project.
Hey archae86, WOW your
)
Hey archae86,
WOW your dual 750ti setup is putting out serious RAC! My system does not run 24/7, probably at best 18 hours on a good day. I don't track it. I just hit 40k for the first time with the changes that Gary has helped me with. IT seems GPU's are the bomb and IF you can put together a dual gpu it is way more powerful for crunching. My board only has one slot, boo Hoo!
I might have to see IF i can pick up a dual slot intel 1150 board. Then I could use my cpu & ram from this board. Maybe later.
Thanks for the info
Mike Brown
This thread about getting
)
This thread about getting started with GPUs, is worth referencing, probably worthy of stickification.
RE: WOW your dual 750ti
)
A word of caution in reviewing multiple-GPU system performance:
You cannot rely on the identification of the GPUs shown on the computer details page to be accurate, as it just lists ONE of the types present, and the total number. For CUDA systems, if there are GPUs at different capability levels, it will show the type of the highest capability--which means the most modern, NOT the fastest.
In my particular case I own three systems, each of which has two GPUs installed, but only one of them actually has two GPUs of the same type--the one shown has having two GTX750 Ti units really does.
However the one shown as being two GTX970 is actually one 970 plus one 750 Ti, and the one shown as being two GTX750 is actually one GTX750 plus one GTX660.
Still, it is true that my dual 750Ti system has substantial output, and that one actually has the components shown on the web page. I have the GPU units overclocked in both core and memory clock quite close to their error limit, the host runs Einstein exclusively 24/7, and has almost no interactive use.
RE: Like you I am into a
)
That was my design decision 3 years ago. I did say that I would most likely go for two GPUs now.
Back then, the GPU apps here were very hungry for PCIe bandwidth. Two or more GPUs would be sharing that bandwidth so the potential was there to have an adverse effect on the crunch times for each GPU. There have been changes to the apps that reduce the bandwidth requirements. Its not a problem any more.
There have been improvements in power consumption for both CPUs and GPUs. You can't really rely on what worked well three years ago. I would have no qualms in agreeing with a design just like archae86 mentioned - a Pentium dual core with two 750Tis.
The motherboard you choose is probably a more important decision than the CPU. With the latest apps now much less reliant on high PCIe bandwidth, you don't need the very expensive boards that allowed the full complement of PCIe lanes (x16) on each slot. I'm sure x8/x8 would give you full performance on both cards. I've noticed some quite cheap boards that are shown as x16/x4 in the specs. I don't know if they could be configured as x8/x8 when two similar cards are installed. Maybe the x4 slot might still be OK for the performance of the second card. These are the sorts of questions you need to address. Maybe archae86 might comment on the boards he uses.
In the future, particularly at Einstein, the importance of CPUs is going to continue to diminish. Eventually, I would expect most (if not all) science runs here to have a high performance GPU app available. It might take a year or so but it should figure in your planning for a crunching box. If you're serious about crunching in the future, save on the CPU cost so that you can spend on a GPU upgrade down the track a bit. That's where you will get most benefit from your spend.
A Pentium dual core is perfectly adequate to service two 750Tis each crunching at least 2x. I would expect that you could crunch CPU tasks on at least one of the cores, maybe even both with little effect on GPU crunch time. It's actually quite interesting to experiment and find out. You would probably get the best credit output to power consumption ratio by disabling CPU crunching completely. So, it really depends on what is most important to you - saving electricity or helping to find a new Gamma-ray Pulsar ;-).
(Pssst - don't look now but I think you've really caught the crunching bug!!) :-).
Cheers,
Gary.
RE: (Pssst - don't look now
)
Looks like it! It slightly depends whether he's caught the generic 'crunching' bug, or the specific 'einsteinian' variant of it.
My personal preference for the last few years has been - for projects where there is a choice, like here - to crunch exclusively on GPUs. And then, to attach to projects which only have CPU applications, and let the CPUs work there. For preference, I run projects which I observe to place a relatively light load on the CPU - perhaps predominantly integer maths, rather than heavily optimised FPU code. I recently gave quite a big share to FightNeglectedDieseases@Home - I liked the style of Ben, the self-described 'lonely sysadmin' on the front page, and was pleased to help where I could. At the moment, that project is in hiatus, while the previous science run is analysed and Ben writes up his PhD thesis. I hope they come back, but I'm sure there are other similar projects out there.
It's probably in the context of that sort of cross-project working that I suggested four full cores, rather than dual-core with HT. I don't think there's a big difference in either purchase price or electricity consumption, these days.
RE: The motherboard you
)
My most recent build in early 2014 used an ASROCK Z87 EXTREME3 motherboard to host a dual-core Haswell. At the time of purchase the price was $115 US, so not a true economy board but not remotely high-end. The full-length PCIe slots on that one configure as x8/x8 with one GTX970 and one GTX750Ti plugged in. I no longer recall what considerations drove me to that board rather than one of the sub $100 boards at the time.
I recently reduced all three of my main crunching machines from running just one CPU job to running zero, expecting to get a power efficiency of computing gain at some modest loss in total Einstein production.
I did indeed get power efficiency improvement from each system, but on the dual core Haswell system in question I was quite surprised to see sufficient output increase from the GPUs to actually get a noticeable net Einstein production increase.
One possible interpretation of that result is that my dual-core Haswell on that motherboard is not quite fully keeping up with the needs of the combination of a 970 and a 750Ti, though I think it is doing pretty well.
Experimentation is key, as there are enough variations in system configurations and applications to make predicting your results from others' experience (or even your own experience) surprisingly iffy.
There are some very low-priced Skylake CPUs sold under the Pentium brand these days. I recently reviewed a list of the capabilities disabled by the manufacturer on these chips as compared to i3/i5/i7 chips sold from the same die, and got the tentative answer that for my purposes they would probably be entirely adequate. "Pentium", "Core", i3, i5, i7, and even Xeon are just brands, and as with, say, Ford/Mercury/Lincoln, the names are not a durably reliable indicator of the detailed characteristics of the products sold under that name in any given year. Anyone telling you what a Skylake Pentium is like based on their memory of ancient products sold under that name is speaking through their hat.
How do you get BOINC to see
)
How do you get BOINC to see the separate GPU's?
Other interesting thought, you have a two 750 and a two 750ti setup and they are showing about the same RAC per day. 750 has 1 gig of ram and 512 cores the 750ti has 2 gigs of ram and 640 cores??? What gives with that? The ti should be significantly faster?
750ti cards are down to 109 bucks, 1150 z97 mb 100 bucks, 69 buck pentium 3258 cpu, 40 for 8 gigs of ram, 50 for a hard drive, 50 buck 500 watt power supply.
I think IF I am going to invest in more hardware, it sounds like going to the above setup with a linux OS might be the cheapest. Plus I like the idea of a system that runs 24/7 doing crunching only.
Thanks again for all the help
Mike Brown