This isn't possible for first generation HBM. Looking at pictures of Fiji, it may be possible to fit another stack on both sides of the die, giving 6 GB of HBM.
In addition, 4-Hi HBM1 16Gb (2GB per chip) and 8-Hi HBM1 32Gb (8GB per chip) is possible and HBM2 will get double the bandwidth and density.
Would the additional memory require a change to Fiji, as in some modifications to the memory controller on the die to address the additional memory? If it were possible, AMD could come out with a professional version of the card which typically have >= 8 GB of memory and a significant cost increase. Otherwise, AMD tends not to release variants of the same architecture.You don't add another stack. You stack the dies higher. 8-Hi is technically possible in HBM1
http://www.fudzilla.com/news/memory/37308-hbm-1-memory-of-amd-fiji-pictured
It isn't very affordable. I don't think it is a linear increase in cost because the complexity is higher also. HBM2 increases in density means can have more memory in a shorter stack ( or just as "short" at 4-Hi).
There is only one 'consumer' of HBM1 now ( so fewer HBM 4-Hi buy the discounts aren't as good) and a substantially more expensive card isn't going to make much headway with Fiji.
It makes far more sense for AMD to work on asynchronous/concurrent main memory to HBM data tranfers using the "extra" bandwidth they have that the current cores can't use. That is a better spend than on super expensive RAM variant.
It turns out that post may be wrong. BIOS in Fury X says the TDP for that card is 275W. So reducing it by 40% gets it to around... 175W. Exactly the same as Fury Nano. What staggers me is reducing TDP by 100W reduced the max clock by only 15 MHz. Also the power consumption is nearly ideally reflected in the effects, because even if Fury X has 275W at stock on average it draws 246W http://tpucdn.com/reviews/AMD/R9_Fury_X/images/power_average.gifRelevant to the discussion is this forum post here, where someone tries to limit the power consumption of a Fury X to potential Fury Nano limits and sees very little decrease in performance. The catch here is I don't see any great measures of power consumption, so who knows if limiting power through the graphics drivers is actually doing anything.
It turns out that post may be wrong. BIOS in Fury X says the TDP for that card is 275W. So reducing it by 40% gets it to around... 175W. Exactly the same as Fury Nano. What staggers me is reducing TDP by 100W reduced the max clock by only 15 MHz. Also the power consumption is nearly ideally reflected in the effects, because even if Fury X has 275W at stock on average it draws 246W http://tpucdn.com/reviews/AMD/R9_Fury_X/images/power_average.gif
Peaks at 280W. It makes sense.
However, I think I know why AMD did not pushed Fiji GPUs even further. It is because I believe they are worried about VRMs on the GPU. They can get to 100 degrees at water cooling. And that can be a problem, even if they are designed to get to 150 degrees C. For those who don't know VRM's are Voltage Regulators. Now you can see why AMD blocked at first the ability to change voltage.
...
There are other manufacturers out there that will happily sell you a powerful workstation.
Maybe you could even go Hackintosh, you obviously know your way around the hardware and firmware, make yourself a custom Pro machine.
...
One of the comments in:Power efficiency is an oft-used negative against the large-die Hawaii chips, but I've been playing with powertune settings and Furmark recently as an experiment to fit a "hot and noisy" AMD card into an SFF with limited cooling.
Actually, I stand by an earlier post I made that says I think AMD pushed Hawaii silicon too far.
With both GPU-Z and Furmark able to report power consumptions, I can see a 100W reduction in power consumption on 290X cards for as little as 5% performance loss.
If you have a Hawaii card, I urge you to crank power limits down in the overdrive tab of CCC and see what the resulting clockspeed is under full load. Even in a worst-case scenario, I'm seeing a typical clockspeed of 850MHz with the slider all the way to the left at -50%
That means that Hawaii (the two samples I personally own, at least) can run at 850+MHz on only 145W (half the 290W TDP). As mentioned, that's a worst-case scenario using a power-virus like Furmark. Under real gaming situations (I was messing around with Alien Isolation on 1440p ultra settings) the clocks averaged about 925MHz yet my PC was inaudible; Fans that normally hum along at 55% were barely spinning at 30% during my gameplay.
As Nvidia has proved, you can make a 28nm chip run efficiently. I think the design of Hawaii holds up very well under vastly reduced power constraints - AMD just pushed it outside its comfort zone in order to get the most out of it.
In saying that, the "underpowered" 290X is around the same performance as my GTX970 and also the same cost - significantly higher than a GTX960 4GB. I don't know if die-harvested 290 cards deal with power limit caps like the cherry-picked 290X cards.
However, the Memory is not OCed. And it brings massive differences to performance.
but definitely include CUDA GPUs as options.
you mean people that can't use apple because of cuda?It would seem to make more sense to keep the people who can't use the MP6,1 in the Apple fold, rather than drive them to Windows, Linux or Hackintoshes.
anyway.. apple can't have these apps being re-written with cuda..
This is one of the most ignorant statements that I've seen on MacRumours, and I've been here for more than 15 years.you mean people that can't use apple because of cuda?
all 6 of them?
probably not a big deal to drive them away..trying to keep those 6 people happy would result in much more serious losses down the line.
i'm talking about rewriting cpu intensive applications into gpu intensive apps.Talking about rewriting OpenCL apps for CUDA is nonsense
likewise, how about you step out of yours.. maybe you can post some stuff about the magical world of cuda that's not an nvidia advertisementStep out of your reality distortion field and look at what is happening in the real world on small systems with +25K CUDA cores. http://info.nvidianews.com/index.php/email/emailWebview?mkt_tok=3RkMMJWWfF9wsRolvavJZKXonjHpfsX57ekrXaayhYkz2EFye+LIHETpodcMScNgPa+TFAwTG5toziV8R7HAKs1v3NsQXBXg
Aiden, do you really see Apple rebranding anything? They're too proud of their own development.
it's like you're arguing "what's better.. python or ruby?".
it's nvidia that doesn't give you the choice.. not apple.No, I'm asking for the having the choice between Python and Ruby.
The MP6,1 doesn't give you the option of choosing between GPGPU programming models.