Apple likes surprises - the GM204 chips have good OpenCL performance and sip electricity. The nnMP could have OpenCL+CUDA!
Apple likes to surprise customers. Apple does not like suprises from their suppliers. It isn't symmetrical. Apple also doesn't like suppliers who don't get on board with their strategic objectives.
While AMD, Intel, Imagination Tech (PowerVR) , and just about everyone except Nvidia is rolling out OpenCL 2.0 GPUs for the late 2014-2015 era, Nvidia is kneecapping OpenCL.
Apple was OK when Nvidia was putting about as much effort into OpenCL as it was CUDA. Now that Nvidia is on the track of "OpenCL has to loose for CUDA to win" tactics, they are not given themselves a competitive edge with Apple. Even if Apple likes the lower power prospects, they aren't going to shoot OpenCL in the head to get to it.
There is an established set of OpenCL 1.1 benchmarks that Nvidia sales marketing will increasing play misdirection games with: "Oh look we're now faster on OpenCL from 5 years ago ... we are 100% on board".
It is just another phase to the lure folks into the proprietary tarpit that want to draw more customers into.
You're responding to someone who's been bad-mouthing the Haswell-EP for months.
The Mac Pro isn't just the Xeon processor ( or the GPU or the Thunderbolt controller(s)). it is a collection. Whether a new system is better than an older system is multidimensional.
The long term trend is that the Xeon E5 (and up ) models are going to be more design optimized for more cores at good speeds far more than fewer cores at the fastest speeds. Intel has kneedcapped the E5 1620 v2 and 1630 v3 a bit. 4 cores (with 4 switched off) and a TDP budget of 140W shouldn't be backsliding on top end Turbo speeds. That appears to be more lack of AMD competition and market segmentation from Intel's mainstream line up than necessary by the micro architecture. It isn't a Haswell design flaw... just marketing.
It does make Apple's justification of a new Mac Pro harder than simply graphing geekbench drag racing scores. But if that is the limit of their marketing abilities they are more problems than just limited Haswell-EP single threaded scores.
We have some integer-heavy apps that are seeing 20% improvements on our E5-v3 systems vs E5-v2. AVX2 brings "loop unrolling" into the hardware. It's huge.
It isn't exactly the same but these two are similar in relationship to each other AVX (new) vs SSE (old) x86-64 vs. x86-32 . The "loop unrolling is going to make it easier for compilers to write good-to-very-good numerical kernels for AVX2 (whereas AVX is far more limited and needs custom hand coding). Old code targeted at compilers from 4-5 years ago... sure no speed increases. Folks using modern tools will see something.
When things are made easier for software developers ( compiler and/or OS library ) does most of the work the uptake over time is usually pretty good.
[ there is a long software development track record to support that . ]
----------
Yes.
All I will say right now is that, what we see from Mantle right now is only being Low-Level API for gaming.
Future is long...
Sometimes the past is repeated in the future. It isn't like there wasn't a round of vendor specific low level APIs before. There some cyclic stuff going on here. Low-level APIs come in and kick the high level stuff in the shins on performance. There is an adjustment and the pendulum swings again. windows 10 and the DirectX updates seems to be answering relatively rapidly on the Windows side. OS X .... never was a sprinter when it comes to graphics. iOS .... actually would like more vendor lock-in so probably not in a hurry either.
AMD has much bigger problems that Mantle really doesn't resolve all that well. As an interim way point, it will probably be helpful. As a long term strategic solution..... it probably is not. The other low level folks have way more money to throw at the issue long term. Each of the console vendors are going to want their own. Windows is adjusting. AMD will have to track all of those closely if it wants to stay alive.