I'm retired from being a Defense Contractor and been to many places and seen them with mine own eyes walking back to my section for a new Unix crypto device worked! So don't assume you know everything!That's comedy gold!
I'm retired from being a Defense Contractor and been to many places and seen them with mine own eyes walking back to my section for a new Unix crypto device worked! So don't assume you know everything!That's comedy gold!
But I dont think they can replace Amazon. Amazon's server scale is beyond Apple's reach.If Apple is doing this, it is so they can run their services on their own systems, and do it more efficiently and at less cost than using outside service providers.
Could an Apple's server chip based on M1 be better than AWS's Graviton?If Apple is doing this, it is so they can run their services on their own systems, and do it more efficiently and at less cost than using outside service providers.
Could an Apple's server chip based on M1 be better than AWS's Graviton?
Maybe once they make Mac Pro, they might be able to ditch x86 server for their own chips to replace.
What do you mean by that? Does it mean that another company could have designed a better SOC than AWS?Even Neoverse N1 could be done better than what Amazon did.
How could server chips for Apple internal use be? Could Apple be designing chips for running Apache Cassandra clusters or the App Store?Anything Apple makes for the server will almost certainly be Apple-user focused first. Apple isn't going to compete with AWS to host your Postgres/web services. They're going to offer Apple Silicon Cloud to Apple users to accelerate their workflows first.
Coming in late to this thread, but wanted to contribute.MLID is reliable? Don't think so.
Is Youtuber "Moore's Law Is Dead" legit?
Randomly found this guys channel a few days ago. He claims to know a bunch of industry insiders and supposedly has a ton of info that he "can't talk about". But he only has 70K subs so I'm a little suspicious. Does anyone know if this guy is for real?linustechtips.com
He's just a clickbait youtube guy who spits out lots of guesses (often contradicting himself), hypes the hits, and pretends the misses never happened.
I largely agree, Cloud computing assumes flexibility. I don’t know how macOS is doing it’s virtualization, but I will tell you that as far as a workstation goes, it’s the best solution I’ve used having used workstation laptops and Macs Out of the many years I’ve experimented with it.MLID uses three degrees of confidence (very high confidence, high confidence, and mostly confident) in his slides. So, I can expect that he misses more when he has less confidence in his info.
Anyway, for the sake of the conversation, let's pretend that MLID is correct and Apple is designing its server chips.
Does it make sense to have CPU, GPU and RAM in the same SOC for a server chip? From my limited knowledge, cloud computing requires a lot of flexibility. The requirements for a rendering farm are different to an Apache Cassandra cluster.
Form Factor (As of 9 Nov 2021) | AMD Zen 4 Epyc 128 core cpu Rival | AMD Zen 4 Epyc 128 core cpu Rival | Mac Pro iMac Pro | Mac mini Pro Mac Pro iMac Pro | MBP 14" MBP 16" Mac mini Pro iMac 24" iMac Pro | 300mm² Silicon Wafer |
Apple silicon chip | M1 Max Jade-16C | M1 Max Jade-9C | M1 Max Jade-4C | M1 Max Jade-2C | M1 Max | M1 Max Jade-49C |
Launch | >Speculation< | >Speculation< | Q2 or Q4 2022 | Q2 or Q4 2022 | Q4 2021 | >Speculation< |
# of dies | 16 | 9 | 4 | 2 | 1 | 49 |
CPU | 160 | 90 | 40 | 20 | 10 | 490 |
performance cores | 128 | 72 | 32 | 16 | 8 | 392 |
efficiency cores | 32 | 18 | 8 | 4 | 2 | 98 |
GPU core | 512 | 288 | 128 | 64 | 32 | 1568 |
Neural Engine core | 256 | 144 | 64 | 32 | 16 | 784 |
memory bandwidth | 6.4TB/s | 3.6TB/s | 1.6TB/s | 800GB/s | 400GB/s | 19.6TB/s |
Max Memory | 1024GB | 576GB | 256GB | 128GB | 64GB | 3,136GB |
Hardware-accelerated H.264, HEVC, ProRes, and ProRes RAW | 16 | 9 | 4 | 2 | 1 | 49 |
Video decode engines | 16 | 9 | 4 | 2 | 1 | 49 |
Video encode engines | 32 | 18 | 8 | 4 | 2 | 98 |
ProRes encode and decode engines | 32 | 18 | 8 | 4 | 2 | 98 |
Peak Quoted Transistor Densities using TMSC's 5nm (2020) at the same die size 171.3 million transistors per mm² | 912 Billion | 513 Billion | 228 Billion | 114 Billion | 57 Billion | 5.19 Trillion |
Estimated Die Size | 17.004cm² | 12.753cm² | 8.502cm² | 6.3765cm² | 4.251cm² | 30cm² |
Peak Quoted Transistor Densities using IBM's 2nm (2025) at the same die size 333.33 million transistors per mm² | 1.774 Trillion | 998.24 Billion | 443.66 Billion | 221.83 Billion | 110.92 Billion | 10.1 Trillion |
Estimated AMD Ryzen 9 5950X Performance | 16x | 9x | 4x | 2x | 1x | 49x |
Estimated RTX 3080 Performance | 16x | 9x | 4x | 2x | 1x | 49x |
Their Mac Pro with a custom chip is still going to be substantively skewed to providing best single user performance as a single user workstation.
The workloads for a high end server in a cloud environment where trying to service workloads of multiple clients concurrently is substantively different.
TLB+CLR thrash. Graviton2 gets to about 100-108 ns latency at over 16MB Likewise full random is fine until past 16MB. ( the log scale graph)
The M1 Max with next generation memory subsystem, next generation process node ( N5 instead of N7) and twice the memory channels is in the same zone for full random above 16MB.
Neoverse N2 will be better. Even Neoverse N1 could be done better than what Amazon did.
Ampere 80
The Ampere Altra Review: 2x 80 Cores Arm Server Performance Monster
www.anandtech.com
Latencies here under 100ns even out to over 525MB. Graviton2 isn't what Apple's has to beat to be competitive.
Amazon probably has the least expensive server CPU costs out there; not the best.
Single thread drag racing on relatively small test depths above ... yes the M1 will be substantively better. It is tuned that way. That isn't a mainstream cloud service heavy workload though.
What if Apple Silicon uses HBM2 or HBM3?Using TSMC's chiplet tech it is possible to mask-stitch 9-16 M1 Max dies together to reach 90-160 CPU cores together.
The 17.004cm² die could fit into the 19.7cm² Mac mini enclosure although it would need at a 1-1.4kW PSU to power
Form Factor (As of 9 Nov 2021) AMD Zen 4 Epyc 128 core cpu Rival AMD Zen 4 Epyc 128 core cpu Rival Mac Pro
iMac Pro Mac mini Pro
Mac Pro
iMac Pro MBP 14"
MBP 16"
Mac mini Pro
iMac 24"
iMac Pro 300mm² Silicon Wafer Apple silicon chip M1 Max Jade-16C M1 Max Jade-9C M1 Max Jade-4C M1 Max Jade-2C M1 Max M1 Max Jade-49C Launch >Speculation< >Speculation< Q2 or Q4 2022 Q2 or Q4 2022 Q4 2021 >Speculation< # of dies 16 9 4 2 1 49 CPU 160 90 40 20 10 490 performance cores 128 72 32 16 8 392 efficiency cores 32 18 8 4 2 98 GPU core 512 288 128 64 32 1568 Neural Engine core 256 144 64 32 16 784 memory bandwidth 6.4TB/s 3.6TB/s 1.6TB/s 800GB/s 400GB/s 19.6TB/s Max Memory 1024GB 576GB 256GB 128GB 64GB 3,136GB Hardware-accelerated H.264, HEVC, ProRes, and ProRes RAW 16 9 4 2 1 49 Video decode engines 16 9 4 2 1 49 Video encode engines 32 18 8 4 2 98 ProRes encode and decode engines 32 18 8 4 2 98 Peak Quoted Transistor Densities using TMSC's 5nm (2020) at the same die size
171.3 million transistors per mm² 912 Billion 513 Billion 228 Billion 114 Billion 57 Billion 5.19 Trillion Estimated Die Size 17.004cm² 12.753cm² 8.502cm² 6.3765cm² 4.251cm² 30cm² Peak Quoted Transistor Densities using IBM's 2nm (2025) at the same die size
333.33 million transistors per mm² 1.774 Trillion 998.24 Billion 443.66 Billion 221.83 Billion 110.92 Billion 10.1 Trillion Estimated AMD Ryzen 9 5950X Performance 16x 9x 4x 2x 1x 49x Estimated RTX 3080 Performance 16x 9x 4x 2x 1x 49x
Then pay for the additional R&D.What if Apple Silicon uses HBM2 or HBM3?
The 17.004cm² die could fit into the 19.7cm² Mac mini enclosure although it would need at a 1-1.4kW PSU to power
Make that Mac mini taller, say 9.8", and you would have room for that PSU & a heat sink filling the remaining interior volume; something about the size of the G4 Cube...! ;^p
$16,000?Make that Mac mini taller, say 9.8", and you would have room for that PSU & a heat sink filling the remaining interior volume; something about the size of the G4 Cube...! ;^p
Using TSMC's chiplet tech it is possible to mask-stitch 9-16 M1 Max dies together to reach 90-160 CPU cores together.
The 17.004cm² die could fit into the 19.7cm² Mac mini enclosure although it would need at a 1-1.4kW PSU to power
Make that Mac mini taller, say 9.8", and you would have room for that PSU & a heat sink filling the remaining interior volume; something about the size of the G4 Cube...! ;^p
$16,000?
AWS's Graviton2 uses Neoverse N1 CPU microarchitecture.
What do you mean by that? Does it mean that another company could have designed a better SOC than AWS?
@deconstruct60 Thank you!
AMD has two types of GPU: CDNA for scientific computing and RDNA for gaming/rendering.
Which type of GPU should Apple's server chips have: CDNA-like GPU or RDNA-like GPU?
Well, I'm not retired. An M1 SOC Mini is not a competitive supercomputer node. Sorry. Too few cores, inadequate memory, limited to ethernet.... I don't think HPE is worried. Vastly different market. I thought you were joking.I'm retired from being a Defense Contractor and been to many places and seen them with mine own eyes walking back to my section for a new Unix crypto device worked! So don't assume you know everything!
Let us double it to $32,000, then. Default storage would be 8TB?For 16 maxed-out M1 Max SoCs that would be a bargain...!
16-way M1 Max MCM
160-core CPU (128P/32E)
512-core GPU
256-core Neural Engine
1TB LPDDR5 RAM
6.4TB/s memory bandwidth