That happens to not be how you design a processor.
But when it comes to a billion dollar marketing campaign,
you expect hyperbole at many levels. If an executive
tries to use the reality distortion field, stuff like
that happens. Doing lines of coke off coffee tables,
that's what executives do. Engineers live in much
more austere environments. A cup of coffee, and smelly
carpeting with too much formaldehyde in it. That's
where engineering is done.
You need to find an article here (subscription based).
As there isn't as much free-lance interest in processor arch
as there once was.
https://www.linleygroup.com/mpr/about_report.php
A good part of hardware design now, comes from software people.
They know the compiler. The compiler is a long lead-time item
of some importance (it takes around ten years to make
a good one from scratch). Preparing a processor architecture requires
spotting patterns where both the compiler and the hardware
could be modified. You're staring at workload traces, but with
an eye to your favorite pet theory. You're not staring at
lines of coke on a glass coffee table.
As for "where do ideas come from", see Intel. Their innovation
came from hiring a team in Israel. Folsom didn't save them,
Israel did. In the case of Apple, I believe they may have
made a bulk purchase of some CPU people, but I don't sit
around tracking bumpf like that. Knowing which company that
was, would tell you why the processor looks the way it does.
I'm kinda curious how so many functional units can be put in
parallel, because it isn't "normal" for exceptionally high
retirement rates on the same clock tick. Four functional units
would be pushing it, as maybe you could arrange three functional
units to retire on the same tick, but squeezing a fourth into
the model is tough (that means, using a fourth functional
unit, that actually gets used occasionally). This is a measure
of "how much parallelism exists in normal code". And that
situation has been stagnant for some time. The AMD Zen3 made
a small move in that direction.
And this is where a journal that specializes in that sort of
analysis, would shred whatever PR campaign was ongoing.
As for Anandtech, one of the web pages mentions "we used what we had",
which means they weren't able to apply their normal benchmark suite.
You know, Anand left Anandtech years ago, to work at Apple.
Anandtech was bought by a magazine company, and that magazine
company also bought Tomshardware (and probably a few others).
Another good supply of analysis used to come from the Russians (
ixbt.com).
They knew some things about the Pentium P4 that nobody else knew
(how hyperthreading worked and how the first hyperthreading
had a bug in its recirculator). Some of their analysis was
from first principles. They could write code and demonstrate
that an idea they had, was real. But they're not around any more,
so scratch another source of analysis potential.
I'm sorry, but if you're expecting any sort of reasoned
(non-NDA) analysis these days, you pretty well have to pay for
it. The kiddies can only regurgitate what they're given.
But there are some people with the chops to do the analysis.
You couldn't even visit comp.arch any more and expect anything
cogent. I think that was spammed out of existence at some point
and the people left.
*******
One thing I've learned over the years, is hardware is useless
without good software. And then the question is, is the software
for a product, what an individual user wants or not. That's
where my interest in the M1 trails right off... The ability
of poorly written software, to squander tiny improvements in
hardware, is legendary. This is why today, you can hardly have
a web browser that isn't railed and non-responding. An M1
won't fix that. Most of the time, my processor components
sit there unused - but that's software for you.
This is also the reason I'm not interested in owning anything
with an NVMe in it. Fine fine hardware. Rendered useless by
software.
Just yesterday, I did an experiment with a ramdisk, where
the file copy rate was 1.8MB/sec. Now think about that for
a moment, just how pathetic that is. A "device" with a 5GB/sec
sequential benchmark, that in a real life test situation,
can only manage 1.8MB/sec performance. (That's the NTFS fuse
file system on a recent Linux distro.) This is why "dreams of M1"
would be tempered with reality, and that reality is
the bloated software we used today. That software is... everywhere.
This is the Year of the Container, as containers and virtualization
threaten to ruin multiple ecosystems at the same time. Try loading
a Snap and see how long it takes before your application is
ready (20 seconds). Does Apple use containers ? Does Apple use
virtualization ? Of course they do. I don't even need to check,
because there's plenty of copying and "me too" in the software
industry, even if critical analysis would tell them the idea
was wrong.
This is why I no longer get excited about "whizzy hardware".
A software guy will always find a way to ruin it.
Paul