Quote from: Mate on March 14, 2021, 15:29:10
@Richardd Castello You are wrong - M1 uses more transistors than Intel/AMD core.
Quote from: riklaunim on March 14, 2021, 19:05:17Unless something has changed, no current eGPU will work on M1. I imagine the problem is simply that Apple isn't interested in releasing drivers. Potentially to keep the market captive.
AFAIK eGPU on Apple is limited to AMD cards only. Newer Nvidia cards are not supported.
Quote from: Mate on March 14, 2021, 15:29:10Well, real processor designs are converging. RISC folk found out they could never compete without more complex instructions. And CISC folk found the sets growing too complex, necessitating a sort of hardware emulation. You can also see wide designs getting faster and fast designs getting wider.
Additionally ARM64 is almost as complex as x86-64. Reduced instruction set but still executes ~1000 different instructions. Also x86 CPUs are not RISC inside. Every CPU(M1 too) is translating instruction sets to micro-ops and then execute.
Quote from: [email protected] on March 14, 2021, 17:54:16AFAIK eGPU on Apple is limited to AMD cards only. Newer Nvidia cards are not supported.
This is all well and good, but you can not use a EGPU box with a 3080 card
Quote from: Richardd Costello on March 13, 2021, 22:54:08
Layman's terms as best I can.
RISC = simple instructions, executed quickly, to do something complicated you need a group of instructions. Simplicity means less transistors, means less power.
CISC = complex instructions, executed slowly, you can do complicated things with one instruction. Complexity means more transistors, means more power (consumption).
Most of the time, CPUs execute simple instructions so RISC is the way to go for performance. BUT years ago, the CPU of choice for the IBM PC was chosen by the accounting department rather than the engineering department. They went for a chip that they could get a good discount on because of existing deals and chose the Intel 8086.
This choice gave Intel lots of money to develop their chip family, making it more and more complex and crucially utilising faster and faster clock speeds. Other companies couldn't keep up, hence Apple dumping the PowerPC RISC design and jumping on the Intel bandwagon because even IBM couldn't keep up with Intels clock speeds.
The simplicity of the RISC design means you need a relatively small number of transistors to achieve 1 instruction per clock cycle. 80x86 needed many (variable number of) cycles per instruction. Intel developed the 80x86 line up to the Pentium 4 when they hit a problem. They simply couldn't get them to run faster than 3.4GHz - they were melting. So at that point they switched the architecture to the Core line which is more RISC like but internally decodes 80x86 to maintain compatibility. Still complex.
Ultimately Intel has managed to achieve 1 instruction per clock cycle (like RISC) but to achieve this they have had to have ever more complicated pipelining of instructions, meaning even more transistors, more power, more heat.
Think of a processor like a production line in a car factory, it might take 3 days to make a car, but if you have many stations all working on cars moving along a production line, after 3 days of production the cars actually come off the production line one per minute. CPUs are like this, with many instructions being worked on at the same time.
What has happened recently with the M1, is Apple has got close to Intel's clock speeds. The M1 is rumoured to be running at 3.2GHz, which is very fast for the ARM architecture. This speed means that the ARM can process its instructions at the same rate as the Intel chips but crucially using a LOT less transistors.
This has a two fold advantage for Apple, firstly their chip uses a lot less power (less transistors), generating a loss less heat. Secondly it makes their processor core very small compared with an Intel core.
The physically small core and low power means they can add more cores but crucially also means they can add more supplemental hardware to the die. They can have memory controllers and the RAM itself on chip - and can subsequently run the RAM a lot faster than your typical DDR RAM units. The whole SOC (system on a chip) thing means everything is faster.
Apple have also been able to produce their chips using smaller manufacturing methods than Intel can reliably achieve. Further adding benefits.
Intel having always been running into a dead end with CISC, sooner or later RISC was going to dominate once manufacturing and high clock speeds became mass market.
Intel have now reached that dead end. It started with mobile where battery life was the main issue. Now its the desktop where clock speed and memory performance is key. Soon, it will be the server market where 128 or 256 core ARM CPUs will trundle along feeding us our data using less electricity and crucially requiring less cooling. Thats why nVidia want to buy ARM.
QuoteMiani came to this estimate by comparing the differences in results between the A12 and A12Z (1.68x faster) and the A14 and M1 (1.87x faster) and then picking an average for the M1X.
Quote from: _MT_ on March 13, 2021, 14:48:02Wow as a layman this was really helpful, Thanks! :)Quote from: george on March 13, 2021, 11:21:10I don't know if this is layman enough, but one thing you need to understand is that frequency impacts efficiency. It stems from the relationship between voltage and frequency. The higher the operating frequency, the higher the voltage required to ensure stable operation. Power rises with the square of voltage. That's why this relationship is a very important characteristic of a processor from efficiency standpoint. Firestorm cores operate at a much lower frequency than their x86 counterparts under heavy load. They operate much closer to optimum efficiency point. As a rule of thumb, 3 GHz is roughly considered a knee point for efficiency in modern x86 processors. Meaning that efficiency quickly deteriorates beyond this point. That's why base frequencies of mobile processors are where they are. Running around 5 GHz is definitely not good for efficiency.
can someone explain in layman terms how is this possible?
There are two basic approaches to increasing performance of a processor. Working faster (frequency) and doing more simultaneously (so called width). Historically, these two approaches were considered mutually exclusive. Modern processors combine them. But they still pull in different directions. M1 is a wider design. Widest on the market. Actually, modern x86 designs are internally not that dissimilar. The challenge is that you've got to feed the core with instructions. In the case of an x86 processor, it means decoding x86 instructions and turning them into internal micro instructions. But the instruction set is very complex. And there is a big complication in the form of variable length of instructions. You don't know where the next instruction starts without looking at the previous instruction. Which complicates the design of decoders. And it makes going as wide more challenging. This is where SMT comes in. By processing two (or more) threads simultaneously, you provide more instructions for the core to chew on, working its magic (like out-of-order execution, optimizing utilization of resources).
Also, the relationship between frequency and performance is not exactly straightforward. A processor is faster than memory. A big factor in performance is how much time you spend waiting for data. The higher the frequency, the more cycles get wasted by waiting. And again, SMT can come to the rescue, masking latency.
Apple's designs are extreme in more ways than one. It's hard to say what's going on if you don't have access to internal information. But those are the two main factors. They are so efficient mainly because they run at such a low frequency compared to x86 processors. And they are competitive because they're very wide yet their frequency isn't too low. There are many little things going on. Consider how Apple focuses on low latency (many benchmarks are latency sensitive). And then there is the fact that Apple uses the most advanced manufacturing process on the market, better than what AMD and Intel are using. Which means higher efficiency and being able to cram more transistors into the same space.
Desktop processors consume 100+ W primarily because they can. Efficiency is more of a consequence of increasing performance rather than a target. Their power budget for a high performance personal computer can be over 500 W. Mobile x86 processors are derivatives of desktop designs (and power saving technologies do trickle back). Here we see the opposite. Designs from highly power-constrained world of mobile phones upscaled into the world of personal computers. A single core in my desktop computer has a higher power budget than an entire iPad. And Apple succeeding in this effort that was considered very difficult if not borderline impossible. And it doesn't look like the design is running out of breath.
Quote from: Lucas on March 13, 2021, 15:49:25
The only problem with the new MAC CPUs i that they use RISC instructions which requires a lot of RAM and if the user buys a machine with 8 GB then a lot of data is written to the internal SSD which in turn reduces the lifespan. This is not theorizing, it's being observed on the new laptops as we speak where SSDs wil potentially be used up in 2-3 years.
It would not be a big problem if Apple did not use soldered SSDs so once one dies you need a new laptop.