Quote from: A on December 28, 2023, 18:38:21More pricey than 2-3-4-8 RTX cards? How many will you require to get 160Gb VRAM?
160Gb VRAM = 20GB VRAM = 1 RTX 4090. Ok, probably your typo.
160GB VRAM. Hm, you cannot even get a MBP with that much storage, so what are you asking?
Suppose a MBP with 96GB storage, of which - I guess optimistically in your favour - up to 90GB can be used for GPU access. Such a MBP costs, e.g., €8500+. This means we are in a PC with 4 * RTX 4090 price territory and VRAM size.
QuoteYou are the one hiding facts actually, only when I've pushed you agreed MBP can run more models than RTX.
1) For the specific AI kinds of LLMs and generative images, you have - if you have informed us correctly - educated us that there are model size that can run on a MBP with, say, 96GB but cannot run on a single consumer RTX with up to 24GB.
2) If you allow us to use a PC with a few consumer GPUs, then both MBP and RTX have enough storage of a usable kind (VRAM of the RTXs).
3) In particular for other AI kinds, it depends on the kind on which hardware it can run. CUDA-dependent AI cannot run on MBPs. Metal-dependent AI cannot run on RTX. (Same for AMD.)
QuoteYour initial post doesn't mention it.
Thanks for reminding us that meanwhile I could learn something (presuming you have informed us correctly).
QuoteTry to enjoy life.
Uh, yes. For that purpose, while I awaited GPUs during the Corona / mining years, I considered every possibility including using Apple M to run Go AI so as to enjoy my life. At that time, my level of information was the usual benchmarks on NBC, in Youtube videos etc. with the rough suggestion of equality of speed or at least of efficiency. With such insufficient information, I might have made the mistake of buying an Apple device for Go AI and would have found that it would not have enabled me to enjoy my life because Apple M is orders of magnitude too slow for Go AI or many other AIs better run with CUDA / TensorRT. Now that I know better, I inform people to make similarly wise hardware and library choices for such softwares so that they can also enjoy their lives well.
QuoteI was thinking you are interested in reality,
I am.
Quotepersonal vendetta.
No more than Apple's performance against its critics. Tell us: why does Apple force users to click "Not Now" thousands of times instead of letting them click "Never" once? Why does Apple disable wireless OS updates just because one does not accept iCloud? Such is malicious intention to harm those of its critics that do not dispose of their Apple devices.
Quoteyou just posted a public video
And when I posted the same kind of information based on my own investigation, you said something similar. You always downplay findings revealing the truth about slow Apple hardware for certain classes of softwares.
QuoteOr just don't get into discussions if you don't know the basics.
The software basis about bits are:
- a PICe bus has a bit width for accessing VRAM
- a dGPU has a bit width
- an iGPU has a bit width
- a mainboard has a bit width for accessing RAM
- unified memory has a bit width of its mainboard
- the GPU of an Apple M SoC has a bit width
- the NPU of an Apple M SoC has a bit width
- the NPU of a new Intel SoC has a bit width
- some kinds of objects can be composed in units of bits and have a particular bit size, which, for some tasks, can be relevant for execution speed
You have left it ambiguous which of these bit values you mean. Therefore, it is not my mistake not to know which you mean.
If you should mean a value of objects of AI models of the image / LLM kinds, I do not know details. In particular, I do not know which bit size requires how much storage to run a related model. I do know such things reasonably well for Go AIs because I use them and observe RAM and VRAM usage - but I do not accuse you now to know such on your own. Everybody has limits on time for education. Then, mutual explanations can help. Go AI fills 64GB RAM in at least ca. 2.5h on an RTX 4070 while ca. 0.8GB VRAM is the level at every moment.