I noticed a similar trends in selling on X. Put a claim, peg on some product A with good sales - Cursor, Claude, Gemini, etc. Then say, the best way to use A is with our best product, guide, being MCP or something else.
For some of these I see something like 15k followers on X, but then no LinkedIn page for example. Website is always a company you cannot contact and they do everything.
Yes. The article is click bait. With such a title I would have expected majority of the area to be dummy, but it is just structurally more silicon, exactly like a picture may be majority of its mass wood.
Your statement is incorrect. The analysis was made by a professional firm - dummy silicon shims are used because the dies are thinned, as per AMD's own disclosures. Those silicon shims are bonded to the compute and SRAM dies.
I ended up disabling copilot. The reason is that the completions do not always integrate with the rest of the code, in particular with non-matching brackets. Often it just repeats some other part of the code. I had much fewer cases of this with Cody. But, arguably, the difference is not huge. But then add on top of this choice of models.
I noticed I had a lot fewer of these problems these last few weeks. I suspect the Copilot team has put a lot more effort into quality-of-life recently.
For instance, I'd often get a problem where I'd type "foo(", and VsCode would auto-close the parenthesis, so my cursor would be in "foo(|)", but Copilot wouldn't be aware of the auto-close, so it would suggest "bar)" as a completion, leading to "foo(bar))" if I accepted it. But I haven't had this problem in recent versions. Other similar papercuts I'd noticed have been fixed.
I haven't used Cody, though, so I don't know how they compare.
It seems recent years give us a lot of licenses (for core infra software) and now for LLMs. They all say in very legalese basically: these top 5-10 tech companies will not compete fairly with us, thus they are banned from using the software. The rest are welcome to use everything.
I wonder if US monopoly regulation actually starts to work well, which I see some signs of happening, will all this license revert back to fully open source?
Exactly. The whole thing reads like some propaganda. It pits interesting topics ahead then to move on and push some agenda that sounds super political to me.
Yes, some languages are underrepresented and there are some thresholds. But exactly, it is well known that putting the threshold just slightly above or below will probably not materially affect the model.
The product they often presented as started in 20% time is Google news. I don't know the actual details, just this is what I remember from my time at Google (2006-2012).
I find the article quite informative. Yes, M2 and the other chips are completely different products with different goals. If one wants to say that something completely trumps the other, it will be wrong.
But here is what is visible:
The M2 core is probably in the same ballpark as Zen 4 core, likely a tiny bit below. That may become very tiny if Zen 4 core runs at lower frequency to equalize the power. This doesn't account for the AVX512 of Zen4.
24 M2 cores manage to beat 16 Zen 4 cores also at lower power, but these are different products. Zen 4 does scale to far more cores, 96 in an EPYC chip. AMD and Intel have far more investments in interconnects and multi-die chips to do these things.
The M2 GPU is in the same league as a 300$ mid-range nVidia card. It is not competitive at all - Apple produces the largest chip it can manufacture to go against a high margin smaller chip that nVidia orders.
Again all of this doesn't mean each product is not good on its own.
Apple's GPU performance is what makes me sceptical about their gaming related advertising. Sure, you can do 1080p gaming with the highest SKU, but you're paying through the nose if you bought an M2 to play games.
It seems strange to me for Apple to advertise something they haven't exactly mastered yet on stage.
Maybe they have some kind of optimization up their sleeves that will roll out later? I can imagine Apple coming out with their own answer to DLSS and FSX2 based on their machine learning hardware, for example. On the other hand, I would've expected them to demonstrate that in the first place when they shoed off their game port toolkit.
With crossover and Apple's latest release of gameportingtoolkit I'm able to maintain over 120FPS on ultra settings at native resolution on Diablo 4 with my M2 Max MBP. It was fair to be skeptical before that release this week, but there's plenty of evidence out there now that Apple silicon can handle gaming just fine. Other users are reporting 50-60 FPS with ultra settings on their 6k Studio displays.
I thought the whole idea of M2 was “exceptional product given the power consumption”.
I don’t mind that it has nothing to show for all the talk once you throw out the need to basically sip power (like a notebook computer).
Is this something inherent with ARM though? Why can’t there be ARM based desktop and server computers that need a kilowatt of power at peak? Like how much more performance can you get for each additional watt of power? (I don’t know. I’m genuinely asking.)
I was one told that memory and bus bandwidth often creates disparity between benchmark and application performances in ARM CPUs. That was years ago and supposedly don’t apply to custom designs like M2, but maybe both Intel and AMD are still advantageous in that region?
> I thought the whole idea of M2 was “exceptional product given the power consumption”.
When running native code.
Look at the performance of Microsoft's ARM Surface Pro when running emulated code.
> My frustration with this computer wasn’t a workload thing. It didn’t start out fast and gradually slow down as I opened more things and started more processes. It was peppered with glitches and freezes from start to finish.
I’d have only Slack open, and switching between channels would still take almost three seconds (yes, I timed it on my phone). Spotify, also with nothing in the background, would take 11 seconds to open, then be frozen for another four seconds before I could finally press play. When I typed in Chrome, I often saw significant lag, which led to all kinds of typos (because my words weren’t coming out until well after I’d written them). I’d try to watch YouTube videos, and the video would freeze while the audio continued. I’d use the Surface Pen to annotate a PDF, and my strokes would either be frustratingly late or not show up at all. I’d try to open Lightroom, and it would freeze multiple times and then crash.
It quickly became clear that I should try to stick to apps that were running natively on Arm.
Not familiar with DLSS at all, does it requires developers to do something in order to take avantage of it too? I had imagined it was automatic but then again I know nothing about it beyond the marketing pitch to consumers.
I am not knowledgeable enough to know how much work it is but I have played games that didn’t initially support it but eventually released an updated that added support.
There are also multiple “levels” for DLSS in games that support it, eg. Quality, performance, etc
> Apple's GPU performance is what makes me sceptical about their gaming related advertising.
The issue is that people compare games running under emulated x86 and emulated graphics APIs, when making claims about what the SOC is capable of.
There's nothing wrong with knowing how well the SOC performs when emulating games, but if you claim to be talking about what the SOC can do, then include the performance of native games as well.
Apple's x86 emulation is otherwise very impressive, and not many games are bottlenecked on the CPU, especially at high resolutions.
Bigger overhead for AAA games is likely due to emulation of DirectX or Vulkan on Metal, but that's just Apple's stubborn choice to have it that way.
In the end, none of that matters. I won't be playing Cyberpunk at 14fps, without RTX, and comforting myself that the SoC could do maybe 28fps without emulation. Lower-tier Nvidia cards perform better, even when paired with slower CPUs.
The major platforms do use the same graphics API, Vulkan. It should be preferred due to more low-level access and wider platform support (Linux, Android, Nintendo, MacOS, Windows).
On another note, problems that keep major AAA games from running on Linux (Anti-cheat solutions for example) will block many games from running ob MacOS, too.
The CPU is rarely a bottleneck for AAA games, so unless the x86 emulation is particularly terrible (Rosetta isn't) it shouldn't be the issue.
WINE on Linux is able to match the performance of games on Windows, so the DirectX translation layer shouldn't be a problem either.
So it's not unreasonable to assume that the M2 just doesn't have a GPU capable of running these games. And it's really not that surprising that an integrated GPU doesn't match the performance of a dedicated GPU.
PC game players tend to believe you can't play a game unless you bought the latest custom hardware for all of it and put all the settings on maximum.
Game developers are much more willing to run their work on lower end machines if they'll get paid for it, or at least they're more capable of tuning for it.
> So it's not unreasonable to assume that the M2 just doesn't have a GPU capable of running these games
Without including comparison data on native games? It's entirely unreasonable.
For instance, The native version of the DirectX 12 game "The Medium" was shown running side by side with the emulated version at WWDC, and the native version had double the frame rate.
> the M2 just doesn't have a GPU capable of running these games.
As long as AAA games are published on the Xbox Series S and shipping with graphics settings they will have no problem when running natively on an M2 chip.
>The M2 core is probably in the same ballpark as Zen 4 core, likely a tiny bit below.
The 7950x is running at 5.7Ghz when only a single thread is saturated. The M2 Ultra caps its cores at 3.5Ghz. A 62% higher clock speed, at a monster power profile, to barely beat it isn't evidence of a core advantage.
>24 M2 cores manage to beat 16 Zen 4 cores also at lower power
The M2 ultra has 16 real cores, with 8 additional efficiency cores that are very low performance. And of course the M2 Ultra could pretty handily trounce the 7950x because the latter has to dramatically scale back the clock speed, as the power profile of all 16 cores at 5.7Ghz would melt the chip. And of course the 7950x has hyper-threading and hardware for mini-versions of 16 more cores, so in a way it has more cores than the Apple chip.
>This doesn't account for the AVX512 of Zen4.
AVX512 is used by a tiny, minuscule fraction of computers ever in their history of existence. It is the most absolute non-factor going.
I mean...in an ideal world Apple would get the GPU off the core. It limits their core and power profile, and takes up a huge amount of die space. They could then individually mega-size the GPU and the CPU. They could investigate mega interconnects like nvidia's latest instead of trying to jam everything together.
Was Apple correct to call it the most powerful chip? Certainly not. And there is a huge price penalty. But they're hugely, ridiculously powerful machines that will never leave the user wanting.
It is true that nobody competes in the low power high efficiency workstation market or maybe such a market does not exist yet and Apple is creating it.
But also as users, some were expecting the M series are so good that they are going to take many markets by storm. And it seems it is not happening.
$300 midrange Nvidia card? Did you get stuck in 2010?
That's way below entry-level at this point. You're likely comparing it with a 1666 cards or something, which is based on a chip from 2012.
I wish Apple silicone was actually competitive on performance. Nvidia needs competition or they'll likely double prices again with the next generation.
> The M2 GPU is in the same league as a 300$ mid-range nVidia card
It still has the advantage of a much larger memory pool.
I did a quick comparison exercise - I priced two workstations with similar configurations, one from Dell, the other from Apple. While there are x86 (and ARM) machines that'll blow the biggest M2 out of the water, the prices, as far as Apple can go, aren't much different.
If you buy anything labeled as "workstation", you're paying twice the price already.
The article describes the M2 being blown out of the water by a 4080 and a 13900KS. That's about $2000 + RAM, motherboard, and power supply. Plus you can use the built in GPU in your CPU for acceleration things like transcodes.
You can get a pre-built gaming PC with a 4090 for about $4000, that'll crush the M2 in compute if you use any kind of GPU acceleration.
Of course the M2 has some other advantages (the unified memory and macOS) and some other disadvantages (you're stuck with the amount of RAM you pick at checkout, macOS, you have to sacrifice system RAM for GPU RAM) so it all depends on your use case.
I think the M2 still reigns supreme for mobile devices, though AMD is getting closer and closer with their mobile chips, but if you've got a machine hooked into the wall you'll have to pay some pretty excessive electricity rates for the M2 to become competitive.
> If you buy anything labeled as "workstation", you're paying twice the price already.
The price of workstation-class machines also includes the cost of higher build-quality and stability, things like same-day support and service - at least the option for a long-term (5-6 year) warranty, and FRUs - you don't get that with consumer-grade computers - and those things matter when a machine is something you depend on professionally.
What the poster means is that a "workstation" is designed with quickly swappable components, often not even needing to use any tools. Businesses may benefit from this.
While it doesn't necessarily mean the swappable components are standardized or easy to procure, they usually are. That's a separate item that "workstation" machines typically offer: longer availability of replacement parts.
I agree with your take. My plugged into the wall machine is a 128GB 13900k 4090 system. My mobile machine is an Apple Silicon Macbook Pro. There are some tasks that are still better on the unified memory of the Macbook, but only a handful. There are many tasks that are more pleasant on the Macbook because of the absurd power efficiency (DAW, Final Cut Pro).
Both machines have a quality that I appreciate: they are never, ever slow.
You’re forgetting the benefit of everything just working and never having to thinking about effing with drivers ever. To me, it’s priceless. Anything truly performance bound (CPU or GPU) is going to be done on HPC systems, not on a fake Windows “workstation”.
> If you buy anything labeled as "workstation", you're paying twice the price already.
We are not comparing MacPros to low-end desktops.
> You can get a pre-built gaming PC with a 4090 for about $4000, that'll crush the M2 in compute if you use any kind of GPU acceleration.
Yes, but the gaming PC will not as well built as the workstation-grade machine. And pretty much any GPU you can install on a gaming PC you can install on a MacPro - it's just that it won't be there out of the (Apple branded) box.
> you're stuck with the amount of RAM you pick at checkout
Sadly, this has been Apple for some time now - you buy the machine as it will be used for its whole intended lifetime. With the MacPro you can at least add internal storage and one or more GPU cards.
AFAIK the 2023 Mac Pro doesn't support PCIe GPUs for the same reason AS Macs don't support eGPUs. It has PCIe slots you can use for other things like capture cards or whatever but not GPUs.
RAM was something you could upgrade with the 2019 Mac Pro and something you could get a lot of. 1.5TB worth. The new Mac Pro caps out at 192GB which is barely better than consumer AMD/Intel systems at the moment.
I agree some MacPro users will be forced to move to workstation or server-grade PCs, but I am sure Apple knows that and they considered having integrated memory inconsequential for the majority of their users.
Also, remember, terabytes of RAM cost A LOT of money. The Dell I priced for comparison can go way higher than 192GB, but it’ll also cost you a lot more than 7K.
> It still has the advantage of a much larger memory pool.
I wonder if given roughly equal power to the GPUs in current gen consoles (PS5/XBSX), it'd yield some advantage in porting console games since those consoles also have a large shared pool of memory (16GB), and neither AMD nor Nvidia want to give up using VRAM as an upsell.
With the M2 Ultra prices, it'd be cheaper to buy a 4090 than to go the Apple route. With the M2 pro you'll probably still be better off with a 4080 unless you really need more than 16GB of VRAM.
I don't know the M2's efficiency for things like machine learning, but the M1's machine learning performance seemed to have been beaten 4-5x by the 3060Ti so I'm pretty sure "more VRAM" is all it's got going for it in ML tasks.
Well yeah, the market here would be people who already have a reasonably powerful Mac and would rather have that fill their gaming needs instead of having to build or buy a separate dedicated device for that.
But what I was really getting at is the trouble that game studios have been encountering lately when porting PS5 and Xbox titles to Windows, which is that these games are so reliant on those consoles' 16GB shared memory pool that they perform terribly on PCs. The impact is double, because not only are most GPUs in usage right now anemic when it comes to VRAM (even my last-gen high end 3080 Ti comes up short at only 12GB), traditional PCs also have to copy data between RAM and VRAM. Significant re-architecting for the Windows port is required to work around this.
M-series Macs are much more similar to current gen consoles with their shared memory pool, which in theory could make porting from console to Mac (at least when targeting Macs with 16GB+ of RAM) more straightforward than porting to Windows. While some work would need to be done to support Metal, the two most popular engines already do much of that legwork and the work that remains can be shared across multiple titles.
I can’t imagine using my work computer for gaming, as maintaining the software install has so many different requirements, but, then, I’m no PC gamer and would rather have a console plugged into the big TV in the living room than on my desktop monitors. It’s also much less of a hassle maintaining a console than a gaming PC.
As a side-note, my living room TV is a rather small 43 inch one (limited in size by the surrounding overflowing book shelves) but, if I were a gamer, I’d probably have gone with a 60+ inch or wall projector.
If I lived alone, I’d get an Apple Vision Pro instead of the humongous TV, as it’d be cheaper.
Cheaper in terms of money, but in terms of time? I have a hard time justifying anything that requires configuration and dicking around. I’m a grown-up and don’t have “free time”. I need things that just work. For me, that’s not Intel and Windows or Intel and Linux. It’s macOS, which is the only true workstation platform left.
My previous rig is approaching 6 years, and the only dead component is a cheap external USB drive. The rig was mining 24/7 when it wasn't used for development or gaming. You must be doing something very wrong.
Yes, it's not price segmentation, it's planned obsolescence.
The 3080 series would be fine for likely beyond the 50x0 series gpu-wise, but current games are already starting to stutter unless you downgrade textures because of its limited VRAM
The performance of the chip is matched to the memory size.
I think it’s a U shaped curve.
Beyond 80GB, today, the larger chip would maybe all of these: yield less, scale worse, take too much power, etc.
Like this matching of compute resources to RAM is partly the difference between CPUs and GPUs.
Anyway, it’s just to say that it isn’t a business decision. The extra RAM in the M2 doesn’t help the GPU much for the same tasks the H100 excels at, because it isn’t performant enough to use that RAM anywhere near the same way an H100 would, and if it were, there would have to be less RAM. The H100 doesn’t even have a graphics engine. It’s complicated.
> The performance of the chip is matched to the memory size.
That may be approximately true if you only look at a single generation of consumer graphics cards at a time. If you compare across generations or include non-gaming workloads the correlation falls apart.
What speed should we expect from the model on consumer hardware? I tried a 8 bit quantized version on 4090 and got it to generate 100 tokens for 13 second, which seems a bit slow to me.
For some of these I see something like 15k followers on X, but then no LinkedIn page for example. Website is always a company you cannot contact and they do everything.