Hacker Newsnew | past | comments | ask | show | jobs | submit | brcmthrowaway's commentslogin

This looks interesting. Anything similar for FreeCAD?

I've been using it with OpenSCAD, which has the advantage of being entirely script-generated and so more easily understandable by AI.

It's shocking!

This doesn't get talked about enough - the GPU is weak, weak, weak. And anyone who can fix them will go to a serious AI company (for 2-3x the salary).

The GPU is monstrously good. Depending on the workload, the M1 series GPU using 120W could beat an RTX 3090 using 420W.

Same with the CPU. Linux compiled faster on an M1 than on the fastest Intel i9 at the time, again using only 25% of the power budget.

And the M-series has only gotten better.

It is kind of sad Apple neglects helping developers optimize games for the M-series because iDevices and MacBooks could be the mobile gaming devices.


>the M1 series GPU using 120W could beat an RTX 3090 using 420W

You're cooked if you actually believe this


I very recently ran the numbers on these GPUs for an upcoming blog post. The token generation performance is bad, but the prefill performance is _really_ bad.

For a Qwen 3.6 35B / 3B MoE, 4-bit quant:

- parsing a 4k prompt on a M4 Macbook Air takes 17 seconds before generating a single token.

- on an M4 Max Mac Studio it's faster at 2.3 seconds

- on an RTX 5090, it's 142ms.

RTX 5090 uses more power than an M4 Max Mac Studio but it's not 16x more power.


That's just a 4k context too. At a realistic context window of 16-32k tokens, the comparison becomes downright unfair.

Somehow Apple has always been able to sell their stuff as somehow Magic. Remember the megahertz myth? Apple hertzes and apple bytes are much better than PC hertzes and bytes because they are made by virgin elves during a full moon.

> Apple hertzes and apple bytes are much better than PC hertzes and bytes because they are made by virgin elves during a full moon.

The thing that Apple has always been excellent at is efficiency - even during the Intel era, MacBooks outclassed their Windows peers. Same CPU, same RAM, same disks, so it definitely wasn't the hardware, it was the software, that allowed Apple to pull much more real-world performance out of the same clock cycles and power usage.

Windows itself, but especially third party drivers, are disastrous when it comes to code quality, and they are much much more generic (and thus inefficient) compared to Apple with its very small amount of different SKUs. Apple insisted on writing all drivers and IIRC even most of the firmware for embedded modules themselves to achieve that tight control... which was (in addition to the 2010-ish lead-free Soldergate) why they fired NVIDIA from making GPUs for Apple - NV didn't want to give Apple the specs any more to write drivers.


> NV didn't want to give Apple the specs any more to write drivers.

I think that's a valid demand, considering Nvidia's budding commitment to CUDA and other GPGPU paradigms. Apple, backing OpenCL, would have every reason to break Nvidia's code and ship half-baked drivers. They did it with AMD's GPUs later down the line, pretending like Vulkan couldn't be implemented so they could promote Metal.

Apple wouldn't have made GeForce more efficient with their own firmware, they would have installed a Sword of Damocles over Nvidia's head.


> They did it with AMD's GPUs later down the line, pretending like Vulkan couldn't be implemented so they could promote Metal.

It was even worse than that, they just stopped updating OpenGL for years before either Vulkan or Metal existed at all. Taking a Macbook and using bootcamp would instantly raise the GPU feature level by several generations just because Apple's GPU drivers were so fucking old & outdated.


On Geekbench 5, the M1 hits 483 FPS and the RTX 3090 hits 504 FPS.

There are other workloads where the M1 actually beats the 3090.

Apple does plenty of hyping but it's always cute when irrational haters like you put them down. The M1 was (well, is) a marvel and absolutely smokes a 3090 in perf per watt.


What geekbench 5 fps are you talking about? Geekbench only has OpenCL and Vulkan scores for the 3090 as far as I can tell, and the M1 Ultra is less than half the OpenCL score of the 3090. And the M1 Ultra was significantly more expensive.

Find or link these workloads you think exist, please

> The M1 was (well, is) a marvel and absolutely smokes a 3090 in perf per watt.

The GTX 1660 also smokes the 3090 in perf per watt. Being more efficient while being dramatically slower is not exactly an achievement, it's pretty typical power consumption scaling in fact. Perf per watt is only meaningful if you're also able to match the perf itself. That's what actually made the M1 CPU notable. M-series GPUs (not just the M1, but even the latest) haven't managed to match or even come close to the perf, so being more efficient is not really any different than, say, Nvidia, AMD, or Intel mobile GPU offerings. Nice for laptops, insignificant otherwise


Here you go[0]. 'Aztek Ruins offscreen'. Although I misremembered the exact FPS, the 3090 is at 506 FPS.

Also note how the M1 Ultra is pushing 2/3 of the FPS of the 3090 despite 1/3 of the power budget and the game itself being poorly optimized for the M-series architecture.

And here[1] you have it smoking an Intel i9 12900K + RTX 3900. The difference doesn't look too impressive until you realize the power envelope for that build is 700-800W.

Also, the GTX 1660 (technically an RTX 2000 series, but whatever) is about 26% less efficient than an 3090[2].

> Being more efficient while being dramatically slower

That's my whole point and what you're refusing to see. The M1 is not dramatically slower than an i9 or 3090 despite having dramatically lower power use.

The proof for this will really start to come once Qualcomm and Mediatek have gotten a handle on their PC ARM chips and Valve decides they're good enough for a Steam Deck 2 or 3. You'll get to see 2-3x the battery life along a modest performance increase.

[0]https://techjourneyman.com/img/blog/m1-ultra-vs-rtx-3090-ben...

[1]https://techjourneyman.com/img/blog/m1-ultra-vs-intel-i9-129...

[2]https://bestvaluegpu.com/comparison/geforce-rtx-3090-vs-gefo...


Apples and limes.

The context of this thread isn't consumer chips, but Apple's analog to an H/B200.


Well Apple is in the consumer computing business.

* Powered by in-house models they've tried to train and in-house M-series inference servers

TFA is literally about a B2B deal, not consumer compute.

The GPUs are bottom-barrel for compute-focused industries. It is mobile-grade hardware that arguably can't even scale to prior Mac Pro workloads.

> The GPU is monstrously good. Depending on the workload, the M1 series GPU using 120W could beat an RTX 3090 using 420W.

You're just listing the TDP max of both chips. If you limit a 3090 to 120W then it would still run laps around an M1 Max in several workloads despite being an 8nm GPU versus a 5nm one.

> It is kind of sad Apple neglects helping developers optimize games for the M-series

Apple directly advocated for ports like Death Stranding, Cyberpunk 2077 and Resident Evil internally. Advocacy and optimization are not the issue, Apple's obsession over reinventing the wheel with Metal is what puts the Steam Deck ahead.

Edit (response to matthewmacleod):

> Bold of them to reinvent something that hadn't been invented yet.

Vulkan was not the first open graphics API, as most Mac developers will happily inform you.


> Vulkan was not the first open graphics API, as most Mac developers will happily inform you.

OpenGL had become too unmanagable which is why devs moved to DirectX.

Unless you meant a different one?


> The GPUs are bottom-barrel for compute-focused industries. It is mobile-grade hardware that arguably can't even scale to prior Mac Pro workloads.

Surprised Apple didn't create a TPU-like architecture. Another misstep from John Gianneadrea.


I'm confused how anyone ever thought the NPU would be a good idea. The GPU is almost always underutilized on Mac and could do the brunt of the work for inference if it embraced GPGPU principles from the start. Creating a dedicated hardware block to alleviate a theoretical congestion issue is... bewildering. That goes for most NPUs I've seen.

Apple had the technology to scale down a GPGPU-focused architecture just like Nvidia did. They had the money to take that risk, and had the chip design chops to take a serious stab at it. On paper, they could have even extended it to iPhone-level edge silicon similar to what Nvidia did with the Jetson and Tegra SOCs.


I think they built the NPU with whatever models they needed to run on the iPhone in mind vs trying to build a general purpose chip, and then got lucky it was also useful for LLMs.

(Like “I want to do object detection for cutting people into stickers on device without blowing a hole in the battery, make me a chip for that”.)


I'm not sure even Apple thought that, given that they don't officially provide access to ANE internals under macOS (barring unsupported hacks). But if that was fixed, it could then be useful for improving the power efficiency of prefill, where the CPU/GPU hardware is quite weak (especially prior to the M5 Neural Accelerators).

Apple's obsession over reinventing the wheel with Metal

Bold of them to reinvent something that hadn't been invented yet.


The managerial class doesn't think that way.

This is not a good Saturday night for humanity

Where does TB5 come into all of this?

- Thunderbolt 3 is a superset of USB 3.1

- USB4 is built on Thunderbolt 3's protocol, implementing a subset of its mandatory features

- Thunderbolt 4 is a strict profile of USB4 (all optional features made mandatory)

- USB4 v2 introduced 80 Gbps signaling

- Thunderbolt 5 is a strict profile of USB4 v2 (again, optional features made mandatory)


I don't see why it would. Thunderbolt is not a USB standard

> Thunderbolt is not a USB standard

Concerning Thunderbolt 3: USB4 is based on the Thunderbolt 3 protocol [1].

Concerning Thunderbolt 4: "In July 2020 Intel announced Thunderbolt 4 as an implementation of USB4 40 Gbit/s with additional requirements, such as mandatory backward compatibility to Thunderbolt 3 and requirement for smaller notebooks to support being charged over Thunderbolt 4 ports.[14] Publications such as AnandTech described Thunderbolt 4 as "superset of TB3 and USB4" and "able to accept TB4, TB3, USB4, and USB 3/2/1 connections"." [2]

Concerning Thunderbolt 5: Intel considers Thunderbolt 5 as an implementation of USB4 Version 2.0. [3]

[1] https://en.wikipedia.org/w/index.php?title=USB4&oldid=134742...

[2] https://en.wikipedia.org/w/index.php?title=USB4&oldid=134742...

[3] https://en.wikipedia.org/w/index.php?title=USB4&oldid=134742...


Thunderbolt 5 and USB4v2 are the same thing now. They both support 80gbps and pcie pass through.

> Thunderbolt 5 and USB4v2 are the same thing now. They both support 80gbps and pcie pass through.

Not completely true: Thunderbolt 5 demands some capabilities that are optional for USB4v2.


From a protocol/bandwidth level, it’s essentially the same though. Thunderbolt 5 has some more guarantees for power and display, but the data rate of the two is the same.

Doesn't it run over a USB-C shaped wire? If you're trying to understand things that plug into USB-shaped ports it seems at least worth mentioning.

To be fair: You should refer to these as Type-C cables, as they carry things that are not USB protocol.

The sole exception should be made for "charge only" cables, which can, and should, be referred to as "wired for USB 2.0". These cables "shouldn't" exist, but I also don't want to buy a $30 cable just to charge my phone.


Thunderbolt 5 is basically just PCI Express, power delivery, and DisplayPort over the same cable, which for reasons passing understanding is terminated with a USB-C connector.

I think most of those cables will also support USB the protocol.


This is a good strategy. Internal competition between Gemini and GCP.

Can we not do growth hacking here?

We totally agree.

That's what I've been heads down, HUNGRY, working on, looking for investors and founding engineers pst: https://heymanniceidea.com (disclaimer: I am not associated with heymanniceidea.com)


HN is owned by a startup accelerator and venture capital firm. They do growth hacking on the front page. And you probably know that since your throwaway account is several years old.

Truly an epic company.

Theyre killing it!

Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: