> We mainly build small closed-down AI compute-chips we can control, sell them for-profit to individual consumers and then orchestrate data-processing on those chips, with setup and operational cost all paid by the consumer
I wish they did but they don't. They have been for decade so stingy on RAM for iPhone and iPad. There are at current point that only small percent of their userbase have iPhone or iPad with 8GB RAM that somehow can run any AI models even open source and be of any use. Not mentioning they don't compare to big Models.
They don't even provide option to sell iPhone with bigger RAM. iPad can have max 16GB RAM. Those mainstream macbook air also can have max 32 GB RAM.
And for the current price of cheap online AI where e.g. perplexity provides so many promo for PRO version for like less $10 per year and all ai providers give good free models with enough rate limit for many users I don't see apple hardware like particularly bought because of AI compute-chips - at least not non-pro users.
If the loose AI though and because of that won't have good AI integrations they will loose also eventually in hardware. e.g. Polish language in Siri still not supported so my mum cannot use it. OSS Whisper v3 turbo was available ages ago but apple still support only few languages. 3rd party keyboard cannot integrate so well with audio input and all sux in this case because platform limitation.
Some lot of good that's done them. The Neural Engine is dark silicon on most devices I've seen, and now we're getting another product segment with M5's matmul GPUs.
To me, it feels like Apple should have supported CUDA from the start. Sell the ARM-hungry datacenter some rackmount Macs with properly fast GPUs, and Apple can eventually bring the successful inference technology to cheaper devices. Apple's current all-or-nothing strategy has produced nothing but redundant hardware accelerators, while Nvidia's vertical integration only gets stronger.
I have a little rust script that uses the built in vision toolkit to do ocr of pdfs, it spins up the ANE to a full 1W compared to 0 as measured by the power profiler. So it is used!
IMO, It’s a very apple strategy, stuff just works and is slowly more accelerated/lower power.
Maybe. But Apple tried the server business and found that they can't compete there.
Not because of Engineering deficiencies, but because datacenters buy based on facts, not fluff.
Now their ARM silicon is top-notch, no doubt about that. But will they earn a higher margin if they put it in a datacenter instead of a consumer device which is then used to consume Apple Services? I don't think so.
> But will they earn a higher margin if they put it in a datacenter
Nvidia is a five trillion dollar business right now. The total sum of Apple's profits from services, hardware and servicing/repair costs all fail to crest Nvidia's total addressable market. We've been past the point of theorizing for almost two years now.
Apple has the means to break into that market, too. They don't need the silicon (iPhone/iPad are way overpowered, Vision Pro and Mac are low-volume), they have thousands of engineers with UNIX experience, and hundreds of billions of dollars in liquid cash waiting to be spent. If the China divestment and monopoly case happen, Apple needs a game plan that guarantees them protection from US politicians and secures an easy cash flow.
From the consumer perspective, it seems simple; stop shipping the latest silicon in the iPhone. Nobody uses it. They're not playing AAA-games or inferencing the latest AI models, and the efficiency gains haven't been noticable for a decade. You don't need TSMC 2nm to browse the App Store, or watch AppleTV. The only opportunity cost comes from selling consumers hardware they can't appreciate.
I wish they did but they don't. They have been for decade so stingy on RAM for iPhone and iPad. There are at current point that only small percent of their userbase have iPhone or iPad with 8GB RAM that somehow can run any AI models even open source and be of any use. Not mentioning they don't compare to big Models.
They don't even provide option to sell iPhone with bigger RAM. iPad can have max 16GB RAM. Those mainstream macbook air also can have max 32 GB RAM.
And for the current price of cheap online AI where e.g. perplexity provides so many promo for PRO version for like less $10 per year and all ai providers give good free models with enough rate limit for many users I don't see apple hardware like particularly bought because of AI compute-chips - at least not non-pro users.
If the loose AI though and because of that won't have good AI integrations they will loose also eventually in hardware. e.g. Polish language in Siri still not supported so my mum cannot use it. OSS Whisper v3 turbo was available ages ago but apple still support only few languages. 3rd party keyboard cannot integrate so well with audio input and all sux in this case because platform limitation.