Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Apple was always going to fail this, and even more so going forward.

LLM are built on data, and copious amounts of it. Apple has been on a decade long marketing campaign to make data radio active. It has now permeated the culture so much so that, Apple CANNOT build a proprietary, world-class AI product without compromising on their outspoken positions.

It is a losing battle because the more apple wants to do it, the users are gonna punish them and meanwhile, other companies (ChatGPT, anthropic) are gonna extract maximum value.



LLMs are mostly trained on public data while Apple's privacy stance applies to private data. There's no conflict between them.

(Meta can probably train on private data but OpenAI and Anthropic seem to be doing OK without it as far as we know.)


All the LLM advances these days are from synthetic or explicitly created data too. You need public data mostly because it contains facts about the world, or because it's easier to talk about a book when it's "read" the book. But for a known topic area (as opposed to open Q&A) it's not critical since you can go and create or license it.


No, Apple’s privacy stance is about giving users control over data in ways they understand. Posting on Reddit or Arxiv is not a blank check to have your words be reused for LLM training, even if it’s technically public.


Apple’s slogan is “what happens on your iPhone stays on your iPhone”. I think “I published a paper” or “I posted on Reddit” are clearly out of scope - those things are happening in public.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: