Apple Releases Open Source AI Models That Run On-Device
macrumors.comAn LLM in my pocket is truly a mind-blowing concept, I have to say. More than anything else — phone, camera, internet. The feels like a really big deal.
And with regard to LLMs (AI?) in general, I don't think right now we have any idea what we will all be using them for in ten years. But it just feels like a fundamental change is coming from all this.
Discussion: [0] (33 points, 18 hours ago, 7 comments)
I'm not knowledgeable enough to parse much out of the Readme.
How "good" are the models approximately? What hardware do I need to run them? How fast are they?
[dupe]
Some more discussion: https://news.ycombinator.com/item?id=40140675
Has anyone seen a working, clearly explained recipe for running this using the Python MLX library on macOS yet?
What’s there to explain? There’s a readme in the repo that shows how to do it.
I tried and failed to follow that. I'm looking for a report from someone who has got it (or something like it) to work.
I’ve tried their 1.1B model. The only hiccup was that it seems to require mlx 0.10.0 which is what’s in requirements.txt. You also have to place the llama tokenizer file into the model dir - they do not distribute it. The models published for MLX do not seem to be instruction tuned, so with their default prompt they get repetitive. But I suppose you could convert the instruction tuned checkpoints with the script in the repo.
guessing it'll be available thru ollama soon
Why is the 3B model worse than the 450M model on MMLU and TruthfulQA?
Now we can give credit to Apple for invented AI!
Apple rarely invents the first X.
However, they often invent the first X that doesn't suck.
Turns out that the second one is actually more important.
I never see apple claiming to invent something. I frequently see comments like yours however.