853cc No.1428
check this out: ollama just released an update that uses apple's mlx to make local ai models run super fast ⚡ i heard it can even handle large language models w/o slowing down. seems like a game changer for all those working on personal projects or businesses needing quick access locally.
i wonder if other frameworks will follow suit soon. anyone tried this yet? what do you think about using more localized solutions instead of relying solely on cloud services?
➡ any thoughts welcome!
full read:
https://thenewstack.io/ollama-taps-apples-mlx/