Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

It's an optimized backend for running LLMs, much like CoreML on macOS, which has been received very positively due to the acceleration it enables (ollama/llama.cpp use it).

Since this uses ONNX you probably won't be able to use ollama directly with it, but conceptually you could use an app like it to run your models in a more optimized way.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: