This PR is an ongoing effort to add a CUDA backend to MLX, very little things work now but you can run the tutorial example already.
To build and test:
$ cmake . -Bbuild -DMLX_BUILD_CUDA=ON -DMLX_B...
This is enabling MLX models to run on things that aren’t apple cpus. Otherwise the format is of no use to anyone not using apple hardware, which is a large portion of the LLM community, and computer users in general.
This is enabling MLX models to run on things that aren’t apple cpus. Otherwise the format is of no use to anyone not using apple hardware, which is a large portion of the LLM community, and computer users in general.