diff --git a/README.md b/README.md index 798c0e951..540b76de7 100644 --- a/README.md +++ b/README.md @@ -4,7 +4,7 @@ [BitNet Model on Hugging Face](https://huggingface.co/microsoft/BitNet-b1.58-2B-4T) -Try it out via this [demo](https://bitnet-demo.azurewebsites.net/), or build and run it on your own [CPU](https://github.com/microsoft/BitNet?tab=readme-ov-file#build-from-source) or [GPU](https://github.com/microsoft/BitNet/blob/main/gpu/README.md). +Try it out via this [demo](https://bitnet-demo.azurewebsites.net/), [Collab notebook](https://colab.research.google.com/drive/1YWEISZD3OTr4L11g8-PK4l0H7dQXarUo#scrollTo=cIFDpd9TzsAM), or [build and run](https://github.com/microsoft/BitNet?tab=readme-ov-file#build-from-source) it on your own CPU. bitnet.cpp is the official inference framework for 1-bit LLMs (e.g., BitNet b1.58). It offers a suite of optimized kernels, that support **fast** and **lossless** inference of 1.58-bit models on CPU and GPU (NPU support will coming next).