# llama-cpp-python — Free-Tier Friendly Wheel This Space provides a **prebuilt `llama-cpp-python` wheel** designed to work **reliably on Hugging Face Free tier Spaces**. No compilation. No system packages. No build failures. If your Space crashes during `pip install llama-cpp-python`, this wheel is the fix. --- ## Optimized for Hugging Face Free Tier Hugging Face Free tier Spaces are: - CPU-only - Limited in memory - Not suitable for native compilation This wheel is built **ahead of time** so it can be installed instantly without triggering CMake, compilers, or BLAS detection. --- ## What this wheel gives you - ✅ Works on **HF Free tier CPU Spaces** - ✅ Linux (ubuntu-22.04 compatible) - ✅ Python 3.10 - ✅ OpenBLAS enabled (`GGML_BLAS=ON`) - ✅ No system dependencies required - ✅ No build step during Space startup - ✅ Fast, reliable `pip install` --- ## How to use in a Space (Free tier) 1. Download the wheel from the GitHub repository 2. Upload it to your Space 3. Install it in your Space startup: pip install llama_cpp_python-*.whl> ## That’s it — your Space will start without build errors. ## Build details This wheel was built using: abetlen/llama-cpp-python (recursive submodules) OpenBLAS (GGML_VENDOR=OpenBLAS) scikit-build-core ninja python -m build --wheel --no-isolation ## Build environment: OS: Ubuntu 22.04 Python: 3.10 ## Why not build from source on HF? On Free tier Spaces, building from source often fails due to: Missing compilers Missing BLAS libraries Memory limits Build timeouts This prebuilt wheel avoids all of those issues. ## Notes CPU-only (no CUDA) Intended for inference workloads Not an official upstream release ## Credits All credit goes to the maintainers of llama-cpp-python and llama.cpp. This Space exists solely to make Free tier usage painless.