llama.cpp 在 intel cpu 執行最好的後端

立体风發表於2024-07-13

在 cpu 下執行 llama.cpp

https://github.com/flame/blis 計算庫是目前最好的選擇。

SYCL

SYCL is a high-level parallel programming model designed to improve developers productivity writing code across various hardware accelerators such as CPUs, GPUs, and FPGAs. It is a single-source language designed for heterogeneous computing and based on standard C++17.

The llama.cpp SYCL backend is designed to support Intel GPU firstly. Based on the cross-platform feature of SYCL, it could support other vendor GPUs: Nvidia GPU (AMD GPU coming).
When targeting Intel CPU, it is recommended to use llama.cpp for Intel oneMKL backend.

oneMKL

When targeting Intel CPU, it is recommended to use llama.cpp for Intel oneMKL backend.

It has the similar design of other llama.cpp BLAS-based paths such as OpenBLAS, cuBLAS, etc... In beginning work, the oneAPI's SYCLomatic open-source migration tool (Commercial release Intel® DPC++ Compatibility Tool) was used for this purpose.
現在這個也在 llama.cpp 上刪除支援說明文件了。

相關文章