DRAFT
https://github.com/ggml-org/llama.cpp
https://github.com/intel/ipex-llm/blob/main/docs/mddocs/Quickstart/llama_cpp_quickstart.md
https://github.com/ollama/ollama
https://ollama.com/library/llama4
https://ollama.com/library/qwen3
https://ollama.com/library/deepseek-r1
https://www.intel.com/content/www/us/en/developer/tools/oneapi/dpc-compiler.html
https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.1-GGUF/tree/main
Intel tools
Intel oneAPI
oneMKL - OneAPI Math Kernel Library, OneDNN - oneAPI Deep Neural Network Library
apt update apt install -y gpg-agent wget wget -O- https://apt.repos.intel.com/intel-gpg-keys/GPG-PUB-KEY-INTEL-SW-PRODUCTS.PUB | gpg --dearmor | tee /usr/share/keyrings/oneapi-archive-keyring.gpg > /dev/null echo "deb [signed-by=/usr/share/keyrings/oneapi-archive-keyring.gpg] https://apt.repos.intel.com/oneapi all main" | tee /etc/apt/sources.list.d/oneAPI.list apt update
apt install -y intel-oneapi-base-toolkit #
check
. /opt/intel/oneapi/2025.1/oneapi-vars.sh oneapi-cli
Intel C++ essentials
#apt -y install intel-cpp-essentials #apt -y install cmake pkg-config build-essential
Intel Ai Framework and tools
AI Tools
wget https://registrationcenter-download.intel.com/akdlm/IRC_NAS/491d5c2a-67fe-48d0-884f-6aecd88f5d8a/ai-tools-2025.0.0.75_offline.sh sh ai-tools-2025.0.0.75_offline.sh
OpenVINO
https://www.intel.com/content/www/us/en/developer/tools/openvino-toolkit/download.html
wget https://apt.repos.intel.com/intel-gpg-keys/GPG-PUB-KEY-INTEL-SW-PRODUCTS.PUB sudo apt-key add GPG-PUB-KEY-INTEL-SW-PRODUCTS.PUB echo "deb https://apt.repos.intel.com/openvino ubuntu24 main" | sudo tee /etc/apt/sources.list.d/intel-openvino.list sudo apt update apt-cache search openvino sudo apt install openvino-2025.1.0
OpenVINO™ Model Server
wget https://github.com/openvinotoolkit/model_server/releases/download/v2025.1/ovms_ubuntu24_python_on.tar.gz
tar -xzvf ovms_ubuntu24_python_on.tar.gz
export LD_LIBRARY_PATH=${PWD}/ovms/lib
export PATH=$PATH:${PWD}/ovms/bin
curl --create-dirs -k https://storage.openvinotoolkit.org/repositories/open_model_zoo/2022.1/models_bin/2/resnet50-binary-0001/FP32-INT1/resnet50-binary-0001.xml -o models/resnet50/1/model.xml
curl --create-dirs -k https://storage.openvinotoolkit.org/repositories/open_model_zoo/2022.1/models_bin/2/resnet50-binary-0001/FP32-INT1/resnet50-binary-0001.bin -o models/resnet50/1/model.bin
chmod -R 755 models
export PYTHONPATH=${PWD}/ovms/lib/python
sudo apt -y install libpython3.12
pip3 install "Jinja2==3.1.6" "MarkupSafe==3.0.2"
ovms --port 9000 --model_name resnet --model_path models/resnet50
llama.cpp
https://github.com/ggml-org/llama.cpp
build with CPU backend
apt install -y libcurl-ocaml-dev git clone https://github.com/ggml-org/llama.cpp cd llama.cpp cmake -B build cmake --build build --config Release cd build make install ldconfig
Intel oneMKL
use
llama-cli -m model.gguf llama-server -m model.gguf --port 8080 llama-bench -m model.gguf llama-run