...
| Code Block |
|---|
sudo apt update sudo apt upgrade sudo add-apt-repository ppa:deadsnakes/ppa sudo apt install python3.11 -y sudo apt install python3.11-venv -y python3.11 -V python3.11 -m venv llm_env source llm_env/bin/activate pip install --pre --upgrade ipex-llm[cpp] mkdir llama-cpp cd llama-cpp # Run Ollama Serve with Intel GPU export OLLAMA_NUM_GPU=128999 export OLLAMA_THREADS=22 export OMP_NUM_THREADS=22 export ZES_ENABLE_SYSMAN=1 export no_proxy=localhost,127.0.0.1 export ZES_ENABLE_SYSMAN=1 source /opt/intel/oneapi/setvars.sh export SYCL_CACHE_PERSISTENT=1 # localhost access # ./ollama serve # for non-localhost access OLLAMA_HOST=0.0.0.0 ./ollama serve |
list models
| Code Block |
|---|
(base) root@server1:~/llama-cpp# ./ollama list NAME ID SIZE MODIFIED phi3:14b cf611a26b048 7.9 GB 3 minutes ago llama3.3:70b a6eb4748fd29 42 GB 16 minutes ago mistral-small3.1:24b b9aaf0c2586a 15 GB 23 minutes ago llama4:scout 4f01ed6b6e01 67 GB 56 minutes ago openchat:7b 537a4e03b649 4.1 GB About an hour ago qwen3:32b e1c9f234c6eb 20 GB 2 hours ago gemma3:27b a418f5838eaf 17 GB 2 hours ago deepseek-r1:70b 0c1615a8ca32 42 GB 3 hours ago |
...