llama.app的编译安装以及模型蒸馏-Ubuntu成功篇
操作系统是Ubuntu 22.04.5,搭配AMD 6700 xt 12G的显卡 本来是Debian系统,可是它对rocm的支持没有Ubuntu好,所以只能洗掉Debian,重新安装了Ubuntu 22.04.5,普通用户是ubuntu 我们在这个平台上学习对模型进行蒸馏,首先安装openssh以及rocm 6.2的最新套件,并克隆llama.app: sudo apt -y install openssh-server sudo apt -y install wget ca-certificates gnupg2 curl sudo apt -y install git cmake build-essential pkg-config sudo apt -y install linux-headers-$(uname -r) gcc-12 sudo mkdir -p /etc/apt/keyrings wget -qO- https://repo.radeon.com/rocm/rocm.gpg.key | sudo gpg --dearmor -o /etc/apt/keyrings/rocm.gpg echo "deb [arch=amd64 signed-by=/etc/apt/keyrings/rocm.gpg] https://repo.radeon.com/rocm/apt/6.2 jammy main" sudo tee /etc/apt/preferences.d/rocm-repo-radeon-pin >/dev/null <<'EOF' Package: * Pin: origin "repo.radeon.com" Pin-Priority: 1001 EOF sudo apt update sudo apt -y install rocm-hip-sdk rocm-device-libs rocminfo rocm-smi libssl-dev sudo usermod -a -G video,render $USER hipcc --version rocminfo | head git clone https://github.com/ggerganov/llama.cpp cd llama.cpp/ cmake -S . -B build -DCMAKE_BUILD_TYPE=Release cmake --build build -j 然后安装PyTorch,为了环境干净,用uv来维护python的环境: ...