mirror of
https://github.com/kvcache-ai/ktransformers.git
synced 2026-03-15 02:47:22 +00:00
fix install issue & add test
This commit is contained in:
29
.github/workflows/install.yml
vendored
29
.github/workflows/install.yml
vendored
@@ -6,11 +6,12 @@ on:
|
||||
job_to_run:
|
||||
description: "Which job to run?"
|
||||
required: true
|
||||
default: "install"
|
||||
default: "install&test"
|
||||
type: choice
|
||||
options:
|
||||
- create&install
|
||||
- install
|
||||
- create&install&&test
|
||||
- install&&test
|
||||
- test
|
||||
jobs:
|
||||
Install-KTransformers:
|
||||
runs-on: self-hosted
|
||||
@@ -22,31 +23,43 @@ jobs:
|
||||
- run: echo "💡 The ${{ github.repository }} repository has been cloned to the runner."
|
||||
- name: Remove old conda environment
|
||||
continue-on-error: true
|
||||
if: ${{ inputs.job_to_run == 'create&install'}}
|
||||
if: contains(inputs.job_to_run, 'create')
|
||||
run: |
|
||||
source /home/qujing3/anaconda3/etc/profile.d/conda.sh
|
||||
conda env remove --name ktransformers-dev -y
|
||||
- name: Create conda environment
|
||||
if: ${{ inputs.job_to_run == 'create&install'}}
|
||||
if: contains(inputs.job_to_run, 'create')
|
||||
run: |
|
||||
source /home/qujing3/anaconda3/etc/profile.d/conda.sh
|
||||
conda create --name ktransformers-dev python=3.11
|
||||
conda activate ktransformers-dev
|
||||
conda install -c conda-forge libstdcxx-ng -y
|
||||
- name: Install dependencies
|
||||
if: ${{ inputs.job_to_run == 'create&install'}}
|
||||
if: contains(inputs.job_to_run, 'create')
|
||||
run: |
|
||||
source /home/qujing3/anaconda3/etc/profile.d/conda.sh
|
||||
conda activate ktransformers-dev
|
||||
pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu126
|
||||
pip3 install packaging ninja cpufeature numpy
|
||||
pip install https://github.com/Dao-AILab/flash-attention/releases/download/v2.7.4.post1/flash_attn-2.7.4.post1+cu12torch2.2cxx11abiTRUE-cp311-cp311-linux_x86_64.whl
|
||||
pip install ~/flash_attn-2.7.4.post1+cu12torch2.6cxx11abiTRUE-cp311-cp311-linux_x86_64.whl
|
||||
- name: Install KTransformers
|
||||
if: contains(inputs.job_to_run, 'install')
|
||||
run: |
|
||||
source /home/qujing3/anaconda3/etc/profile.d/conda.sh
|
||||
conda activate ktransformers-dev
|
||||
pip3 uninstall ktransformers -y
|
||||
cd ${{ github.workspace }}
|
||||
git submodule init
|
||||
git submodule update
|
||||
USE_NUMA=1 bash install.sh
|
||||
bash install.sh
|
||||
- name: Test Local Chat
|
||||
run: |
|
||||
source /home/qujing3/anaconda3/etc/profile.d/conda.sh
|
||||
conda activate ktransformers-dev
|
||||
export PATH=/usr/local/cuda-12.4/bin:$PATH
|
||||
export LD_LIBRARY_PATH=/usr/local/cuda-12.4/lib64:$LD_LIBRARY_PATH
|
||||
export CUDA_HOME=/usr/local/cuda-12.4
|
||||
cd ${{ github.workspace }}
|
||||
python ktransformers/local_chat.py --model_path /home/qujing3/models/DeepSeek-R1-Q4_K_M/config --gguf_path /home/qujing3/models/DeepSeek-R1-Q4_K_M/ --max_new_tokens 100 --cache_len 1536 --cpu_infer 64 --prompt_file /home/qujing3/promptsbook.txt
|
||||
DeepSeek-R1-Q4_K_M/config --gguf_path /home/qujing3/models/DeepSeek-R1-Q4_K_M/ --max_new_tokens 100 --cache_len 1536 --cpu_infer 64 --prompt_file /home/qujing3/prompts/chinese.txt
|
||||
- run: echo "This job's status is ${{ job.status }}."
|
||||
|
||||
Reference in New Issue
Block a user