wget https://repo.anaconda.com/miniconda/Miniconda3-latest-Linux-x86_64.sh bash Miniconda3-latest-Linux-x86_64.sh
conda create -n vllm python=3.12 conda activate vllm
pip install vllm -i https://pypi.tuna.tsinghua.edu.cn/simple
vllm serve /home/ctq/Huggingface/Qwen2.5-1.5B-Instruct
curl http://localhost:8000/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{
"model": "/home/ctq/Huggingface/Qwen2.5-1.5B-Instruct",
"prompt": "San Francisco is a",
"max_tokens": 7,
"temperature": 0
}'
vllm serve /home/ctq/Huggingface/Qwen2.5-1.5B-Instruct
git clone https://github.com/hiyouga/LLaMA-Factory.git
cd LLaMA-Factory pip install -r requirements.txt
conda activate llamafactory bash qw15.sh
watch -n 2 nvidia-smi
pip install huggingface_hub
huggingface-cli login
huggingface-cli download Qwen/Qwen2.5-3B-Instruct --local-dir Qwen2.5-3B-Instruct
pip install modelscope
modelscope download Qwen/Qwen2.5-3B-Instruct --local_dir Qwen2.5-3B-Instruct
python src/export_model_lora.py \
--model_name_or_path /home/ctq/Huggingface/Qwen2.5-1.5B-Instruct \
--adapter_name_or_path /home/ctq/Pyproject/LLaMA-Factory/outmoxi/Qwen2.5-1.5B-1 \
--template qwen \
--export_dir ./outmoxi/Qwen2.5-1.5B-lora
export http_proxy=192.168.2.218:2023 export https_proxy=192.168.2.218:2023