wget https://repo.anaconda.com/miniconda/Miniconda3-latest-Linux-x86_64.sh bash Miniconda3-latest-Linux-x86_64.sh
conda create -n vllm python=3.12 conda activate vllm
pip install vllm -i https://pypi.tuna.tsinghua.edu.cn/simple
vllm serve /home/ctq/Huggingface/Qwen2.5-1.5B-Instruct
curl http://localhost:8000/v1/chat/completions \ -H "Content-Type: application/json" \ -d '{ "model": "/home/ctq/Huggingface/Qwen2.5-1.5B-Instruct", "prompt": "San Francisco is a", "max_tokens": 7, "temperature": 0 }'
vllm serve /home/ctq/Huggingface/Qwen2.5-1.5B-Instruct
git clone https://github.com/hiyouga/LLaMA-Factory.git
cd LLaMA-Factory pip install -r requirements.txt
conda activate llamafactory bash qw15.sh
watch -n 2 nvidia-smi
pip install huggingface_hub
huggingface-cli login
huggingface-cli download Qwen/Qwen2.5-3B-Instruct --local-dir Qwen2.5-3B-Instruct
pip install modelscope
modelscope download Qwen/Qwen2.5-3B-Instruct --local_dir Qwen2.5-3B-Instruct
python src/export_model_lora.py \ --model_name_or_path /home/ctq/Huggingface/Qwen2.5-1.5B-Instruct \ --adapter_name_or_path /home/ctq/Pyproject/LLaMA-Factory/outmoxi/Qwen2.5-1.5B-1 \ --template qwen \ --export_dir ./outmoxi/Qwen2.5-1.5B-lora
export http_proxy=192.168.2.218:2023 export https_proxy=192.168.2.218:2023