llama.cpp/Dockerfile-llama
zhoupengcheng ee553ce5c4 - rename /usr/bin/convert.py
- update long-term yum.repo in dockerfile
2023-09-21 10:35:52 +08:00

25 lines
1.0 KiB
Plaintext

#Usage:
#1.build image:
# docker build -f Dockerfile-llama -t llama_image .
#2.run image:
# docker run -it --security-opt seccomp=unconfined llama_image:latest
#base image
FROM openeuler/openeuler:22.03
#update openEuler2309 source and install chatglm
RUN echo '[everything]' > /etc/yum.repos.d/openEuler.repo && \
echo 'name=everything' >> /etc/yum.repos.d/openEuler.repo && \
echo 'baseurl=http://121.36.84.172/dailybuild/EBS-openEuler-23.09/rc4_openeuler-2023-09-13-21-46-47/everything/$basearch/' >> /etc/yum.repos.d/openEuler.repo && \
echo 'enabled=1' >> /etc/yum.repos.d/openEuler.repo && \
echo 'gpgcheck=0' >> /etc/yum.repos.d/openEuler.repo && \
yum install -y llama.cpp wget
#download ggml model
WORKDIR /model_path
RUN wget -P /model_path https://huggingface.co/TheBloke/Llama-2-13B-chat-GGML/resolve/main/llama-2-13b-chat.ggmlv3.q4_0.bin
# run ggml model
CMD /usr/bin/llama_cpp_main -m /model_path/llama-2-13b-chat.ggmlv3.q4_0.bin --color --ctx_size 2048 -n -1 -ins -b 256 --top_k 10000 --temp 0.2 --repeat_penalty 1.1 -t 8