Last active
July 25, 2023 16:16
-
-
Save tempusfrangit/7b9495c57458a774fa442640598dd97b to your computer and use it in GitHub Desktop.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
# Clone repo | |
git clone https://github.com/ggerganov/llama.cpp | |
cd llama.cpp | |
# Build | |
mkdir build | |
cd build | |
cmake .. | |
cmake --build . --config Release | |
cd .. | |
# Download model | |
export MODEL=llama-2-13b-chat.ggmlv3.q4_0.bin | |
if [ ! -f models/${MODEL} ]; then | |
curl -L "https://huggingface.co/TheBloke/Llama-2-13B-chat-GGML/resolve/main/${MODEL}" -o models/${MODEL} | |
fi | |
# Set prompt | |
PROMPT="Hello! How are you?" | |
# Run in interactive mode | |
./build/bin/main -m ./models/llama-2-13b-chat.ggmlv3.q4_0.bin \ | |
--color \ | |
--ctx_size 2048 \ | |
-n -1 \ | |
-ins -b 256 \ | |
--top_k 10000 \ | |
--temp 0.2 \ | |
--repeat_penalty 1.1 \ | |
-t 8 |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment