-
Notifications
You must be signed in to change notification settings - Fork 59
Description
llm_load_tensors: ggml ctx size = 0.41 MiB
ggml_backend_metal_buffer_from_ptr: error: failed to allocate buffer, size = 0.00 MiB
llama_model_load: error loading model: failed to allocate buffer
llama_load_model_from_file: failed to load model
llama_init_from_gpt_params: error: failed to load model '/Users/zhangyixin/Desktop/llama.cpp/TheBloke/Yi-34B-Chat-GGUF/yi-34b-chat.Q8_0.gguf'
main: error: unable to load model
1021 huggingface-cli download --token YOUR_TOKEN --resume-download --local-dir-use-symlinks False TheBloke/Yi-34B-Chat-GGUF --include "yi-34b-chat.Q8_0.gguf" --local-dir TheBloke/Yi-34B-Chat-GGUF\n
1022 chmod +x /Users/zhangyixin/Desktop/llama.cpp/TheBloke/Yi-34B-Chat-GGUF/yi-34b-chat.Q8_0.gguf
1023 /Users/zhangyixin/Desktop/llama.cpp/TheBloke/Yi-34B-Chat-GGUF/yi-34b-chat.Q8_0.gguf
1024 ./main --frequency-penalty 0.5 --frequency-penalty 0.5 --top-k 5 --top-p 0.9 -m /Users/zhangyixin/Desktop/llama.cpp/TheBloke/Yi-34B-Chat-GGUF/yi-34b-chat.Q8_0.gguf -p "Building a website can be done in 10 simple steps:\nStep 1:" -n 400 -e\n
1025 hist