-
Couldn't load subscription status.
- Fork 215
lm-eval for llama.cpp enhancement. #1543
base: main
Are you sure you want to change the base?
Conversation
⛈️ Required checks status: Has failure 🔴
Groups summary🔴 Format Scan Tests workflow
These checks are required after the changes to 🔴 Optimize Unit Test workflow
These checks are required after the changes to 🟢 NeuralChat Unit Test
These checks are required after the changes to 🟢 Engine Unit Test workflow
These checks are required after the changes to 🟢 Chat Bot Test workflow
These checks are required after the changes to Thank you for your contribution! 💜
|
for more information, see https://pre-commit.ci
usages:CPUGPU |
Type of Change
enable lm-eval for llama.cpp models
API not changed
Description
refer to the lm-eval official code and llama-cpp-python
improvements:
bos_idfor qwen models because thellama-cpp-pythondoesn't addbos_idsuccessfully. Even though some changes for qwen, I still find that the tokenizer results are different between llama.cpp and huggingface/transformers. I will verify this further.llama-cpp-python, I implement it with a custom class, which can accelerate the post-process.