-
Create a new conda environment and install the basic dependencies
conda create -n mllm-efficiency python=3.10 conda activate mllm-efficiency pip install -r requirements.txt pip install ninja pip install omegaconf pip install flash-attention-softmax-n conda install pytorch==2.3.0 torchvision==0.18.0 torchaudio==2.3.0 pytorch-cuda=12.1 -c pytorch -c nvidia conda install nvidia/label/cuda-12.1.1::cuda-nvcc
-
Change the env path
mkdir -p $CONDA_PREFIX/etc/conda/activate.d mkdir -p $CONDA_PREFIX/etc/conda/deactivate.d
Create a new file in the activate.d directory and add the following content:
#!/bin/bash export CUDA_HOME=$(dirname $(dirname $(which nvcc)))
Create a new file in the deactivate.d directory and add the following content:
#!/bin/bash unset CUDA_HOME
-
Install the flash-attn
conda activate mllm-efficiency echo $CUDA_HOME which nvcc pip install flash-attn --no-build-isolation
-
use lmms-eval
cd lmms-eval pip install -e . cd ../llava/ pip install -e . pip install numpy==2.2.0
MUST use scripts in base folder, or will raise ModuleNotFoundError
-
use qwen2_vl for develop
cd qwen2vl pip install -e .
The KV cache cluster (managed by the
xxxKVClusterclass) is responsible for selecting and compressing the KV states (keys and values), while the actual storage and dynamic updates of the compressed KV states are still handled by thepast_key_values(aka.Dynamiccacheclass intransformerslibraries), which remains the central mechanism for autoregressive decoding in Transformers. This separation allows for flexible and efficient cache management, critical for scaling LLMs in constrained environments.
Tip
Step-by-Step Code construction(take exampleKVCache as example.)
-
Initialization of the KV Cache Compression System.
Code:
init_exampleMLLMdef init_exampleMLLM(self): # some configs setting self.kv_cluster = exampleMLLMKVCluster( # some configs setting )
-
Forward Pass in Attention.
Code:
qwen_attn_forward_exampleMLLMdef llama_attn_forward_exampleMLLM(...): bsz, q_len, _ = hidden_states.size() init_exampleMLLM(self) ... key_states_compress, value_states_compress = self.kv_cluster.update_kv( key_states, query_states, value_states, attention_mask, self.num_key_value_groups ) past_key_value.update(key_states_compress, value_states_compress, self.layer_idx, cache_kwargs)
if the method need allocate budgets from all layers, initialize in decoder layer class or do some change(like self.budgets in init_exampleMLLM). Take dynamickv from llm as example to handle this problem.
-
Managing and Compressing KV Cache
Code:
exampleMLLMKVClusterandupdate_kvMethod in this class
Important
Above steps is the basic implementation from kv-cache-factory. BUT the disadvantage is that we can not evict cache during decoding since only current qkv are in
./scripts/run_eval.sh textvqa_val /share/home/mhma/models/llava-onevision-qwen2-7b-ov streamingllm 0.5- hyperparams: model_args:pretrained,conv_template,model_name,method,...
def __init__(
self,
pretrained: str = "/share/home/mhma/models/llava-onevision-qwen2-7b-ov",
truncation: Optional[bool] = True,
device: Optional[str] = "cuda:0",
batch_size: Optional[Union[int, str]] = 1,
model_name: Optional[str] = 'llava_qwen',
attn_implementation: Optional[str] = best_fit_attn_implementation,
device_map: Optional[str] = "cuda:0",
conv_template: Optional[str] = "qwen_1_5",
use_cache: Optional[bool] = True,
truncate_context: Optional[bool] = False, # whether to truncate the context in generation, set it False for LLaVA-1.6
customized_config: Optional[str] = None, # ends in json
max_frames_num: Optional[int] = 32,
mm_spatial_pool_stride: Optional[int] = 2,
mm_spatial_pool_mode: Optional[str] = "bilinear",
token_strategy: Optional[str] = "single", # could be "single" or "multiple", "multiple" denotes adding multiple <image> tokens for each frame
video_decode_backend: str = "decord",
method: Optional[str] = None, # None is not used kv cache.
**kwargs,
)If you are trying to use large LLaVA models such as LLaVA-NeXT-Qwen1.5-72B, you can try adding device_map=auto in model_args and change num_processes to 1.
-
command line use
This mode supports a number of command-line arguments, the details of which can be also be seen via running with
-hor--help:--model: Selects which model type or provider is evaluated. Must be a string corresponding to the name of the model type/provider being used. See the main README for a full list of enabled model names and supported libraries or APIs.
-
--model_args: Controls parameters passed to the model constructor. Accepts a string containing comma-separated keyword arguments to the model class of the format"arg1=val1,arg2=val2,...", such as, for example--model_args pretrained=liuhaotian/llava-v1.5-7b,batch_size=1. For a full list of what keyword arguments, see the initialization of the corresponding model class inlmms_eval/models/. -
--tasks: Determines which tasks or task groups are evaluated. Accepts a comma-separated list of task names or task group names. Must be solely comprised of valid tasks/groups. You can use--tasks listto see all the available tasks. If you add your own tasks but not shown on the list, you can try to set--verbosity=DEBUGto view the error message. You can also use--tasks list_with_numto check every tasks and the number of question each task contains. However,list_with_numwill download all the available datasets and may require lots of memory and time.
-
--num_fewshot: Sets the number of few-shot examples to place in context. Must be an integer. -
--gen_kwargs: takes an arg string in same format as--model_argsand creates a dictionary of keyword arguments. These will be passed to the models for all calledgenerate_until(free-form or greedy generation task) tasks, to set options such as the sampling temperature ortop_p/top_k. For a list of what args are supported for each model type, reference the respective library's documentation (for example, the documentation fortransformers.AutoModelForCausalLM.generate().) These kwargs will be applied to allgenerate_untiltasks called--we do not currently support unique gen_kwargs or batch_size values per task in a single run of the library. To control these on a per-task level, set them in that task's YAML file. -
--batch_size: Sets the batch size used for evaluation. Can be a positive integer or"auto"to automatically select the largest batch size that will fit in memory, speeding up evaluation. One can pass--batch_size auto:Nto re-select the maximum batch sizeNtimes during evaluation. This can help accelerate evaluation further, sincelm-evalsorts documents in descending order of context length. -
--max_batch_size: Sets the maximum batch size to try to fit in memory, if--batch_size autois passed. -
--device: Sets which device to place the model onto. Must be a string, for example,"cuda", "cuda:0", "cpu", "mps". Defaults to "cuda", and can be ignored if running multi-GPU or running a non-local model type. -
--output_path: A string of the formdir/file.jsonlordir/. Provides a path where high-level results will be saved, either into the file named or into the directory named. If--log_samplesis passed as well, then per-document outputs and metrics will be saved into the directory as well. -
--log_samples: If this flag is passed, then the model's outputs, and the text fed into the model, will be saved at per-document granularity. Must be used with--output_path. -
--limit: Accepts an integer, or a float between 0.0 and 1.0 . If passed, will limit the number of documents to evaluate to the first X documents (if an integer) per task or first X% of documents per task. Useful for debugging, especially on costly API models. -
--use_cache: Should be a path where a sqlite db file can be written to. Takes a string of format/path/to/sqlite_cache_in order to create a cache db at/path/to/sqlite_cache_rank{i}.dbfor each process (0-NUM_GPUS). This allows results of prior runs to be cached, so that there is no need to re-run results in order to re-score or re-run a given (model, task) pair again. -
--cache_requests: Can be "true", "refresh", or "delete". "true" means that the cache should be used. "refresh" means that you wish to regenerate the cache, which you should run if you change your dataset configuration for a given task. "delete" will delete the cache. Cached files are stored under lm_eval/cache/.cache unless you specify a different path via the environment variable:LM_HARNESS_CACHE_PATH. e.g.LM_HARNESS_CACHE_PATH=~/Documents/cache_for_lm_harness. -
--check_integrity: If this flag is used, the library tests for each task selected are run to confirm task integrity. -
--write_out: Used for diagnostic purposes to observe the format of task documents passed to a model. If this flag is used, then prints the prompt and gold target string for the first document of each task. -
--show_config: If used, prints the fulllm_eval.api.task.TaskConfigcontents (non-default settings the task YAML file) for each task which was run, at the completion of an evaluation. Useful for when one is modifying a task's configuration YAML locally to transmit the exact configurations used for debugging or for reproducibility purposes. -
--include_path: Accepts a path to a folder. If passed, then all YAML files containinglm-evalcompatible task configurations will be added to the task registry as available tasks. Used for when one is writing config files for their own task in a folder other thanlm_eval/tasks/. -
--system_instruction: Specifies a system instruction string to prepend to the prompt. -
--apply_chat_template: This flag specifies whether to apply a chat template to the prompt. It can be used in the following ways:--apply_chat_template: When used without an argument, applies the only available chat template to the prompt. For Hugging Face models, if no dedicated chat template exists, the default chat template will be applied.--apply_chat_template template_name: If the model has multiple chat templates, apply the specified template to the prompt.
For Hugging Face models, the default chat template can be found in the
default_chat_templateproperty of the Transformers Tokenizer. -
--fewshot_as_multiturn: If this flag is on, the Fewshot examples are treated as a multi-turn conversation. Questions are provided as user content and answers are provided as assistant responses. Requires--num_fewshotto be set to be greater than 0, and--apply_chat_templateto be on. -
--predict_only: Generates the model outputs without computing metrics. Use with--log_samplesto retrieve decoded results.
-
--seed: Set seed for python's random, numpy and torch. Accepts a comma-separated list of 3 values for python's random, numpy, and torch seeds, respectively, or a single integer to set the same seed for all three. The values are either an integer or 'None' to not set the seed. Default is0,1234,1234(for backward compatibility). E.g.--seed 0,None,8setsrandom.seed(0)andtorch.manual_seed(8). Here numpy's seed is not set since the second value isNone. E.g,--seed 42sets all three seeds to 42. -
--wandb_args: Tracks logging to Weights and Biases for evaluation runs and includes args passed towandb.init, such asprojectandjob_type. Full list here. e.g.,--wandb_args project=test-project,name=test-run -
--hf_hub_log_args: Logs evaluation results to Hugging Face Hub. Accepts a string with the arguments separated by commas. Available arguments:hub_results_org- organization name on Hugging Face Hub, e.g.,EleutherAI. If not provided, the results will be pushed to the owner of the Hugging Face token,hub_repo_name- repository name on Hugging Face Hub (deprecated,details_repo_nameandresults_repo_nameshould be used instead), e.g.,lm-eval-results,details_repo_name- repository name on Hugging Face Hub to store details, e.g.,lm-eval-results,results_repo_name- repository name on Hugging Face Hub to store results, e.g.,lm-eval-results,push_results_to_hub- whether to push results to Hugging Face Hub, can beTrueorFalse,push_samples_to_hub- whether to push samples results to Hugging Face Hub, can beTrueorFalse. Requires--log_samplesto be set,public_repo- whether the repository is public, can beTrueorFalse,leaderboard_url- URL to the leaderboard, e.g.,https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard.point_of_contact- Point of contact for the results dataset, e.g.,yourname@example.com.gated- whether to gate the details dataset, can beTrueorFalse.