Skip to content

Fix model unloading issue on AMD GPUs#183

Open
YellowRoseCx wants to merge 1 commit intoZyphra:mainfrom
YellowRoseCx:patch-1
Open

Fix model unloading issue on AMD GPUs#183
YellowRoseCx wants to merge 1 commit intoZyphra:mainfrom
YellowRoseCx:patch-1

Conversation

@YellowRoseCx
Copy link

Currently, switching from the Transformer backend to the Hybrid backend loads the Hybrid model into VRAM without removing the
Transformer model from VRAM on AMD GPUs, eating up about 13gb of VRAM instead of 6.5-7gb

This patch corrects that issue and unloads the model in a cleaner way

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant