Last active
December 9, 2022 08:00
-
-
Save younesbelkada/073f0b7902cbed2cbff662996a74162e to your computer and use it in GitHub Desktop.
An example script to run bnb int8 models using `bitsandbytes` and `transformers`
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
import torch | |
from transformers import AutoModelForCausalLM, AutoTokenizer | |
MAX_NEW_TOKENS = 128 | |
model_name = 'facebook/opt-66b' | |
text = """ | |
Q: On average Joe throws 25 punches per minute. A fight lasts 5 rounds of 3 minutes. | |
How many punches did he throw?\n | |
A: Let’s think step by step.\n""" | |
tokenizer = AutoTokenizer.from_pretrained(model_name) | |
input_ids = tokenizer(text, return_tensors="pt").input_ids | |
free_in_GB = int(torch.cuda.mem_get_info()[0]/1024**3) | |
max_memory = f'{free_in_GB-2}GB' | |
n_gpus = torch.cuda.device_count() | |
max_memory = {i: max_memory for i in range(n_gpus)} | |
model = AutoModelForCausalLM.from_pretrained( | |
model_name, | |
device_map='auto', | |
load_in_8bit=True, | |
max_memory=max_memory | |
) | |
generated_ids = model.generate(input_ids, max_length=MAX_NEW_TOKENS) | |
print(tokenizer.decode(generated_ids[0], skip_special_tokens=True)) |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment