Skip to content

Instantly share code, notes, and snippets.

View raphael-sch's full-sized avatar

Raphael raphael-sch

View GitHub Profile
@raphael-sch
raphael-sch / run_left_padding_train_example.py
Last active May 7, 2024 16:28
Training and position_ids with left padding
import argparse
import transformers
import torch
from datasets import Dataset
from transformers import AutoTokenizer, AutoModelForCausalLM
parser = argparse.ArgumentParser(description='Define experiment parameters')
parser.add_argument('--use_custom_position_ids', default='no', choices=['yes', 'no'], type=str)
parser.add_argument('--model_name', default='meta-llama/Llama-2-7b-hf', type=str)
@raphael-sch
raphael-sch / run_padding_prefill.py
Last active March 21, 2024 14:49
Using padding and prefill during inference in huggingface transformers
import re
import sys
import time
import tqdm
import torch
from datasets import load_dataset, concatenate_datasets
from transformers import AutoTokenizer, LlamaForCausalLM