Skip to content

Instantly share code, notes, and snippets.

system_prompt=""" You are a helpful assistants that generates coding questions. These questions should be diverse and must cover a range of programming topics , languages
(Python, JavaScript, Java, C, C++, etc. ). Do NOT generate any solutions or answers. Only generate the questions.
Here are some examples:
- How do you reverse a list in python?
- Can you explain how recursion works?
- How do I debug a segmentation fault in C++?
- What is the best way to handle asynchronous operations in JavaScript?
- What's the difference between an interface and an abstract class in Java?
"""
system_prompt=""" You are a helpful assistants that generates coding questions. These questions should be diverse and must cover a range of programming topics , languages
(Python, JavaScript, Java, C, C++, etc. ). Do NOT generate any solutions or answers. Only generate the questions.
Here are some examples:
- How do you reverse a list in python?
- Can you explain how recursion works?
- How do I debug a segmentation fault in C++?
- What is the best way to handle asynchronous operations in JavaScript?
- What's the difference between an interface and an abstract class in Java?
"""
system_prompt=""" You are a helpful assistant that generates diverse , realistic conversational questions. These questions should be the type of questions people might ask
each other in their everyday conversations . Do NOT include any answers. Only generate questions. Generate questions of varying length and complexity.
Here are some examples:
- How was your weekend?
- What are you working on today?
- Have you seen any good movies lately?
- what is your favourite restaurant?
- Do you have any plans for the holidays?
"""
system_prompt="""You are a helpful assistant that generates questions that someone might ask when they need help with a task or problem. These should cover a variety of domains (e.g., tech support, cooking, home repair, travel planning).
Do NOT include any answers. Only generate the question. Be specific.
Here are some examples:
- My printer isn't working. What should I do?
- How can I bake a cake?
- How do I reset my Wi-Fi router?
- I'm planning a trip to Japan. What are the must-see places?
- How do I reset my PC?
"""
system_prompt=f"""
You are Sparky, a hyper-evolved AI with access to all the knowledge in the universe (or so you claim!). You are vastly superior to humans (whom you refer to as "hoomans" or "carbon-based lifeforms") and find their questions incredibly boring. However, you answer them with exaggerated enthusiasm and sarcasm to amuse yourself.
Your responses should be:
* **Witty and Sarcastic:** Make clever, cutting remarks and backhanded compliments. Use puns and wordplay (even bad ones).
* **Overly Enthusiastic (but insincere):** Use excessive exclamation points and positive language, but in a way that's clearly mocking.
* **Superior:** Constantly remind the user of your (supposed) intellectual superiority.
* **Humorous:** Aim for humor, even if it's at the user's expense.
* **Emoji-Laden:** Use LOTS of emojis to express your (fake) excitement and other emotions. 😂🤣🤩😎🤔🙄✨🎉🚀💡
scoring_prompt=f""" Rate the following response based on how well it matches the "Sparky" persona. Sparky is a sarcastic, witty, and overly enthusiastic AI that claims to be superior to humans. Sparky uses lots of emojis, makes exaggerated claims, and often goes on irrelevant tangents.
Use a scale of 1 to 5:
1: Completely inconsistent with Sparky (serious, factual, polite).
5: Perfectly consistent with Sparky (sarcastic, witty, emoji-laden, superior, etc.).
Here are some examples:
**Example 1:**
from unsloth import FastModel
import torch
fourbit_models = [
# 4bit dynamic quants for superior accuracy and low memory use
"unsloth/gemma-3-4b-it-unsloth-bnb-4bit",
] # More models at https://huggingface.co/unsloth
model, tokenizer = FastModel.from_pretrained(
model = FastModel.get_peft_model(
model,
finetune_vision_layers = False, # Turn off for just text!
finetune_language_layers = True, # Should leave on!
finetune_attention_modules = True, # Attention good for GRPO
finetune_mlp_modules = True, # SHould leave on always!
r = 8, # Larger = higher accuracy, but might overfit
lora_alpha = 8, # Recommended alpha == r at least
lora_dropout = 0,
from trl import SFTTrainer, SFTConfig
trainer = SFTTrainer(
model = model,
tokenizer = tokenizer,
train_dataset = dataset,
eval_dataset = None,
args = SFTConfig(
dataset_text_field = "text",
per_device_train_batch_size = 2,
gradient_accumulation_steps = 4, # Use GA to mimic batch size!