This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
base_model: google/gemma-3-27b-it | |
#model_type: AutoModelForCausalLM | |
#tokenizer_type: AutoTokenizer | |
deepspeed: deepspeed_config.json | |
load_in_8bit: false | |
load_in_4bit: true | |
#strict: false |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
base_model: google/gemma-2-27b | |
lora_model_dir: /path/to/previous/axolotl/run/output/ | |
#model_type: AutoModelForCausalLM | |
#tokenizer_type: AutoTokenizer | |
save_safetensors: true | |
load_in_8bit: false | |
load_in_4bit: true | |
strict: false |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
""" | |
Declarative Prompts, Generative program generators, oh my... | |
Takes a minimal Prompt Declaration Language (PDL) file and generates a finite state generative machine | |
as Python objects for a subset of the PDL language [1], [2]. These objects (the "programs" in particular) | |
can be executed, and their Model class can be extended to incorporate the functionality for | |
evaluating the prompts against the models specified in PDL using any accumulated conversational | |
context, prompts, and generation parameters (sampling parameters, for example), (optionally) updating | |
the context as the programs execution continues. A generative, conversational workflow system |