Created
April 4, 2025 17:58
-
-
Save rakki194/d58aec059d3a7c0cef911fedbb8da763 to your computer and use it in GitHub Desktop.
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
#!/usr/bin/env python | |
# -*- coding: utf-8 -*- | |
""" | |
JoyCaption Alpha Two | |
This module provides functionality for generating captions for images using a | |
combination of CLIP, LLM, and custom image adapters. It supports various | |
caption types, tones, and lengths. | |
The main components include: | |
- Loading and initializing models (CLIP, LLM, image adapter) | |
- Processing images and generating captions | |
- Command-line interface for batch processing images in a directory | |
""" | |
import os | |
import argparse | |
import re | |
import random | |
import math | |
from pathlib import Path | |
from typing import List, Tuple, Dict | |
from PIL import Image | |
import pillow_jxl # type: ignore | |
import torch | |
import torchvision.transforms.functional as TVF | |
from transformers import ( | |
AutoModel, | |
AutoTokenizer, | |
AutoModelForCausalLM, | |
PreTrainedTokenizer, | |
PreTrainedTokenizerFast, | |
) | |
from torch import nn | |
from e6db_reader import TagSetNormalizer, tag_category2id, tag_rank_to_freq | |
import logging | |
import gettext | |
import locale | |
from caption_utils import ( | |
CaptionValidator, | |
setup_logging, | |
STOP_WORDS, | |
ColoredFormatter, | |
calculate_entropy | |
) | |
try: | |
import pillow_avif | |
except ImportError: | |
pass | |
# Set up localization | |
try: | |
# Add the directory containing the locale files to the path | |
localedir = os.path.join(os.path.dirname(__file__), 'locales') | |
locale.setlocale(locale.LC_ALL, '') | |
except locale.Error: | |
locale.setlocale(locale.LC_ALL, 'C') | |
# Initialize gettext with the correct path | |
gettext.bindtextdomain('joy', localedir) | |
gettext.textdomain('joy') | |
_ = gettext.gettext | |
CLIP_PATH = "google/siglip-so400m-patch14-384" | |
MODEL_PATH = "meta-llama/Meta-Llama-3.1-8B" | |
CHECKPOINT_PATH = Path(__file__).resolve().parent / "cgrkzexw-599808" | |
CAPTION_TYPE_MAP = { | |
"descriptive": [ | |
"Write a descriptive caption for this image in a formal tone.", | |
"Write a descriptive caption for this image in a formal tone within {word_count} words.", | |
"Write a {length} descriptive caption for this image in a formal tone.", | |
], | |
"descriptive (informal)": [ | |
"Write a descriptive caption for this image in a casual tone.", | |
"Write a descriptive caption for this image in a casual tone within {word_count} words.", | |
"Write a {length} descriptive caption for this image in a casual tone.", | |
], | |
"training prompt": [ | |
"Write a stable diffusion prompt for this image.", | |
"Write a stable diffusion prompt for this image within {word_count} words.", | |
"Write a {length} stable diffusion prompt for this image.", | |
], | |
"midjourney": [ | |
"Write a MidJourney prompt for this image.", | |
"Write a MidJourney prompt for this image within {word_count} words.", | |
"Write a {length} MidJourney prompt for this image.", | |
], | |
"booru tag list": [ | |
"Write a list of Booru tags for this image.", | |
"Write a list of Booru tags for this image within {word_count} words.", | |
"Write a {length} list of Booru tags for this image.", | |
], | |
"booru-like tag list": [ | |
"Write a list of Booru-like tags for this image.", | |
"Write a list of Booru-like tags for this image within {word_count} words.", | |
"Write a {length} list of Booru-like tags for this image.", | |
], | |
"art critic": [ | |
"Analyze this image like an art critic would with information about its composition, style, symbolism, the use of color, light, any artistic movement it might belong to, etc.", | |
"Analyze this image like an art critic would with information about its composition, style, symbolism, the use of color, light, any artistic movement it might belong to, etc. Keep it within {word_count} words.", | |
"Analyze this image like an art critic would with information about its composition, style, symbolism, the use of color, light, any artistic movement it might belong to, etc. Keep it {length}.", | |
], | |
"product listing": [ | |
"Write a caption for this image as though it were a product listing.", | |
"Write a caption for this image as though it were a product listing. Keep it under {word_count} words.", | |
"Write a {length} caption for this image as though it were a product listing.", | |
], | |
"social media post": [ | |
"Write a caption for this image as if it were being used for a social media post.", | |
"Write a caption for this image as if it were being used for a social media post. Limit the caption to {word_count} words.", | |
"Write a {length} caption for this image as if it were being used for a social media post.", | |
], | |
} | |
HF_TOKEN = os.environ.get("HF_TOKEN", None) | |
class ImageAdapter(nn.Module): | |
""" | |
Custom image adapter module for processing CLIP vision outputs. | |
This module adapts the output of a CLIP vision model to be compatible with | |
a text model. It supports optional layer normalization, positional | |
embeddings, and deep feature extraction. | |
Args: | |
input_features (int): | |
Number of input features from the vision model. | |
output_features (int): | |
Number of output features to match the text model. | |
ln1 (bool): | |
Whether to use layer normalization. | |
pos_emb (bool): | |
Whether to use positional embeddings. | |
num_image_tokens (int): | |
Number of image tokens. | |
deep_extract (bool): | |
Whether to use deep feature extraction. | |
""" | |
def __init__( | |
self, | |
input_features: int, | |
output_features: int, | |
ln1: bool, | |
pos_emb: bool, | |
num_image_tokens: int, | |
deep_extract: bool, | |
): | |
super().__init__() | |
self.deep_extract = deep_extract | |
if self.deep_extract: | |
input_features = input_features * 5 | |
self.linear1 = nn.Linear(input_features, output_features) | |
self.activation = nn.GELU() | |
self.linear2 = nn.Linear(output_features, output_features) | |
self.ln1 = nn.Identity() if not ln1 else nn.LayerNorm(input_features) | |
self.pos_emb = ( | |
None | |
if not pos_emb | |
else nn.Parameter(torch.zeros(num_image_tokens, input_features)) | |
) | |
self.other_tokens = nn.Embedding(3, output_features) | |
self.other_tokens.weight.data.normal_(mean=0.0, std=0.02) | |
def forward(self, vision_outputs: torch.Tensor): | |
""" | |
Forward pass of the image adapter. | |
Args: | |
vision_outputs (torch.Tensor): | |
Output tensor from the CLIP vision model. | |
Returns: | |
torch.Tensor: Adapted image features. | |
""" | |
if self.deep_extract: | |
x = torch.concat( | |
( | |
vision_outputs[-2], | |
vision_outputs[3], | |
vision_outputs[7], | |
vision_outputs[13], | |
vision_outputs[20], | |
), | |
dim=-1, | |
) | |
assert len(x.shape) == 3, f"Expected 3, got {len(x.shape)}" | |
expected_shape = vision_outputs[-2].shape[-1] * 5 | |
assert ( | |
x.shape[-1] == expected_shape | |
), f"Expected {expected_shape}, got {x.shape[-1]}" | |
else: | |
x = vision_outputs[-2] | |
x = self.ln1(x) | |
if self.pos_emb is not None: | |
assert ( | |
x.shape[-2:] == self.pos_emb.shape | |
), f"Expected {self.pos_emb.shape}, got {x.shape[-2:]}" | |
x = x + self.pos_emb | |
x = self.linear1(x) | |
x = self.activation(x) | |
x = self.linear2(x) | |
other_tokens = self.other_tokens( | |
torch.tensor([0, 1], device=self.other_tokens.weight.device).expand( | |
x.shape[0], -1 | |
) | |
) | |
assert other_tokens.shape == ( | |
x.shape[0], | |
2, | |
x.shape[2], | |
), f"Expected {(x.shape[0], 2, x.shape[2])}, got {other_tokens.shape}" | |
x = torch.cat((other_tokens[:, 0:1], x, other_tokens[:, 1:2]), dim=1) | |
return x | |
def get_eot_embedding(self): | |
""" | |
Get the end-of-text embedding. | |
Returns: | |
torch.Tensor: The end-of-text embedding. | |
""" | |
return self.other_tokens( | |
torch.tensor([2], device=self.other_tokens.weight.device) | |
).squeeze(0) | |
class JoyCaptionModel: | |
""" | |
A class for generating captions for images using CLIP, LLM, | |
and custom image adapters. | |
This class encapsulates the functionality to load and initialize | |
various models (CLIP, LLM, image adapter) and use them to process | |
images and generate captions. | |
It supports different caption types, tones, and lengths. | |
Attributes: | |
clip_model: The CLIP vision model for processing images. | |
text_model: The language model for generating captions. | |
image_adapter: Custom adapter for processing CLIP vision outputs. | |
tokenizer: Tokenizer for the language model. | |
Methods: | |
load_models(): Load and initialize all required models. | |
process_image(input_image, caption_type, caption_length): | |
Process an input image and generate a caption | |
based on specified parameters. | |
""" | |
def __init__(self): | |
self.clip_model = None | |
self.text_model = None | |
self.image_adapter = None | |
self.tokenizer = None | |
def load_models(self): | |
""" | |
Load and initialize all required models (CLIP, LLM, image adapter). | |
""" | |
logging.info(_("Loading CLIP")) | |
self.clip_model = AutoModel.from_pretrained(CLIP_PATH) | |
self.clip_model = self.clip_model.vision_model | |
if (CHECKPOINT_PATH / "clip_model.pt").exists(): | |
logging.info(_("Loading VLM's custom vision model")) | |
checkpoint = torch.load( | |
CHECKPOINT_PATH / "clip_model.pt", map_location="cpu" | |
) | |
checkpoint = { | |
k.replace("_orig_mod.module.", ""): v for k, v in checkpoint.items() | |
} | |
self.clip_model.load_state_dict(checkpoint) | |
del checkpoint | |
self.clip_model.eval() | |
self.clip_model.requires_grad_(False) | |
self.clip_model.to("cuda") | |
logging.info(_("Loading tokenizer")) | |
self.tokenizer = AutoTokenizer.from_pretrained( | |
CHECKPOINT_PATH / "text_model", use_fast=True | |
) | |
assert isinstance( | |
self.tokenizer, (PreTrainedTokenizer, PreTrainedTokenizerFast) | |
) | |
logging.info(_("Loading LLM")) | |
if (CHECKPOINT_PATH / "text_model").exists(): | |
logging.info(_("Loading VLM's custom text model")) | |
self.text_model = AutoModelForCausalLM.from_pretrained( | |
CHECKPOINT_PATH / "text_model", device_map=0, torch_dtype=torch.bfloat16 | |
) | |
else: | |
self.text_model = AutoModelForCausalLM.from_pretrained( | |
MODEL_PATH, device_map="auto", torch_dtype=torch.bfloat16 | |
) | |
self.text_model.eval() | |
logging.info(_("Loading image adapter")) | |
self.image_adapter = ImageAdapter( | |
self.clip_model.config.hidden_size, | |
self.text_model.config.hidden_size, | |
False, | |
False, | |
38, | |
False, | |
) | |
self.image_adapter.load_state_dict( | |
torch.load(CHECKPOINT_PATH / "image_adapter.pt", map_location="cpu") | |
) | |
self.image_adapter.eval() | |
self.image_adapter.to("cuda") | |
@torch.no_grad() | |
def process_image( | |
self, | |
input_image: Image.Image, | |
prompt_str: str, | |
) -> Tuple[str, float]: | |
""" | |
Process an input image and generate a caption based on specified parameters. | |
Also calculates the entropy of the generated caption. | |
Returns: | |
Tuple[str, float]: The generated caption and its entropy. | |
""" | |
torch.cuda.empty_cache() | |
pixel_values = self._preprocess_image(input_image) | |
embedded_images = self._embed_image(pixel_values) | |
inputs_embeds, input_ids, attention_mask = self._construct_inputs( | |
embedded_images, prompt_str | |
) | |
generate_ids = self._generate_caption(inputs_embeds, input_ids, attention_mask) | |
caption = self._decode_caption(generate_ids, input_ids) | |
# Calculate entropy using the shared function | |
token_ids = generate_ids[0].tolist() | |
entropy = calculate_entropy(token_ids) | |
return caption.strip(), entropy | |
def generate_valid_caption( | |
self, | |
input_image: Image.Image, | |
prompt: str, | |
*, | |
limited_words: Dict[str, int] = {"fluffy": 2}, | |
min_sentence_count: int = 3, | |
max_word_repetitions: int = 5, | |
min_entropy: float = 1.75, | |
stop_words: set[str] = STOP_WORDS, | |
) -> str: | |
""" | |
Generate a valid caption, retrying if certain conditions are not met. | |
Args: | |
input_image (Image.Image): The input image to caption. | |
prompt (str): Prompt for caption generation. | |
limited_words (Dict[str, int]): Dictionary of words with their maximum allowed occurrences. | |
min_sentence_count (int): Minimum required number of sentences. | |
max_word_repetitions (int): Maximum allowed repetitions for words longer than 4 characters. | |
min_entropy (float): Minimum required entropy of the caption. | |
stop_words (set[str]): Set of words to ignore in word count calculations. | |
Returns: | |
str: A valid caption meeting all specified criteria. | |
""" | |
while True: | |
caption, entropy = self.process_image(input_image, prompt) | |
# Add entropy as an attribute for validation | |
setattr(caption, "entropy", entropy) | |
is_valid, error_msg = CaptionValidator.validate_caption( | |
caption, | |
limited_words=limited_words, | |
min_sentence_count=min_sentence_count, | |
max_word_repetitions=max_word_repetitions, | |
min_entropy=min_entropy, | |
stop_words=stop_words, | |
) | |
if is_valid: | |
return caption | |
else: | |
logging.info(_(error_msg)) | |
@staticmethod | |
def get_prompt_string(caption_type, caption_length): | |
length = None if caption_length == "any" else caption_length | |
if isinstance(length, str): | |
try: | |
length = int(length) | |
except ValueError: | |
pass | |
# Build prompt | |
if length is None: | |
map_idx = 0 | |
elif isinstance(length, int): | |
map_idx = 1 | |
elif isinstance(length, str): | |
map_idx = 2 | |
else: | |
raise ValueError(f"Invalid caption length: {length}") | |
caption_type = caption_type.lower() | |
if caption_type not in CAPTION_TYPE_MAP: | |
raise ValueError(f"Invalid caption type: {caption_type}") | |
prompt_str = CAPTION_TYPE_MAP[caption_type][map_idx] | |
prompt_str = prompt_str.format(length=caption_length, word_count=caption_length) | |
return prompt_str | |
@staticmethod | |
def _preprocess_image(input_image: Image.Image) -> torch.Tensor: | |
""" | |
Preprocess the input image for the CLIP model. | |
Args: | |
input_image (Image.Image): The input PIL image. | |
Returns: | |
torch.Tensor: Preprocessed image tensor. | |
""" | |
image = input_image.resize((384, 384), Image.LANCZOS) | |
pixel_values = TVF.pil_to_tensor(image).unsqueeze(0) / 255.0 | |
pixel_values = TVF.normalize(pixel_values, [0.5], [0.5]) | |
return pixel_values.to("cuda") | |
def _embed_image(self, pixel_values: torch.Tensor) -> torch.Tensor: | |
""" | |
Embed the preprocessed image using CLIP and the image adapter. | |
Args: | |
pixel_values (torch.Tensor): Preprocessed image tensor. | |
Returns: | |
torch.Tensor: Embedded image tensor. | |
""" | |
with torch.amp.autocast_mode.autocast("cuda", enabled=True): | |
vision_outputs = self.clip_model( | |
pixel_values=pixel_values, output_hidden_states=True | |
) | |
embedded_images = self.image_adapter(vision_outputs.hidden_states) | |
return embedded_images.to("cuda") | |
def _construct_inputs( | |
self, embedded_images: torch.Tensor, prompt_str: str | |
) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor]: | |
""" | |
Construct the inputs for the text model's generate method. | |
Args: | |
embedded_images (torch.Tensor): Embedded image tensor. | |
prompt_str (str): The prompt string for captioning. | |
Returns: | |
tuple: (input_embeds, input_ids, attention_mask) | |
""" | |
# Build the conversation | |
convo = [ | |
{"role": "system", "content": "You are a helpful image captioner."}, | |
{"role": "user", "content": prompt_str}, | |
] | |
# Format and tokenize the conversation | |
convo_string = self.tokenizer.apply_chat_template( | |
convo, tokenize=False, add_generation_prompt=True | |
) | |
logging.debug(f"Convo:\n{convo_string}") | |
convo_tokens = self.tokenizer.encode( | |
convo_string, | |
return_tensors="pt", | |
add_special_tokens=False, | |
truncation=False, | |
) | |
prompt_tokens = self.tokenizer.encode( | |
prompt_str, return_tensors="pt", add_special_tokens=False, truncation=False | |
) | |
convo_tokens = convo_tokens.squeeze(0) | |
prompt_tokens = prompt_tokens.squeeze(0) | |
# Calculate where to inject the image | |
eot_id_indices = ( | |
(convo_tokens == self.tokenizer.convert_tokens_to_ids("<|eot_id|>")) | |
.nonzero(as_tuple=True)[0] | |
.tolist() | |
) | |
preamble_len = eot_id_indices[1] - prompt_tokens.shape[0] | |
# Embed the tokens | |
convo_embeds = self.text_model.model.embed_tokens( | |
convo_tokens.unsqueeze(0).to("cuda") | |
) | |
# Construct the input | |
input_embeds = torch.cat( | |
[ | |
convo_embeds[:, :preamble_len], | |
embedded_images.to(dtype=convo_embeds.dtype), | |
convo_embeds[:, preamble_len:], | |
], | |
dim=1, | |
).to("cuda") | |
input_ids = torch.cat( | |
[ | |
convo_tokens[:preamble_len].unsqueeze(0), | |
torch.zeros((1, embedded_images.shape[1]), dtype=torch.long), | |
convo_tokens[preamble_len:].unsqueeze(0), | |
], | |
dim=1, | |
).to("cuda") | |
attention_mask = torch.ones_like(input_ids) | |
return input_embeds, input_ids, attention_mask | |
def _generate_caption(self, inputs_embeds, input_ids, attention_mask): | |
generate_ids = self.text_model.generate( | |
input_ids, | |
inputs_embeds=inputs_embeds, | |
attention_mask=attention_mask, | |
max_new_tokens=300, | |
# max_length=4096, | |
do_sample=True, | |
suppress_tokens=None, | |
repetition_penalty=1.2, | |
) | |
return generate_ids | |
def _decode_caption(self, generate_ids, input_ids): | |
generate_ids = generate_ids[:, input_ids.shape[1] :] | |
if generate_ids[0][-1] == self.tokenizer.eos_token_id or generate_ids[0][ | |
-1 | |
] == self.tokenizer.convert_tokens_to_ids("<|eot_id|>"): | |
generate_ids = generate_ids[:, :-1] | |
caption = self.tokenizer.batch_decode( | |
generate_ids, skip_special_tokens=False, clean_up_tokenization_spaces=False | |
)[0] | |
return caption | |
def main(): | |
""" | |
Generate captions for images in a directory | |
and save them as .caption files. | |
""" | |
parser = argparse.ArgumentParser( | |
description=_("Generate captions for images in a directory and save them as .caption files.") | |
) | |
parser.add_argument( | |
"directory", type=str, help=_("Target directory containing images.") | |
) | |
parser.add_argument( | |
"--caption_type", | |
type=str, | |
default="descriptive", | |
choices=CAPTION_TYPE_MAP.keys(), | |
help=_("Type of caption to generate."), | |
) | |
parser.add_argument( | |
"--caption_length", type=str, default="any", help=_("Length of the caption.") | |
) | |
parser.add_argument( | |
"--dont-strip-commas", | |
action="store_true", | |
help=_("If set, commas will not be stripped from the generated captions."), | |
) | |
parser.add_argument( | |
"--custom_prompt", | |
type=str, | |
help=_("Custom prompt for the captioner. Use with --caption_type custom."), | |
) | |
parser.add_argument( | |
"--add-commas-to-sentence-ends", | |
action="store_true", | |
help=_("Add commas after periods in sentences"), | |
) | |
parser.add_argument( | |
"--feed-from-tags", | |
type=int, | |
nargs="?", | |
const=-1, | |
help=_( | |
"Use .tags files with the same base filename as the images as input to the captioner. Optionally specify the number of tags to use." | |
), | |
) | |
parser.add_argument( | |
"--artist-from-folder", | |
action="store_true", | |
help=_("Get the artist name from the parent folder"), | |
) | |
parser.add_argument( | |
"--random-tags", | |
type=int, | |
help=_( | |
"Randomly select n number of tags. Only works if --feed-from-tags is enabled." | |
), | |
) | |
parser.add_argument( | |
"--dry-run", | |
action="store_true", | |
help=_("Run in dry-run mode without loading models or generating captions."), | |
) | |
parser.add_argument( | |
"-v", | |
"--verbose", | |
action="count", | |
default=0, | |
help=_("Increase output verbosity (can be repeated)"), | |
) | |
args = parser.parse_args() | |
setup_logging(args.verbose) | |
tasks = [] | |
image_extensions = {".webp", ".png", ".jpeg", ".jpg", ".jxl", ".avif"} | |
for image_path in Path(args.directory).rglob("*"): | |
# Skip analysis_output directories | |
if "analysis_output" in str(image_path.parent).split(os.sep): | |
continue | |
if image_path.suffix.lower() in image_extensions: | |
caption_file = image_path.with_suffix(".caption") | |
# Skip if the caption file already exists | |
if caption_file.exists(): | |
logging.info( | |
_("Skipping {}: Caption file already exists.").format(image_path) | |
) | |
continue | |
tasks.append((image_path, caption_file)) | |
if not tasks: | |
logging.error("No input file found.") | |
return | |
# Validate random-tags usage | |
if args.random_tags is not None and args.feed_from_tags is None: | |
parser.error(_("--random-tags can only be used when --feed-from-tags is enabled")) | |
if args.feed_from_tags is not None and args.artist_from_folder: | |
raise ValueError("feed-from-tags and artist-from-folder can't be used together") | |
if args.feed_from_tags is not None: | |
logging.info(_("Loading e621 tag data")) | |
tagset_normalizer = make_tagset_normalizer() | |
# Initialize and load models only if not in dry-run mode | |
if not args.dry_run: | |
joy_caption_model = JoyCaptionModel() | |
joy_caption_model.load_models() | |
else: | |
logging.info(_("Running in dry-run mode. Models will not be loaded.")) | |
for image_path, caption_file in tasks: | |
if not args.dry_run: | |
input_image = Image.open(image_path).convert("RGB") | |
# Use custom prompt if specified | |
prompt = args.custom_prompt or JoyCaptionModel.get_prompt_string( | |
args.caption_type, args.caption_length | |
) | |
if args.feed_from_tags is not None: | |
prompt = prompt_from_tags(args, image_path, tagset_normalizer, prompt) | |
elif args.artist_from_folder: | |
prompt = prompt_from_folder(prompt, image_path.resolve()) | |
if args.dry_run: | |
logging.info( | |
_(f"Dry run: Skipping caption generation for {image_path} with prompt:\n\t{prompt}") | |
) | |
continue | |
else: | |
logging.info( | |
_("Prompt for {}:\n\t{}").format(image_path, prompt) | |
) | |
caption = joy_caption_model.generate_valid_caption(input_image, prompt) | |
caption = CaptionValidator.clean_caption( | |
caption, | |
strip_commas=not args.dont_strip_commas, | |
add_commas_to_sentence_ends=args.add_commas_to_sentence_ends | |
) | |
logging.info( | |
_("Caption for {}:\n\t{}{}{}").format( | |
image_path, | |
ColoredFormatter.COLOR_GREEN, | |
caption, | |
ColoredFormatter.RESET | |
) | |
) | |
# Save the caption to a .caption file | |
with open(caption_file, "w", encoding="utf-8") as f: | |
f.write(caption) | |
logging.info( | |
_(f"Caption saved to {caption_file}") | |
) | |
RE_PARENS_SUFFIX = re.compile(r"_\([^)]+\)$") | |
E6DB_DATA = Path(__file__).resolve().parent / "data" | |
def make_tagset_normalizer(): | |
""" | |
Create a TagSetNormalizer for encoding/decoding tags to and from integers. | |
Configures it based on the provided config. | |
""" | |
# This loads all the aliases and implications | |
tagset_normalizer = TagSetNormalizer(E6DB_DATA) | |
tagid2cat = tagset_normalizer.tag_normalizer.tag_categories | |
cat_artist = tag_category2id["artist"] | |
cat2suffix = { | |
tag_category2id["character"]: "_(character)", | |
tag_category2id["lore"]: "_(lore)", | |
tag_category2id["species"]: "_(species)", | |
tag_category2id["copyright"]: "_(copyright)", | |
} | |
# Create additional aliases for tags using simple rules | |
def input_map(tag, tid): | |
# Make an alias without parentheses, it might conflict but we'll handle | |
# it depending on `on_alias_conflict` config value. | |
without_suffix = RE_PARENS_SUFFIX.sub("", tag) | |
had_suffix = tag != without_suffix | |
if had_suffix: | |
yield without_suffix | |
# Add an alias with the suffix (special case for artist) | |
cat = tagid2cat[tid] if tid is not None else -1 | |
if cat == cat_artist: | |
artist = without_suffix.removeprefix("by_") | |
if artist != without_suffix: | |
yield artist | |
if not had_suffix: | |
yield f"{artist}_(artist)" | |
else: | |
yield f"by_{artist}" | |
if not had_suffix: | |
yield f"by_{artist}_(artist)" | |
elif not had_suffix: | |
suffix = cat2suffix.get(cat) | |
if suffix is not None: | |
yield f"{without_suffix}{suffix}" | |
# Recognize tags where ':' were replaced by a space (aspect ratio) | |
if ":" in tag: | |
yield tag.replace(":", "_") | |
return tagset_normalizer.map_inputs(input_map, on_conflict="ignore") | |
def format_nl_list(word_list): | |
""" | |
Takes a list of words and generates a natural language output. | |
""" | |
n = len(word_list) | |
assert n > 0 | |
if n == 1: | |
return word_list[0] | |
if n == 2: | |
return f"{word_list[0]} and {word_list[1]}" | |
# n > 2 | |
*head, last = word_list | |
return ", ".join(head) + ", and " + last | |
TAG_SPECIES = tag_category2id["species"] | |
TAG_CHARACTER = tag_category2id["character"] | |
TAG_ARTIST = tag_category2id["artist"] | |
TAG_COPYRIGHT = tag_category2id["copyright"] | |
TAG_META = tag_category2id["meta"] | |
def prompt_from_tags( | |
args, | |
image_path: Path, | |
tagset_normalizer: TagSetNormalizer, | |
base_prompt: str = "Write a descriptive caption for this image in a formal tone.", | |
tag_freq_threshold: int = 0, | |
tag_string_prefix: str = "Use these words to construct your caption:", | |
): | |
""" | |
Generates a prompt from tags associated with the given image. | |
Args: | |
args: Additional arguments for the function. | |
image_path (Path): | |
The path to the image file. | |
tagset_normalizer (TagSetNormalizer): | |
An instance to normalize the tag set. | |
""" | |
# Find and read the corresponding tag file | |
tag_file = find_tag_file(image_path) | |
if tag_file is None: | |
logging.warning( | |
_(f"No tag file found for {image_path}") | |
) | |
return base_prompt | |
with open(tag_file, "r", encoding="utf-8") as f: | |
tags = f.read().lower().split(",") | |
# Get helper functions from the tagset_normalizer | |
tag_id_to_cat_id = tagset_normalizer.tag_normalizer.tag_categories | |
encode = tagset_normalizer.tag_normalizer.encode | |
# Initialize dictionaries and lists to store categorized tags | |
# These lists will contain tuples (freq, tag, tag_id) | |
tag_by_category: Dict[int, List[Tuple[int, str, int]]] = { | |
cat: [] for cat in [TAG_ARTIST, TAG_CHARACTER, TAG_COPYRIGHT, TAG_SPECIES] | |
} | |
other_tags: List[Tuple[int, str, int]] = [] | |
implied: set = set() | |
# Process each tag | |
for tag in tags: | |
tag = tag.strip() | |
# Encode the tag into a numerical id | |
tag_id = encode(tag.replace(" ", "_")) | |
if tag_id is None: | |
# If tag is not recognized, add it to other_tags | |
other_tags.append((0, tag, 0)) | |
implied.update(tagset_normalizer.implications_rej.get(0, ())) | |
continue | |
# Get the category of the tag | |
cat_id = tag_id_to_cat_id[tag_id] | |
# Skip meta tags | |
if cat_id == TAG_META: | |
continue | |
# Update implied tags | |
implied.update(tagset_normalizer.implications.get(tag_id, ())) | |
# Get the frequency of the tag | |
freq = tag_rank_to_freq(tag_id) | |
if freq < tag_freq_threshold: | |
continue | |
# Add the tag to its category, or other_tags | |
tag_by_category.get(cat_id, other_tags).append((int(freq), tag, tag_id)) | |
# Sort other_tags by frequency (descending) and filter out implied tags | |
other_tags = sorted( | |
(-freq, tag, tag_id) | |
for freq, tag, tag_id in other_tags | |
if tag_id not in implied | |
) | |
# Sort tags within each category, prefering non implied tags | |
for cat_id, cat_list in tag_by_category.items(): | |
tag_by_category[cat_id] = sorted( | |
((tag_id in implied, -freq), tag, tag_id) for freq, tag, tag_id in cat_list | |
) | |
# Handle random tag selection or tag limit if specified | |
if args.random_tags is not None: | |
# Randomly select tags if --random-tags is specified | |
num_tags = min(args.random_tags, len(other_tags)) | |
other_tags = random.sample( | |
[ | |
(i, tag, 0) | |
for i, tag in enumerate(tags[: round(args.random_tags * 1.5)]) | |
], | |
num_tags, | |
) | |
elif args.feed_from_tags > 0: | |
# Use specified number of tags if --feed-from-tags has a positive value | |
other_tags = other_tags[: args.feed_from_tags] | |
# Prepare sentence pieces for each category | |
artist_tag = tag_by_category[TAG_ARTIST] | |
if artist_tag: | |
artist_list = [str(tp[1]).removeprefix("by ") for tp in artist_tag[:4]] | |
artist_txt = f"by {format_nl_list(artist_list)}" | |
else: | |
artist_txt = "" | |
character_tag = tag_by_category[TAG_CHARACTER] | |
if character_tag: | |
tags = [tag for _, tag, _ in character_tag[:4]] | |
character_txt = f"named {format_nl_list(tags)}" | |
else: | |
character_txt = "" | |
species_tag = tag_by_category[TAG_SPECIES] | |
if species_tag: | |
species_txt = ( | |
"of a " if len(character_tag) <= 1 and len(species_tag) <= 1 else "of " | |
) | |
species_txt += format_nl_list([tp[1] for tp in species_tag[:4]]) | |
else: | |
if character_tag: | |
species_txt = ( | |
"of a character" if len(character_tag) <= 1 else "of characters" | |
) | |
else: | |
species_txt = "" | |
copyright_tag = tag_by_category[TAG_COPYRIGHT] | |
if copyright_tag: | |
tags = [tag for _, tag, *_ in copyright_tag[:4]] | |
copyright_txt = f"from {format_nl_list(tags)}" | |
else: | |
copyright_txt = "" | |
# Prepare the remaining tags as a string | |
tag_string = ", ".join(tp[1] for tp in other_tags) | |
# Extract the prefix and suffix around the word "image" from the base prompt | |
image_pos = base_prompt.find("image") | |
if image_pos < 0: | |
raise ValueError("Base prompt must contain the word 'image'") | |
image_pos += len("image") | |
base_prompt_prefix = base_prompt[:image_pos].rstrip() | |
base_prompt_suffix = base_prompt[image_pos:].lstrip() | |
pieces = [ | |
base_prompt_prefix, | |
artist_txt, | |
species_txt, | |
character_txt, | |
copyright_txt, | |
base_prompt_suffix, | |
tag_string_prefix, | |
tag_string, | |
".", | |
] | |
logging.debug("Prompt pieces: %r", pieces) | |
custom_prompt = " ".join(p for p in pieces if p) | |
custom_prompt = custom_prompt.replace(" .", ".").replace(" ,", ",") | |
return custom_prompt | |
def find_tag_file(image_path): | |
""" | |
Find the corresponding .tags file for the given image path. | |
Handles cases where the image has a -(number) suffix. | |
""" | |
base_name = image_path.stem | |
tag_file = image_path.with_suffix(".tags") | |
if tag_file.exists(): | |
return tag_file | |
# Handle -(number) suffix | |
match = re.match(r"(.+)-\d+$", base_name) | |
if match: | |
base_name = match.group(1) | |
tag_file = image_path.with_name(base_name).with_suffix(".tags") | |
if tag_file.exists(): | |
return tag_file | |
return None | |
def prompt_from_folder(prompt, path): | |
artist = ( | |
path.parent.name.replace("_", " ").replace("-", " ").replace(".", " ").title() | |
) | |
return prompt.replace("image", f"image by {artist}") | |
if __name__ == "__main__": | |
main() |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment