mirror of
https://github.com/turboderp-org/exllamav2.git
synced 2026-03-15 00:07:26 +00:00
208 lines
6.4 KiB
Python
208 lines
6.4 KiB
Python
import sys, os
|
|
sys.path.append(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
|
|
|
|
from exllamav2 import (
|
|
ExLlamaV2,
|
|
ExLlamaV2Config,
|
|
ExLlamaV2Cache,
|
|
ExLlamaV2Tokenizer,
|
|
ExLlamaV2VisionTower,
|
|
)
|
|
|
|
from exllamav2.generator import (
|
|
ExLlamaV2DynamicGenerator,
|
|
ExLlamaV2DynamicJob,
|
|
ExLlamaV2Sampler,
|
|
)
|
|
|
|
from PIL import Image
|
|
import requests
|
|
|
|
import torch
|
|
torch.set_printoptions(precision = 5, sci_mode = False, linewidth=200)
|
|
|
|
# Models used:
|
|
#
|
|
# Pixtral:
|
|
# https://huggingface.co/mistral-community/pixtral-12b/
|
|
# https://huggingface.co/turboderp/pixtral-12b-exl2
|
|
# Mistral-Small 3.1:
|
|
# https://huggingface.co/prince-canuma/Mistral-Small-3.1-24B-Instruct-2503
|
|
# Qwen2-VL:
|
|
# https://huggingface.co/Qwen/Qwen2-VL-7B-Instruct
|
|
# https://huggingface.co/turboderp/Qwen2-VL-7B-Instruct-exl2
|
|
# Gemma3:
|
|
# https://huggingface.co/google/gemma-3-27b-it
|
|
# https://huggingface.co/turboderp/gemma-3-27b-it-exl2
|
|
|
|
# mode = "pixtral"
|
|
mode = "mistral3"
|
|
# mode = "qwen2"
|
|
# mode = "gemma3"
|
|
|
|
streaming = True
|
|
greedy = True
|
|
|
|
if mode == "pixtral":
|
|
model_directory = "/mnt/str/models/pixtral-12b-exl2/6.0bpw"
|
|
elif mode == "qwen2":
|
|
model_directory = "/mnt/str/models/qwen2.5-vl-7b-instruct-exl2/5.0bpw"
|
|
elif mode == "gemma3":
|
|
model_directory = "/mnt/str/models/gemma3-12b-it-exl2/6.0bpw"
|
|
elif mode == "mistral3":
|
|
model_directory = "/mnt/str/models/mistral-small-3.1-24b-instruct/exl2/4.5bpw"
|
|
|
|
images = [
|
|
# {"file": "media/test_image_1.jpg"},
|
|
# {"file": "media/test_image_2.jpg"},
|
|
{"url": "https://media.istockphoto.com/id/1212540739/photo/mom-cat-with-kitten.jpg?s=612x612&w=0&k=20&c=RwoWm5-6iY0np7FuKWn8FTSieWxIoO917FF47LfcBKE="},
|
|
# {"url": "https://i.dailymail.co.uk/1s/2023/07/10/21/73050285-12283411-Which_way_should_I_go_One_lady_from_the_US_shared_this_incredibl-a-4_1689019614007.jpg"},
|
|
# {"url": "https://images.fineartamerica.com/images-medium-large-5/metal-household-objects-trevor-clifford-photography.jpg"}
|
|
]
|
|
|
|
# instruction = "Compare and contrast the two experiments."
|
|
instruction = "Describe the image."
|
|
# instruction = "Find the alarm clock." # Qwen2 seems to support this but unsure of how to prompt correctly
|
|
|
|
# Initialize model
|
|
|
|
config = ExLlamaV2Config(model_directory)
|
|
config.max_seq_len = 8192 # Pixtral default is 1M
|
|
|
|
# Load vision model and multimodal projector and initialize preprocessor
|
|
|
|
vision_model = ExLlamaV2VisionTower(config)
|
|
vision_model.load(progress = True)
|
|
|
|
# Load EXL2 model
|
|
|
|
model = ExLlamaV2(config)
|
|
cache = ExLlamaV2Cache(model, max_seq_len = 8192, lazy = True)
|
|
model.load_autosplit(progress = True, cache = cache)
|
|
tokenizer = ExLlamaV2Tokenizer(config)
|
|
|
|
# Create generator
|
|
|
|
generator = ExLlamaV2DynamicGenerator(
|
|
model = model,
|
|
cache = cache,
|
|
tokenizer = tokenizer,
|
|
)
|
|
|
|
# Util function to get a PIL image from a URL or from a file in the script's directory
|
|
|
|
def get_image(file = None, url = None):
|
|
assert (file or url) and not (file and url)
|
|
if file:
|
|
script_dir = os.path.dirname(os.path.abspath(__file__))
|
|
file_path = os.path.join(script_dir, file)
|
|
return Image.open(file_path)
|
|
elif url:
|
|
return Image.open(requests.get(url, stream = True).raw)
|
|
|
|
# Convert image(s) to embeddings. Aliases can be given explicitly with the text_alias argument, but here we
|
|
# use automatically assigned unique identifiers, then concatenate them into a string
|
|
|
|
image_embeddings = [
|
|
vision_model.get_image_embeddings(
|
|
model = model,
|
|
tokenizer = tokenizer,
|
|
image = get_image(**img_args),
|
|
)
|
|
for img_args in images
|
|
]
|
|
|
|
placeholders = "\n".join([ie.text_alias for ie in image_embeddings]) + "\n"
|
|
|
|
# Define a prompt using the aliases above as placeholders for image tokens. The tokenizer will replace each alias
|
|
# with a range of temporary token IDs, and the model will embed those temporary IDs from their respective sources
|
|
# rather than the model's text embedding table.
|
|
#
|
|
# The temporary IDs are unique for the lifetime of the process and persist as long as a reference is held to the
|
|
# corresponding ExLlamaV2Embedding object. This way, images can be reused between generations, or used multiple
|
|
# for multiple jobs in a batch, and the generator will be able to apply prompt caching and deduplication to image
|
|
# tokens as well as text tokens.
|
|
#
|
|
# Image token IDs are assigned sequentially, however, so two ExLlamaV2Embedding objects created from the same
|
|
# source image will not be recognized as the same image for purposes of prompt caching etc.
|
|
|
|
if mode in ["pixtral", "mistral3"]:
|
|
prompt = (
|
|
"[INST]" +
|
|
placeholders +
|
|
instruction +
|
|
"[/INST]"
|
|
)
|
|
stop_conditions = [tokenizer.eos_token_id]
|
|
|
|
elif mode == "qwen2":
|
|
prompt = (
|
|
"<|im_start|>system\n" +
|
|
"You are a helpful assistant.<|im_end|>\n" +
|
|
"<|im_start|>user\n" +
|
|
placeholders +
|
|
instruction +
|
|
"<|im_end|>\n" +
|
|
"<|im_start|>assistant\n"
|
|
)
|
|
stop_conditions = [tokenizer.eos_token_id]
|
|
|
|
elif mode == "gemma3":
|
|
prompt = (
|
|
"<start_of_turn>user\nYou are a helpful assistant.\n\n\n\n" +
|
|
placeholders +
|
|
"\n" +
|
|
instruction +
|
|
"<end_of_turn>\n" +
|
|
"<start_of_turn>model\n"
|
|
)
|
|
stop_conditions = [tokenizer.single_id("<end_of_turn>")]
|
|
|
|
# Generate
|
|
|
|
if streaming:
|
|
|
|
input_ids = tokenizer.encode(
|
|
prompt,
|
|
add_bos = True,
|
|
encode_special_tokens = True,
|
|
embeddings = image_embeddings,
|
|
)
|
|
|
|
job = ExLlamaV2DynamicJob(
|
|
input_ids = input_ids,
|
|
max_new_tokens = 500,
|
|
decode_special_tokens = True,
|
|
stop_conditions = stop_conditions,
|
|
gen_settings = ExLlamaV2Sampler.Settings.greedy() if greedy else None,
|
|
embeddings = image_embeddings,
|
|
)
|
|
|
|
generator.enqueue(job)
|
|
|
|
print()
|
|
print(prompt, end = ""); sys.stdout.flush()
|
|
|
|
eos = False
|
|
while generator.num_remaining_jobs():
|
|
results = generator.iterate()
|
|
for result in results:
|
|
text = result.get("text", "")
|
|
print(text, end = ""); sys.stdout.flush()
|
|
|
|
print()
|
|
|
|
else:
|
|
|
|
output = generator.generate(
|
|
prompt = prompt,
|
|
max_new_tokens = 500,
|
|
add_bos = True,
|
|
encode_special_tokens = True,
|
|
decode_special_tokens = True,
|
|
stop_conditions = [tokenizer.eos_token_id],
|
|
gen_settings = ExLlamaV2Sampler.Settings.greedy() if greedy else None,
|
|
embeddings = image_embeddings,
|
|
)
|
|
|
|
print(output) |