diff --git a/README.md b/README.md index 4c51e30..06fb2c8 100644 --- a/README.md +++ b/README.md @@ -1,6 +1,7 @@ --- license: apache-2.0 pipeline_tag: image-text-to-text +library_name: transformers --- Chat @@ -93,7 +94,7 @@ model = Qwen3VLForConditionalGeneration.from_pretrained( # device_map="auto", # ) -processor = AutoProcessor.from_pretrained("Qwen/Qwen/Qwen3-VL-8B-Instruct") +processor = AutoProcessor.from_pretrained("Qwen/Qwen3-VL-8B-Instruct") messages = [ { @@ -116,6 +117,7 @@ inputs = processor.apply_chat_template( return_dict=True, return_tensors="pt" ) +inputs = inputs.to(model.device) # Inference: Generation of the output generated_ids = model.generate(**inputs, max_new_tokens=128)