Skip to content

Commit b5a3050

Browse files
authored
[Fix] Add padding_side="left" for Qwen2.5 to enable flash_attention (#620)
* Add padding_side="left" to enable flash_attention * Fix linting errors
1 parent 4abf677 commit b5a3050

File tree

1 file changed

+2
-3
lines changed

1 file changed

+2
-3
lines changed

lmms_eval/models/qwen2_5_vl.py

Lines changed: 2 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -82,12 +82,11 @@ def __init__(
8282
).eval()
8383
else:
8484
self._model = Qwen2_5_VLForConditionalGeneration.from_pretrained(pretrained, torch_dtype="auto", device_map=self.device_map).eval()
85-
self.processor = AutoProcessor.from_pretrained(pretrained, max_pixels=max_pixels, min_pixels=min_pixels)
8685
self.max_pixels = max_pixels
8786
self.min_pixels = min_pixels
8887
self.max_num_frames = max_num_frames
89-
self.processor = AutoProcessor.from_pretrained(pretrained, max_pixels=max_pixels, min_pixels=min_pixels)
90-
self._tokenizer = AutoTokenizer.from_pretrained(pretrained)
88+
self.processor = AutoProcessor.from_pretrained(pretrained, max_pixels=max_pixels, min_pixels=min_pixels, padding_side="left")
89+
self._tokenizer = AutoTokenizer.from_pretrained(pretrained, padding_side="left")
9190

9291
self._config = self.model.config
9392
self.batch_size_per_gpu = int(batch_size)

0 commit comments

Comments
 (0)