Skip to content

Commit e9b1649

Browse files
authored
Fix Gemma3n inference without cache (#323)
Closes #322
1 parent b60cec8 commit e9b1649

File tree

2 files changed

+2
-2
lines changed

2 files changed

+2
-2
lines changed

mlx_lm/_version.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,3 @@
11
# Copyright © 2023-2025 Apple Inc.
22

3-
__version__ = "0.26.1"
3+
__version__ = "0.26.2"

mlx_lm/models/gemma3n.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -476,7 +476,7 @@ def __call__(
476476
per_layer_inputs = self.project_per_layer_inputs(h, per_layer_inputs)
477477

478478
if cache is None:
479-
cache = [None] * len(self.layers)
479+
cache = self.make_cache()
480480

481481
if mask is None:
482482
full_mask = create_attention_mask(

0 commit comments

Comments
 (0)