From 715ad2f7e2540ebeff04ff78f27e23e4cd6c528f Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Jonatan=20K=C5=82osko?= Date: Sat, 24 Feb 2024 03:12:40 +0700 Subject: [PATCH] Fix loading Mistral configuration with attention window disabled --- lib/bumblebee/text/mistral.ex | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/lib/bumblebee/text/mistral.ex b/lib/bumblebee/text/mistral.ex index c886deca..a045943a 100644 --- a/lib/bumblebee/text/mistral.ex +++ b/lib/bumblebee/text/mistral.ex @@ -393,7 +393,7 @@ defmodule Bumblebee.Text.Mistral do num_blocks: {"num_hidden_layers", number()}, num_attention_heads: {"num_attention_heads", number()}, num_key_value_heads: {"num_key_value_heads", number()}, - attention_window_size: {"sliding_window", number()}, + attention_window_size: {"sliding_window", optional(number())}, intermediate_size: {"intermediate_size", number()}, activation: {"hidden_act", activation()}, rotary_embedding_base: {"rope_theta", number()},