Skip to content

Commit 2669a5c

Browse files
committed
Fix q_proj and o_proj shape
1 parent a5907fa commit 2669a5c

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

backends/candle/src/models/flash_qwen3.rs

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,7 @@ impl Qwen3Attention {
3838
let hidden_size = config.hidden_size;
3939

4040
let query_weight = vb.pp("q_proj").get(
41-
(num_attention_heads * attention_head_size, hidden_size),
41+
(hidden_size, num_attention_heads * attention_head_size),
4242
"weight",
4343
)?;
4444
let query_bias = if config.attention_bias {
@@ -77,7 +77,7 @@ impl Qwen3Attention {
7777
let v_proj = Linear::new(value_weight, value_bias, None);
7878

7979
let o_proj_weight = vb.pp("o_proj").get(
80-
(num_attention_heads * attention_head_size, hidden_size),
80+
(hidden_size, num_attention_heads * attention_head_size),
8181
"weight",
8282
)?;
8383
let o_proj = Linear::new(o_proj_weight, None, None);

0 commit comments

Comments
 (0)