We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
q_proj
o_proj
1 parent a5907fa commit 2669a5cCopy full SHA for 2669a5c
backends/candle/src/models/flash_qwen3.rs
@@ -38,7 +38,7 @@ impl Qwen3Attention {
38
let hidden_size = config.hidden_size;
39
40
let query_weight = vb.pp("q_proj").get(
41
- (num_attention_heads * attention_head_size, hidden_size),
+ (hidden_size, num_attention_heads * attention_head_size),
42
"weight",
43
)?;
44
let query_bias = if config.attention_bias {
@@ -77,7 +77,7 @@ impl Qwen3Attention {
77
let v_proj = Linear::new(value_weight, value_bias, None);
78
79
let o_proj_weight = vb.pp("o_proj").get(
80
81
82
83
let o_proj = Linear::new(o_proj_weight, None, None);
0 commit comments