Qwen
/

yangapku commited on
Commit
328fbcc
1 Parent(s): e0b4ff6

remove unused config

Browse files
Files changed (2) hide show
  1. config.json +0 -1
  2. modeling_qwen.py +1 -1
config.json CHANGED
@@ -1,5 +1,4 @@
1
  {
2
- "activation": "swiglu",
3
  "apply_residual_connection_post_layernorm": false,
4
  "architectures": [
5
  "QWenLMHeadModel"
 
1
  {
 
2
  "apply_residual_connection_post_layernorm": false,
3
  "architectures": [
4
  "QWenLMHeadModel"
modeling_qwen.py CHANGED
@@ -245,7 +245,7 @@ class QWenAttention(nn.Module):
245
  math.log(i, self.seq_length) if i > self.seq_length else 1
246
  for i in range(1, 32768)
247
  ]
248
- self.logn_tensor = torch.Tensor(logn_list)[None, :, None, None]
249
  self._ntk_cached = 1.0
250
 
251
  self.attn_dropout = nn.Dropout(config.attn_pdrop)
 
245
  math.log(i, self.seq_length) if i > self.seq_length else 1
246
  for i in range(1, 32768)
247
  ]
248
+ self.logn_tensor = torch.tensor(logn_list)[None, :, None, None]
249
  self._ntk_cached = 1.0
250
 
251
  self.attn_dropout = nn.Dropout(config.attn_pdrop)