Skip to content

Commit 144c8ce

Browse files
authored
Fix modular docstring for Mixtral (#42041)
* Fix modular docstring for Mixtral * fixes all docstrings
1 parent 069684e commit 144c8ce

File tree

18 files changed

+38
-38
lines changed

18 files changed

+38
-38
lines changed

src/transformers/models/deepseek_v2/modeling_deepseek_v2.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -59,8 +59,8 @@ def forward(
5959
"""
6060
Args:
6161
hidden_states: (batch_size * sequence_length, hidden_dim)
62-
selected_experts: (batch_size * sequence_length, top_k)
63-
routing_weights: (batch_size * sequence_length, top_k)
62+
top_k_index: (batch_size * sequence_length, top_k)
63+
top_k_weights: (batch_size * sequence_length, top_k)
6464
Returns:
6565
(batch_size * sequence_length, hidden_dim)
6666
"""

src/transformers/models/deepseek_v3/modeling_deepseek_v3.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -166,8 +166,8 @@ def forward(
166166
"""
167167
Args:
168168
hidden_states: (batch_size * sequence_length, hidden_dim)
169-
selected_experts: (batch_size * sequence_length, top_k)
170-
routing_weights: (batch_size * sequence_length, top_k)
169+
top_k_index: (batch_size * sequence_length, top_k)
170+
top_k_weights: (batch_size * sequence_length, top_k)
171171
Returns:
172172
(batch_size * sequence_length, hidden_dim)
173173
"""

src/transformers/models/dots1/modeling_dots1.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -322,8 +322,8 @@ def forward(
322322
"""
323323
Args:
324324
hidden_states: (batch_size * sequence_length, hidden_dim)
325-
selected_experts: (batch_size * sequence_length, top_k)
326-
routing_weights: (batch_size * sequence_length, top_k)
325+
top_k_index: (batch_size * sequence_length, top_k)
326+
top_k_weights: (batch_size * sequence_length, top_k)
327327
Returns:
328328
(batch_size * sequence_length, hidden_dim)
329329
"""

src/transformers/models/flex_olmo/modeling_flex_olmo.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -310,8 +310,8 @@ def forward(
310310
"""
311311
Args:
312312
hidden_states: (batch_size * sequence_length, hidden_dim)
313-
selected_experts: (batch_size * sequence_length, top_k)
314-
routing_weights: (batch_size * sequence_length, top_k)
313+
top_k_index: (batch_size * sequence_length, top_k)
314+
top_k_weights: (batch_size * sequence_length, top_k)
315315
Returns:
316316
(batch_size * sequence_length, hidden_dim)
317317
"""

src/transformers/models/glm4_moe/modeling_glm4_moe.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -347,8 +347,8 @@ def forward(
347347
"""
348348
Args:
349349
hidden_states: (batch_size * sequence_length, hidden_dim)
350-
selected_experts: (batch_size * sequence_length, top_k)
351-
routing_weights: (batch_size * sequence_length, top_k)
350+
top_k_index: (batch_size * sequence_length, top_k)
351+
top_k_weights: (batch_size * sequence_length, top_k)
352352
Returns:
353353
(batch_size * sequence_length, hidden_dim)
354354
"""

src/transformers/models/glm4v_moe/modeling_glm4v_moe.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -368,8 +368,8 @@ def forward(
368368
"""
369369
Args:
370370
hidden_states: (batch_size * sequence_length, hidden_dim)
371-
selected_experts: (batch_size * sequence_length, top_k)
372-
routing_weights: (batch_size * sequence_length, top_k)
371+
top_k_index: (batch_size * sequence_length, top_k)
372+
top_k_weights: (batch_size * sequence_length, top_k)
373373
Returns:
374374
(batch_size * sequence_length, hidden_dim)
375375
"""

src/transformers/models/hunyuan_v1_moe/modeling_hunyuan_v1_moe.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -261,8 +261,8 @@ def forward(
261261
"""
262262
Args:
263263
hidden_states: (batch_size * sequence_length, hidden_dim)
264-
selected_experts: (batch_size * sequence_length, top_k)
265-
routing_weights: (batch_size * sequence_length, top_k)
264+
top_k_index: (batch_size * sequence_length, top_k)
265+
top_k_weights: (batch_size * sequence_length, top_k)
266266
Returns:
267267
(batch_size * sequence_length, hidden_dim)
268268
"""

src/transformers/models/jamba/modeling_jamba.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -575,8 +575,8 @@ def forward(
575575
"""
576576
Args:
577577
hidden_states: (batch_size * sequence_length, hidden_dim)
578-
selected_experts: (batch_size * sequence_length, top_k)
579-
routing_weights: (batch_size * sequence_length, top_k)
578+
top_k_index: (batch_size * sequence_length, top_k)
579+
top_k_weights: (batch_size * sequence_length, top_k)
580580
Returns:
581581
(batch_size * sequence_length, hidden_dim)
582582
"""

src/transformers/models/lfm2_moe/modeling_lfm2_moe.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -161,8 +161,8 @@ def forward(
161161
"""
162162
Args:
163163
hidden_states: (batch_size * sequence_length, hidden_dim)
164-
selected_experts: (batch_size * sequence_length, top_k)
165-
routing_weights: (batch_size * sequence_length, top_k)
164+
top_k_index: (batch_size * sequence_length, top_k)
165+
top_k_weights: (batch_size * sequence_length, top_k)
166166
Returns:
167167
(batch_size * sequence_length, hidden_dim)
168168
"""

src/transformers/models/minimax/modeling_minimax.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -488,8 +488,8 @@ def forward(
488488
"""
489489
Args:
490490
hidden_states: (batch_size * sequence_length, hidden_dim)
491-
selected_experts: (batch_size * sequence_length, top_k)
492-
routing_weights: (batch_size * sequence_length, top_k)
491+
top_k_index: (batch_size * sequence_length, top_k)
492+
top_k_weights: (batch_size * sequence_length, top_k)
493493
Returns:
494494
(batch_size * sequence_length, hidden_dim)
495495
"""

0 commit comments

Comments
 (0)