|
7 | 7 | from typing import Any, Callable, Literal |
8 | 8 |
|
9 | 9 | import torch |
10 | | -import transformers |
11 | 10 | from tensordict import lazy_stack, TensorDict, TensorDictBase |
12 | 11 | from torch.utils.data import DataLoader |
13 | 12 | from torchrl.data import Composite, NonTensor |
@@ -116,7 +115,7 @@ def __init__( |
116 | 115 | batch_size: tuple | torch.Size | None = None, |
117 | 116 | system_prompt: str | None = None, |
118 | 117 | apply_template: bool | None = None, |
119 | | - tokenizer: transformers.AutoTokenizer | None = None, |
| 118 | + tokenizer: transformers.AutoTokenizer | None = None, # noqa: F821 |
120 | 119 | template_kwargs: dict[str, Any] | None = None, |
121 | 120 | system_role: str = "system", |
122 | 121 | user_role: str = "user", |
@@ -309,7 +308,7 @@ def __init__( |
309 | 308 | batch_size_dl: int = 1, |
310 | 309 | seed: int | None = None, |
311 | 310 | group_repeats: bool = False, |
312 | | - tokenizer: transformers.AutoTokenizer | None = None, |
| 311 | + tokenizer: transformers.AutoTokenizer | None = None, # noqa: F821 |
313 | 312 | device: torch.device | None = None, |
314 | 313 | template_kwargs: dict[str, Any] | None = None, |
315 | 314 | apply_template: bool | None = None, |
|
0 commit comments