Spaces:
Running
on
Zero
Running
on
Zero
Update src/transformer.py
Browse files- src/transformer.py +0 -7
src/transformer.py
CHANGED
|
@@ -7,7 +7,6 @@ from diffusers.models.transformers.transformer_flux import (
|
|
| 7 |
FluxTransformer2DModel,
|
| 8 |
Transformer2DModelOutput,
|
| 9 |
USE_PEFT_BACKEND,
|
| 10 |
-
is_torch_version,
|
| 11 |
scale_lora_layers,
|
| 12 |
unscale_lora_layers,
|
| 13 |
logger,
|
|
@@ -155,9 +154,6 @@ def tranformer_forward(
|
|
| 155 |
|
| 156 |
return custom_forward
|
| 157 |
|
| 158 |
-
ckpt_kwargs: Dict[str, Any] = (
|
| 159 |
-
{"use_reentrant": False} if is_torch_version(">=", "1.11.0") else {}
|
| 160 |
-
)
|
| 161 |
encoder_hidden_states, hidden_states = torch.utils.checkpoint.checkpoint(
|
| 162 |
create_custom_forward(block),
|
| 163 |
hidden_states,
|
|
@@ -204,9 +200,6 @@ def tranformer_forward(
|
|
| 204 |
|
| 205 |
return custom_forward
|
| 206 |
|
| 207 |
-
ckpt_kwargs: Dict[str, Any] = (
|
| 208 |
-
{"use_reentrant": False} if is_torch_version(">=", "1.11.0") else {}
|
| 209 |
-
)
|
| 210 |
hidden_states = torch.utils.checkpoint.checkpoint(
|
| 211 |
create_custom_forward(block),
|
| 212 |
hidden_states,
|
|
|
|
| 7 |
FluxTransformer2DModel,
|
| 8 |
Transformer2DModelOutput,
|
| 9 |
USE_PEFT_BACKEND,
|
|
|
|
| 10 |
scale_lora_layers,
|
| 11 |
unscale_lora_layers,
|
| 12 |
logger,
|
|
|
|
| 154 |
|
| 155 |
return custom_forward
|
| 156 |
|
|
|
|
|
|
|
|
|
|
| 157 |
encoder_hidden_states, hidden_states = torch.utils.checkpoint.checkpoint(
|
| 158 |
create_custom_forward(block),
|
| 159 |
hidden_states,
|
|
|
|
| 200 |
|
| 201 |
return custom_forward
|
| 202 |
|
|
|
|
|
|
|
|
|
|
| 203 |
hidden_states = torch.utils.checkpoint.checkpoint(
|
| 204 |
create_custom_forward(block),
|
| 205 |
hidden_states,
|