File tree 1 file changed +4
-4
lines changed
ldm/modules/diffusionmodules 1 file changed +4
-4
lines changed Original file line number Diff line number Diff line change @@ -464,7 +464,7 @@ def __init__(
464
464
use_spatial_transformer = False , # custom transformer support
465
465
transformer_depth = 1 , # custom transformer support
466
466
context_dim = None , # custom transformer support
467
- n_embed = None , # custom support for prediction of discrete ids into codebook of first stage vq model
467
+ n_embed = None , # custom support for prediction of discrete ids into codebook of first stage vq model
468
468
legacy = True ,
469
469
):
470
470
super ().__init__ ()
@@ -545,7 +545,7 @@ def __init__(
545
545
num_heads = ch // num_head_channels
546
546
dim_head = num_head_channels
547
547
if legacy :
548
- num_heads = 1
548
+ # num_heads = 1
549
549
dim_head = ch // num_heads if use_spatial_transformer else num_head_channels
550
550
layers .append (
551
551
AttentionBlock (
@@ -592,7 +592,7 @@ def __init__(
592
592
num_heads = ch // num_head_channels
593
593
dim_head = num_head_channels
594
594
if legacy :
595
- num_heads = 1
595
+ # num_heads = 1
596
596
dim_head = ch // num_heads if use_spatial_transformer else num_head_channels
597
597
self .middle_block = TimestepEmbedSequential (
598
598
ResBlock (
@@ -646,7 +646,7 @@ def __init__(
646
646
num_heads = ch // num_head_channels
647
647
dim_head = num_head_channels
648
648
if legacy :
649
- num_heads = 1
649
+ # num_heads = 1
650
650
dim_head = ch // num_heads if use_spatial_transformer else num_head_channels
651
651
layers .append (
652
652
AttentionBlock (
You can’t perform that action at this time.
0 commit comments