model state_dict 내부

heyme·2023년 3월 23일

Load checkpoint from: /root/share/heyme_share_239/wav2lip/syncnet_checkpoint/wav2lip.pth
000 k module.face_encoder_blocks.0.0.conv_block.0.weight
v torch.Size([16, 6, 7, 7])
001 k module.face_encoder_blocks.0.0.conv_block.0.bias
v torch.Size([16])
002 k module.face_encoder_blocks.0.0.conv_block.1.weight
v torch.Size([16])
003 k module.face_encoder_blocks.0.0.conv_block.1.bias
v torch.Size([16])
004 k module.face_encoder_blocks.0.0.conv_block.1.running_mean
v torch.Size([16])
005 k module.face_encoder_blocks.0.0.conv_block.1.running_var
v torch.Size([16])
006 k module.face_encoder_blocks.0.0.conv_block.1.num_batches_tracked
v torch.Size([])
007 k module.face_encoder_blocks.1.0.conv_block.0.weight
v torch.Size([32, 16, 3, 3])
008 k module.face_encoder_blocks.1.0.conv_block.0.bias
v torch.Size([32])
009 k module.face_encoder_blocks.1.0.conv_block.1.weight
v torch.Size([32])
010 k module.face_encoder_blocks.1.0.conv_block.1.bias
v torch.Size([32])
011 k module.face_encoder_blocks.1.0.conv_block.1.running_mean
v torch.Size([32])
012 k module.face_encoder_blocks.1.0.conv_block.1.running_var
v torch.Size([32])
013 k module.face_encoder_blocks.1.0.conv_block.1.num_batches_tracked
v torch.Size([])
014 k module.face_encoder_blocks.1.1.conv_block.0.weight
v torch.Size([32, 32, 3, 3])
015 k module.face_encoder_blocks.1.1.conv_block.0.bias
v torch.Size([32])
016 k module.face_encoder_blocks.1.1.conv_block.1.weight
v torch.Size([32])
017 k module.face_encoder_blocks.1.1.conv_block.1.bias
v torch.Size([32])
018 k module.face_encoder_blocks.1.1.conv_block.1.running_mean
v torch.Size([32])
019 k module.face_encoder_blocks.1.1.conv_block.1.running_var
v torch.Size([32])
020 k module.face_encoder_blocks.1.1.conv_block.1.num_batches_tracked
v torch.Size([])
021 k module.face_encoder_blocks.1.2.conv_block.0.weight
v torch.Size([32, 32, 3, 3])
022 k module.face_encoder_blocks.1.2.conv_block.0.bias
v torch.Size([32])
023 k module.face_encoder_blocks.1.2.conv_block.1.weight
v torch.Size([32])
024 k module.face_encoder_blocks.1.2.conv_block.1.bias
v torch.Size([32])
025 k module.face_encoder_blocks.1.2.conv_block.1.running_mean
v torch.Size([32])
026 k module.face_encoder_blocks.1.2.conv_block.1.running_var
v torch.Size([32])
027 k module.face_encoder_blocks.1.2.conv_block.1.num_batches_tracked
v torch.Size([])
028 k module.face_encoder_blocks.2.0.conv_block.0.weight
v torch.Size([64, 32, 3, 3])
029 k module.face_encoder_blocks.2.0.conv_block.0.bias
v torch.Size([64])
030 k module.face_encoder_blocks.2.0.conv_block.1.weight
v torch.Size([64])
031 k module.face_encoder_blocks.2.0.conv_block.1.bias
v torch.Size([64])
032 k module.face_encoder_blocks.2.0.conv_block.1.running_mean
v torch.Size([64])
033 k module.face_encoder_blocks.2.0.conv_block.1.running_var
v torch.Size([64])
034 k module.face_encoder_blocks.2.0.conv_block.1.num_batches_tracked
v torch.Size([])
035 k module.face_encoder_blocks.2.1.conv_block.0.weight
v torch.Size([64, 64, 3, 3])
036 k module.face_encoder_blocks.2.1.conv_block.0.bias
v torch.Size([64])
037 k module.face_encoder_blocks.2.1.conv_block.1.weight
v torch.Size([64])
038 k module.face_encoder_blocks.2.1.conv_block.1.bias
v torch.Size([64])
039 k module.face_encoder_blocks.2.1.conv_block.1.running_mean
v torch.Size([64])
040 k module.face_encoder_blocks.2.1.conv_block.1.running_var
v torch.Size([64])
041 k module.face_encoder_blocks.2.1.conv_block.1.num_batches_tracked
v torch.Size([])
042 k module.face_encoder_blocks.2.2.conv_block.0.weight
v torch.Size([64, 64, 3, 3])
043 k module.face_encoder_blocks.2.2.conv_block.0.bias
v torch.Size([64])
044 k module.face_encoder_blocks.2.2.conv_block.1.weight
v torch.Size([64])
045 k module.face_encoder_blocks.2.2.conv_block.1.bias
v torch.Size([64])
046 k module.face_encoder_blocks.2.2.conv_block.1.running_mean
v torch.Size([64])
047 k module.face_encoder_blocks.2.2.conv_block.1.running_var
v torch.Size([64])
048 k module.face_encoder_blocks.2.2.conv_block.1.num_batches_tracked
v torch.Size([])
049 k module.face_encoder_blocks.2.3.conv_block.0.weight
v torch.Size([64, 64, 3, 3])
050 k module.face_encoder_blocks.2.3.conv_block.0.bias
v torch.Size([64])
.
.
.

0개의 댓글