Skip to content

weights problem of downloaded SiT_XL_2-256x256.pt #39

@HOHKUIK0217

Description

@HOHKUIK0217

Why the three mismatches occur (observations and consequences)

Checkpoint contains:
y_embedder.embedding_table.weight shape = (1001, 1152)
final_layer.linear.weight shape = (32, 1152)
final_layer.linear.bias shape = (32,)

RuntimeError: Error(s) in loading state_dict for SiT:
size mismatch for y_embedder.embedding_table.weight: copying a param with shape torch.Size([1001, 1152]) from checkpoint, the shape in current model is torch.Size([33, 1152]).
size mismatch for final_layer.linear.weight: copying a param with shape torch.Size([32, 1152]) from checkpoint, the shape in current model is torch.Size([16, 1152]).
size mismatch for final_layer.linear.bias: copying a param with shape torch.Size([32]) from checkpoint, the shape in current model is torch.Size([16]).

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions