colossalai
52 строки · 1.7 Кб
1import torch2from torch.fx import symbolic_trace3
4from colossalai.fx._compatibility import is_compatible_with_meta5from colossalai.fx.passes.adding_split_node_pass import split_with_split_nodes_pass, uniform_split_pass6from colossalai.fx.passes.meta_info_prop import MetaInfoProp7from colossalai.fx.passes.utils import get_comm_size8from colossalai.testing import clear_cache_before_run9
10is_compatible = is_compatible_with_meta()11if is_compatible:12from colossalai.fx.profiler import MetaTensor13
14MODEL_DIM = 1615BATCH_SIZE = 816PIPELINE_SIZE = 217
18
19class MLP(torch.nn.Module):20def __init__(self, dim: int):21super().__init__()22self.linear1 = torch.nn.Linear(dim, dim)23self.linear2 = torch.nn.Linear(dim, dim)24self.linear3 = torch.nn.Linear(dim, dim)25self.linear4 = torch.nn.Linear(dim, dim)26
27def forward(self, x):28x = self.linear1(x)29x = self.linear2(x)30x = self.linear3(x)31x = self.linear4(x)32return x33
34
35@clear_cache_before_run()36def test_comm_size_compute():37model = MLP(MODEL_DIM)38input_sample = torch.rand(BATCH_SIZE, MODEL_DIM, device="meta")39gm = symbolic_trace(model)40if is_compatible:41input_sample = MetaTensor(input_sample, fake_device=next(gm.parameters()).device)42MetaInfoProp(gm).run(input_sample)43annotated_model = uniform_split_pass(gm, PIPELINE_SIZE)44split_model, split_submodules = split_with_split_nodes_pass(annotated_model)45submodule_list = list(split_model.children())46comm_size = get_comm_size(submodule_list[0], submodule_list[1])47# the shape of tensor send from partition 0 to partition 1 is (8, 16)48assert comm_size == 12849
50
51if __name__ == "__main__":52test_comm_size_compute()53