Coverage report: 68%

Files Functions Classes

coverage.py v7.13.1, created at 2026-03-12 13:53 +0800

File   statements missing excluded   coverage
hyper_parallel / __init__.py   13 0 0   100%
hyper_parallel / collectives / __init__.py   0 0 0   100%
hyper_parallel / collectives / cc.py   18 1 0   94%
hyper_parallel / core / __init__.py   0 0 0   100%
hyper_parallel / core / activation_checkpoint / __init__.py   2 0 0   100%
hyper_parallel / core / activation_checkpoint / activation_checkpoint.py   17 0 0   100%
hyper_parallel / core / activation_checkpoint / swap.py   252 89 0   65%
hyper_parallel / core / checkpoint / __init__.py   11 0 0   100%
hyper_parallel / core / checkpoint / api.py   83 11 0   87%
hyper_parallel / core / checkpoint / filesystem_storage.py   172 23 0   87%
hyper_parallel / core / checkpoint / layout.py   46 6 0   87%
hyper_parallel / core / checkpoint / loader.py   11 1 0   91%
hyper_parallel / core / checkpoint / metadata.py   29 0 0   100%
hyper_parallel / core / checkpoint / planner.py   83 2 0   98%
hyper_parallel / core / checkpoint / reshard.py   159 8 0   95%
hyper_parallel / core / checkpoint / saver.py   11 1 0   91%
hyper_parallel / core / checkpoint / standard_planner.py   212 23 0   89%
hyper_parallel / core / checkpoint / storage.py   42 0 0   100%
hyper_parallel / core / checkpoint / util.py   82 12 0   85%
hyper_parallel / core / device_mesh.py   442 107 0   76%
hyper_parallel / core / dtensor.py   114 7 0   94%
hyper_parallel / core / fully_shard / __init__.py   2 0 0   100%
hyper_parallel / core / fully_shard / api.py   224 64 0   71%
hyper_parallel / core / fully_shard / hsdp_async_grad_hook.py   88 78 0   11%
hyper_parallel / core / fully_shard / hsdp_grad_buffer.py   164 145 0   12%
hyper_parallel / core / fully_shard / hsdp_grad_hook.py   113 102 0   10%
hyper_parallel / core / fully_shard / hsdp_param.py   60 27 0   55%
hyper_parallel / core / fully_shard / hsdp_param_buffer.py   84 74 0   12%
hyper_parallel / core / fully_shard / hsdp_scheduler.py   114 35 0   69%
hyper_parallel / core / fully_shard / hsdp_state.py   124 68 0   45%
hyper_parallel / core / fully_shard / hsdp_utils.py   71 12 0   83%
hyper_parallel / core / hsdp / __init__.py   2 0 0   100%
hyper_parallel / core / hsdp / api.py   101 42 0   58%
hyper_parallel / core / hsdp / hsdp_async_grad_hook.py   88 78 0   11%
hyper_parallel / core / hsdp / hsdp_grad_buffer.py   164 145 0   12%
hyper_parallel / core / hsdp / hsdp_grad_hook.py   113 58 0   49%
hyper_parallel / core / hsdp / hsdp_param.py   237 41 0   83%
hyper_parallel / core / hsdp / hsdp_param_buffer.py   66 57 0   14%
hyper_parallel / core / hsdp / hsdp_scheduler.py   89 30 0   66%
hyper_parallel / core / hsdp / hsdp_state.py   111 72 0   35%
hyper_parallel / core / hsdp / hsdp_utils.py   23 0 0   100%
hyper_parallel / core / init_weights.py   11 2 0   82%
hyper_parallel / core / layout.py   342 31 0   91%
hyper_parallel / core / parameter_init.py   4 0 0   100%
hyper_parallel / core / pipeline_parallel / __init__.py   4 0 0   100%
hyper_parallel / core / pipeline_parallel / scheduler.py   476 232 0   51%
hyper_parallel / core / pipeline_parallel / stage.py   229 75 0   67%
hyper_parallel / core / pipeline_parallel / utils.py   33 11 0   67%
hyper_parallel / core / placement_types.py   61 16 0   74%
hyper_parallel / core / random.py   184 24 0   87%
hyper_parallel / core / redistribute_infer.py   331 128 0   61%
hyper_parallel / core / shard / __init__.py   0 0 0   100%
hyper_parallel / core / shard / _op_dispatch.py   478 109 0   77%
hyper_parallel / core / shard / api.py   305 83 0   73%
hyper_parallel / core / shard / local_func.py   48 8 0   83%
hyper_parallel / core / shard / ops / __init__.py   0 0 0   100%
hyper_parallel / core / shard / ops / parallel_activation_with_axis.py   30 3 0   90%
hyper_parallel / core / shard / ops / parallel_argmax_with_value_ops.py   34 31 0   9%
hyper_parallel / core / shard / ops / parallel_concat.py   32 28 0   12%
hyper_parallel / core / shard / ops / parallel_cumsum.py   30 4 0   87%
hyper_parallel / core / shard / ops / parallel_elementwise.py   245 63 0   74%
hyper_parallel / core / shard / ops / parallel_embedding.py   26 1 0   96%
hyper_parallel / core / shard / ops / parallel_expand.py   88 12 1   86%
hyper_parallel / core / shard / ops / parallel_expand_dims.py   26 6 0   77%
hyper_parallel / core / shard / ops / parallel_flash_attention_score.py   111 21 0   81%
hyper_parallel / core / shard / ops / parallel_gather.py   131 65 0   50%
hyper_parallel / core / shard / ops / parallel_isin.py   24 2 0   92%
hyper_parallel / core / shard / ops / parallel_masked_scatter.py   12 1 0   92%
hyper_parallel / core / shard / ops / parallel_matmul.py   193 46 0   76%
hyper_parallel / core / shard / ops / parallel_ms_flash_attention_score.py   502 128 0   75%
hyper_parallel / core / shard / ops / parallel_multinomial.py   19 1 0   95%
hyper_parallel / core / shard / ops / parallel_new_ones.py   16 2 0   88%
hyper_parallel / core / shard / ops / parallel_norm.py   70 38 0   46%
hyper_parallel / core / shard / ops / parallel_npu_flash_attention_score.py   492 124 0   75%
hyper_parallel / core / shard / ops / parallel_one_hot_ext.py   169 149 0   12%
hyper_parallel / core / shard / ops / parallel_ops.py   18 6 0   67%
hyper_parallel / core / shard / ops / parallel_ops_register.py   5 0 0   100%
hyper_parallel / core / shard / ops / parallel_pad.py   32 7 0   78%
hyper_parallel / core / shard / ops / parallel_reduce.py   155 24 0   85%
hyper_parallel / core / shard / ops / parallel_repeat_interleave.py   33 5 0   85%
hyper_parallel / core / shard / ops / parallel_reshape.py   141 41 0   71%
hyper_parallel / core / shard / ops / parallel_scaled_dot_product_attention.py   177 59 0   67%
hyper_parallel / core / shard / ops / parallel_scatter.py   32 5 0   84%
hyper_parallel / core / shard / ops / parallel_scatter_update.py   41 11 0   73%
hyper_parallel / core / shard / ops / parallel_slice.py   34 29 0   15%
hyper_parallel / core / shard / ops / parallel_slice_ext.py   9 6 0   33%
hyper_parallel / core / shard / ops / parallel_sort.py   25 4 0   84%
hyper_parallel / core / shard / ops / parallel_split.py   81 39 0   52%
hyper_parallel / core / shard / ops / parallel_squeeze.py   91 20 0   78%
hyper_parallel / core / shard / ops / parallel_topk.py   18 2 0   89%
hyper_parallel / core / shard / ops / parallel_transpose.py   38 8 0   79%
hyper_parallel / core / shard / ops / parallel_tuple_elementwise.py   6 1 0   83%
hyper_parallel / core / shard / ops / parallel_unbind.py   23 1 0   96%
hyper_parallel / core / shard / sharding_plan.py   8 0 0   100%
hyper_parallel / core / tensor_redistribution.py   209 32 0   85%
hyper_parallel / core / utils / __init__.py   3 0 0   100%
hyper_parallel / core / utils / clip_grad.py   6 0 0   100%
hyper_parallel / core / utils / shape_utils.py   14 0 0   100%
hyper_parallel / platform / __init__.py   1 0 0   100%
hyper_parallel / platform / mindspore / __init__.py   0 0 0   100%
hyper_parallel / platform / mindspore / custom_pass / __init__.py   7 7 0   0%
hyper_parallel / platform / mindspore / custom_pass / build_plugin.py   56 56 0   0%
hyper_parallel / platform / mindspore / dtensor.py   89 20 0   78%
hyper_parallel / platform / mindspore / hsdp / __init__.py   0 0 0   100%
hyper_parallel / platform / mindspore / hsdp / async_grad_hook.py   9 6 0   33%
hyper_parallel / platform / mindspore / hsdp / grad_hook.py   31 10 0   68%
hyper_parallel / platform / mindspore / hsdp / param.py   29 0 0   100%
hyper_parallel / platform / mindspore / hsdp / scheduler.py   109 68 0   38%
hyper_parallel / platform / mindspore / hsdp / state.py   36 5 0   86%
hyper_parallel / platform / mindspore / init_weights.py   31 26 0   16%
hyper_parallel / platform / mindspore / parameter_init.py   32 3 0   91%
hyper_parallel / platform / mindspore / pipeline_parallel / __init__.py   0 0 0   100%
hyper_parallel / platform / mindspore / pipeline_parallel / _utils.py   80 80 0   0%
hyper_parallel / platform / mindspore / pipeline_parallel / stage.py   79 66 0   16%
hyper_parallel / platform / mindspore / platform.py   367 81 0   78%
hyper_parallel / platform / mindspore / platform_graph.py   27 10 0   63%
hyper_parallel / platform / platform.py   284 100 0   65%
hyper_parallel / platform / torch / __init__.py   0 0 0   100%
hyper_parallel / platform / torch / activation_checkpoint / __init__.py   1 0 0   100%
hyper_parallel / platform / torch / activation_checkpoint / activation_swap.py   86 17 0   80%
hyper_parallel / platform / torch / activation_checkpoint / sac.py   105 41 0   61%
hyper_parallel / platform / torch / clip_grad.py   243 58 0   76%
hyper_parallel / platform / torch / dtensor.py   122 45 0   63%
hyper_parallel / platform / torch / fully_shard / __init__.py   0 0 0   100%
hyper_parallel / platform / torch / fully_shard / async_grad_hook.py   10 10 0   0%
hyper_parallel / platform / torch / fully_shard / grad_hook.py   10 10 0   0%
hyper_parallel / platform / torch / fully_shard / hook_function.py   38 11 0   71%
hyper_parallel / platform / torch / fully_shard / param.py   389 135 0   65%
hyper_parallel / platform / torch / fully_shard / scheduler.py   82 23 0   72%
hyper_parallel / platform / torch / fully_shard / state.py   138 51 0   63%
hyper_parallel / platform / torch / fully_shard / state_dict_utils.py   45 3 0   93%
hyper_parallel / platform / torch / fully_shard / utils.py   48 3 0   94%
hyper_parallel / platform / torch / function_override.py   44 12 0   73%
hyper_parallel / platform / torch / group_utils.py   88 18 0   80%
hyper_parallel / platform / torch / hsdp / __init__.py   0 0 0   100%
hyper_parallel / platform / torch / hsdp / async_grad_hook.py   10 7 0   30%
hyper_parallel / platform / torch / hsdp / grad_hook.py   10 4 0   60%
hyper_parallel / platform / torch / hsdp / hook_function.py   37 8 0   78%
hyper_parallel / platform / torch / hsdp / param.py   22 14 0   36%
hyper_parallel / platform / torch / hsdp / scheduler.py   60 14 0   77%
hyper_parallel / platform / torch / hsdp / state.py   14 2 0   86%
hyper_parallel / platform / torch / init_weights.py   26 0 0   100%
hyper_parallel / platform / torch / pipeline_parallel / __init__.py   0 0 0   100%
hyper_parallel / platform / torch / pipeline_parallel / _utils.py   49 14 0   71%
hyper_parallel / platform / torch / pipeline_parallel / stage.py   75 10 0   87%
hyper_parallel / platform / torch / platform.py   397 78 0   80%
Total   13208 4289 1   68%

No items found using the specified filter.