Coverage report: 68%

Files Functions Classes

coverage.py v7.13.1, created at 2026-02-21 06:04 +0800

File class   statements missing excluded   coverage
hyper_parallel / __init__.py (no class)   12 0 0   100%
hyper_parallel / collectives / __init__.py (no class)   0 0 0   100%
hyper_parallel / collectives / cc.py (no class)   14 0 0   100%
hyper_parallel / core / __init__.py (no class)   0 0 0   100%
hyper_parallel / core / activation_checkpoint / __init__.py (no class)   2 0 0   100%
hyper_parallel / core / activation_checkpoint / activation_checkpoint.py CheckpointPolicy   0 0 0   100%
hyper_parallel / core / activation_checkpoint / activation_checkpoint.py (no class)   17 0 0   100%
hyper_parallel / core / activation_checkpoint / swap.py SwapTensor   54 37 0   31%
hyper_parallel / core / activation_checkpoint / swap.py Storage   30 14 0   53%
hyper_parallel / core / activation_checkpoint / swap.py SwapGroup   37 16 0   57%
hyper_parallel / core / activation_checkpoint / swap.py SwapManager   82 22 0   73%
hyper_parallel / core / activation_checkpoint / swap.py (no class)   49 0 0   100%
hyper_parallel / core / checkpoint / __init__.py (no class)   11 0 0   100%
hyper_parallel / core / checkpoint / api.py (no class)   83 11 0   87%
hyper_parallel / core / checkpoint / filesystem_storage.py FileSystemWriter   66 8 0   88%
hyper_parallel / core / checkpoint / filesystem_storage.py FileSystemReader   44 11 0   75%
hyper_parallel / core / checkpoint / filesystem_storage.py (no class)   62 4 0   94%
hyper_parallel / core / checkpoint / layout.py (no class)   46 6 0   87%
hyper_parallel / core / checkpoint / loader.py (no class)   11 1 0   91%
hyper_parallel / core / checkpoint / metadata.py MetadataIndex   0 0 0   100%
hyper_parallel / core / checkpoint / metadata.py ChunkStorageMetadata   0 0 0   100%
hyper_parallel / core / checkpoint / metadata.py TensorProperties   0 0 0   100%
hyper_parallel / core / checkpoint / metadata.py BytesStorageMetadata   0 0 0   100%
hyper_parallel / core / checkpoint / metadata.py TensorStorageMetadata   0 0 0   100%
hyper_parallel / core / checkpoint / metadata.py Metadata   0 0 0   100%
hyper_parallel / core / checkpoint / metadata.py (no class)   29 0 0   100%
hyper_parallel / core / checkpoint / planner.py WriteItemType   0 0 0   100%
hyper_parallel / core / checkpoint / planner.py LoadItemType   0 0 0   100%
hyper_parallel / core / checkpoint / planner.py WriteItem   21 2 0   90%
hyper_parallel / core / checkpoint / planner.py ReadItem   0 0 0   100%
hyper_parallel / core / checkpoint / planner.py SavePlan   0 0 0   100%
hyper_parallel / core / checkpoint / planner.py LoadPlan   0 0 0   100%
hyper_parallel / core / checkpoint / planner.py SavePlanner   0 0 0   100%
hyper_parallel / core / checkpoint / planner.py LoadPlanner   0 0 0   100%
hyper_parallel / core / checkpoint / planner.py (no class)   62 0 0   100%
hyper_parallel / core / checkpoint / reshard.py ReshardHandler   80 2 0   98%
hyper_parallel / core / checkpoint / reshard.py (no class)   79 6 0   92%
hyper_parallel / core / checkpoint / saver.py (no class)   11 1 0   91%
hyper_parallel / core / checkpoint / standard_planner.py StandardSavePlanner   109 9 0   92%
hyper_parallel / core / checkpoint / standard_planner.py StandardLoadPlanner   58 13 0   78%
hyper_parallel / core / checkpoint / standard_planner.py (no class)   45 1 0   98%
hyper_parallel / core / checkpoint / storage.py StorageInfo   0 0 0   100%
hyper_parallel / core / checkpoint / storage.py WriteResult   0 0 0   100%
hyper_parallel / core / checkpoint / storage.py StorageWriter   0 0 0   100%
hyper_parallel / core / checkpoint / storage.py StorageReader   0 0 0   100%
hyper_parallel / core / checkpoint / storage.py (no class)   42 0 0   100%
hyper_parallel / core / checkpoint / util.py (no class)   82 12 0   85%
hyper_parallel / core / device_mesh.py DeviceMesh   331 129 0   61%
hyper_parallel / core / device_mesh.py (no class)   101 3 0   97%
hyper_parallel / core / dtensor.py SkipDTensorDispatch   4 0 0   100%
hyper_parallel / core / dtensor.py DTensor   41 7 0   83%
hyper_parallel / core / dtensor.py (no class)   60 0 0   100%
hyper_parallel / core / fully_shard / __init__.py (no class)   2 0 0   100%
hyper_parallel / core / fully_shard / api.py _UnshardHandle   4 4 0   0%
hyper_parallel / core / fully_shard / api.py HSDPModule   122 117 0   4%
hyper_parallel / core / fully_shard / api.py (no class)   119 37 0   69%
hyper_parallel / core / fully_shard / hsdp_async_grad_hook.py HSDPAsyncGradHook   78 78 0   0%
hyper_parallel / core / fully_shard / hsdp_async_grad_hook.py (no class)   10 0 0   100%
hyper_parallel / core / fully_shard / hsdp_grad_buffer.py HSDPGradBuffer   145 145 0   0%
hyper_parallel / core / fully_shard / hsdp_grad_buffer.py (no class)   19 0 0   100%
hyper_parallel / core / fully_shard / hsdp_grad_hook.py HSDPGradHook   102 102 0   0%
hyper_parallel / core / fully_shard / hsdp_grad_hook.py (no class)   11 0 0   100%
hyper_parallel / core / fully_shard / hsdp_param.py HSDPParamV2   27 27 0   0%
hyper_parallel / core / fully_shard / hsdp_param.py (no class)   33 0 0   100%
hyper_parallel / core / fully_shard / hsdp_param_buffer.py HSDPParamBuffer   74 74 0   0%
hyper_parallel / core / fully_shard / hsdp_param_buffer.py (no class)   10 0 0   100%
hyper_parallel / core / fully_shard / hsdp_scheduler.py HSDPSchedulerContext   5 0 0   100%
hyper_parallel / core / fully_shard / hsdp_scheduler.py HSDPSchedulerV2   84 50 0   40%
hyper_parallel / core / fully_shard / hsdp_scheduler.py (no class)   29 0 0   100%
hyper_parallel / core / fully_shard / hsdp_state.py HSDPState   113 82 0   27%
hyper_parallel / core / fully_shard / hsdp_state.py (no class)   18 0 0   100%
hyper_parallel / core / fully_shard / hsdp_utils.py OptimizerLevel   0 0 0   100%
hyper_parallel / core / fully_shard / hsdp_utils.py GroupInfo   3 3 0   0%
hyper_parallel / core / fully_shard / hsdp_utils.py HSDPConfigV2   10 0 0   100%
hyper_parallel / core / fully_shard / hsdp_utils.py ShardedState   0 0 0   100%
hyper_parallel / core / fully_shard / hsdp_utils.py FSDPSchedulerState   0 0 0   100%
hyper_parallel / core / fully_shard / hsdp_utils.py ParamModuleInfo   0 0 0   100%
hyper_parallel / core / fully_shard / hsdp_utils.py ExtensionsData   2 2 0   0%
hyper_parallel / core / fully_shard / hsdp_utils.py (no class)   56 7 0   88%
hyper_parallel / core / hsdp / __init__.py (no class)   2 0 0   100%
hyper_parallel / core / hsdp / api.py HSDPCell   35 9 0   74%
hyper_parallel / core / hsdp / api.py (no class)   66 14 0   79%
hyper_parallel / core / hsdp / hsdp_async_grad_hook.py HSDPAsyncGradHook   78 21 0   73%
hyper_parallel / core / hsdp / hsdp_async_grad_hook.py (no class)   10 0 0   100%
hyper_parallel / core / hsdp / hsdp_grad_buffer.py HSDPGradBuffer   145 76 0   48%
hyper_parallel / core / hsdp / hsdp_grad_buffer.py (no class)   19 0 0   100%
hyper_parallel / core / hsdp / hsdp_grad_hook.py HSDPGradHook   102 39 0   62%
hyper_parallel / core / hsdp / hsdp_grad_hook.py (no class)   11 0 0   100%
hyper_parallel / core / hsdp / hsdp_param.py HSDPParam   219 16 0   93%
hyper_parallel / core / hsdp / hsdp_param.py (no class)   20 0 0   100%
hyper_parallel / core / hsdp / hsdp_param_buffer.py HSDPParamBuffer   57 10 0   82%
hyper_parallel / core / hsdp / hsdp_param_buffer.py (no class)   9 0 0   100%
hyper_parallel / core / hsdp / hsdp_scheduler.py FSDPSchedulerState   0 0 0   100%
hyper_parallel / core / hsdp / hsdp_scheduler.py HSDPScheduler   62 9 0   85%
hyper_parallel / core / hsdp / hsdp_scheduler.py (no class)   27 0 0   100%
hyper_parallel / core / hsdp / hsdp_state.py HSDPState   98 23 0   77%
hyper_parallel / core / hsdp / hsdp_state.py (no class)   13 0 0   100%
hyper_parallel / core / hsdp / hsdp_utils.py OptimizerLevel   0 0 0   100%
hyper_parallel / core / hsdp / hsdp_utils.py GroupInfo   3 0 0   100%
hyper_parallel / core / hsdp / hsdp_utils.py HSDPConfig   11 0 0   100%
hyper_parallel / core / hsdp / hsdp_utils.py (no class)   9 0 0   100%
hyper_parallel / core / layout.py Layout   218 29 0   87%
hyper_parallel / core / layout.py (no class)   116 0 0   100%
hyper_parallel / core / parameter_init.py (no class)   4 0 0   100%
hyper_parallel / core / pipeline_parallel / __init__.py (no class)   4 0 0   100%
hyper_parallel / core / pipeline_parallel / scheduler.py MetaStepType   0 0 0   100%
hyper_parallel / core / pipeline_parallel / scheduler.py MetaStep   16 10 0   38%
hyper_parallel / core / pipeline_parallel / scheduler.py PipelineScheduleRuntime   92 15 0   84%
hyper_parallel / core / pipeline_parallel / scheduler.py ScheduleGPipe   17 17 0   0%
hyper_parallel / core / pipeline_parallel / scheduler.py Schedule1F1B   44 44 0   0%
hyper_parallel / core / pipeline_parallel / scheduler.py ScheduleInterleaved1F1B   63 8 0   87%
hyper_parallel / core / pipeline_parallel / scheduler.py (no class)   244 138 0   43%
hyper_parallel / core / pipeline_parallel / stage.py SharedParameterInfo   11 11 0   0%
hyper_parallel / core / pipeline_parallel / stage.py PipelineStage   183 53 0   71%
hyper_parallel / core / pipeline_parallel / stage.py (no class)   35 0 0   100%
hyper_parallel / core / pipeline_parallel / utils.py BatchDimSpec   11 11 0   0%
hyper_parallel / core / pipeline_parallel / utils.py _RecvInfo   5 0 0   100%
hyper_parallel / core / pipeline_parallel / utils.py (no class)   17 0 0   100%
hyper_parallel / core / placement_types.py Placement   4 4 0   0%
hyper_parallel / core / placement_types.py Shard   11 7 0   36%
hyper_parallel / core / placement_types.py Replicate   4 4 0   0%
hyper_parallel / core / placement_types.py Partial   13 10 0   23%
hyper_parallel / core / placement_types.py (no class)   29 0 0   100%
hyper_parallel / core / random.py _PhiloxState   8 3 0   62%
hyper_parallel / core / random.py _RNGStateTracker   9 2 0   78%
hyper_parallel / core / random.py OffsetBasedRNGTracker   40 2 0   95%
hyper_parallel / core / random.py (no class)   128 35 0   73%
hyper_parallel / core / redistribute_infer.py Status   0 0 0   100%
hyper_parallel / core / redistribute_infer.py TensorMap   14 8 0   43%
hyper_parallel / core / redistribute_infer.py DevMat   47 25 0   47%
hyper_parallel / core / redistribute_infer.py RedistributionOperatorInfer   236 95 0   60%
hyper_parallel / core / redistribute_infer.py (no class)   34 0 0   100%
hyper_parallel / core / shard / __init__.py (no class)   0 0 0   100%
hyper_parallel / core / shard / _op_dispatch.py LayoutCacheKey   9 1 0   89%
hyper_parallel / core / shard / _op_dispatch.py LayoutCacheManager   9 0 0   100%
hyper_parallel / core / shard / _op_dispatch.py OpDispatcher   374 97 0   74%
hyper_parallel / core / shard / _op_dispatch.py (no class)   49 0 0   100%
hyper_parallel / core / shard / api.py parallelize_value_and_grad.CellWrapper   4 1 0   75%
hyper_parallel / core / shard / api.py (no class)   301 68 0   77%
hyper_parallel / core / shard / local_func.py (no class)   48 8 0   83%
hyper_parallel / core / shard / ops / __init__.py (no class)   0 0 0   100%
hyper_parallel / core / shard / ops / parallel_activation_with_axis.py ActivationWithAxisDistributedOp   26 3 0   88%
hyper_parallel / core / shard / ops / parallel_activation_with_axis.py (no class)   4 0 0   100%
hyper_parallel / core / shard / ops / parallel_argmax_with_value_ops.py ArgMaxWithValueDistributedOp   31 31 0   0%
hyper_parallel / core / shard / ops / parallel_argmax_with_value_ops.py (no class)   3 0 0   100%
hyper_parallel / core / shard / ops / parallel_concat.py ConcatDistributedOp   28 28 0   0%
hyper_parallel / core / shard / ops / parallel_concat.py (no class)   4 0 0   100%
hyper_parallel / core / shard / ops / parallel_cumsum.py CumsumDistributedOp   26 4 0   85%
hyper_parallel / core / shard / ops / parallel_cumsum.py (no class)   4 0 0   100%
hyper_parallel / core / shard / ops / parallel_elementwise.py ElementWiseDistributedOp   201 62 0   69%
hyper_parallel / core / shard / ops / parallel_elementwise.py ElementWiseWithPartialDistributedOp   2 0 0   100%
hyper_parallel / core / shard / ops / parallel_elementwise.py AddDistributedOp   19 1 0   95%
hyper_parallel / core / shard / ops / parallel_elementwise.py (no class)   23 0 0   100%
hyper_parallel / core / shard / ops / parallel_embedding.py EmbeddingDistributedOp   22 1 0   95%
hyper_parallel / core / shard / ops / parallel_embedding.py (no class)   4 0 0   100%
hyper_parallel / core / shard / ops / parallel_expand.py ExpandDistributedOp   41 5 1   88%
hyper_parallel / core / shard / ops / parallel_expand.py ExpandAsDistributedOp   41 7 0   83%
hyper_parallel / core / shard / ops / parallel_expand.py (no class)   6 0 0   100%
hyper_parallel / core / shard / ops / parallel_expand_dims.py ExpandDimsDistributedOp   22 6 0   73%
hyper_parallel / core / shard / ops / parallel_expand_dims.py (no class)   4 0 0   100%
hyper_parallel / core / shard / ops / parallel_flash_attention_score.py ParallelFlashAttention   92 21 0   77%
hyper_parallel / core / shard / ops / parallel_flash_attention_score.py (no class)   19 0 0   100%
hyper_parallel / core / shard / ops / parallel_gather.py IndexSelectDistributedOp   21 21 0   0%
hyper_parallel / core / shard / ops / parallel_gather.py GatherDistributedOp   21 21 0   0%
hyper_parallel / core / shard / ops / parallel_gather.py GatherNdDistributedOp   75 23 0   69%
hyper_parallel / core / shard / ops / parallel_gather.py (no class)   14 0 0   100%
hyper_parallel / core / shard / ops / parallel_isin.py IsinDistributedOp   20 2 0   90%
hyper_parallel / core / shard / ops / parallel_isin.py (no class)   4 0 0   100%
hyper_parallel / core / shard / ops / parallel_masked_scatter.py MaskedScatterDistributedOp   9 1 0   89%
hyper_parallel / core / shard / ops / parallel_masked_scatter.py (no class)   3 0 0   100%
hyper_parallel / core / shard / ops / parallel_matmul.py MatMulExtDistributedOp   26 6 0   77%
hyper_parallel / core / shard / ops / parallel_matmul.py MatMulDistributedOp   33 10 0   70%
hyper_parallel / core / shard / ops / parallel_matmul.py BaseBatchMatMulDistributedOp   30 7 0   77%
hyper_parallel / core / shard / ops / parallel_matmul.py BatchMatMulExtDistributedOp   15 3 0   80%
hyper_parallel / core / shard / ops / parallel_matmul.py BatchMatMulDistributedOp   24 8 0   67%
hyper_parallel / core / shard / ops / parallel_matmul.py LinearDistributedOp   47 10 0   79%
hyper_parallel / core / shard / ops / parallel_matmul.py (no class)   18 0 0   100%
hyper_parallel / core / shard / ops / parallel_multinomial.py MultinomialDistributedOp   15 1 0   93%
hyper_parallel / core / shard / ops / parallel_multinomial.py (no class)   4 0 0   100%
hyper_parallel / core / shard / ops / parallel_new_ones.py NewOnesDistributedOp   12 2 0   83%
hyper_parallel / core / shard / ops / parallel_new_ones.py (no class)   4 0 0   100%
hyper_parallel / core / shard / ops / parallel_norm.py NormDistributedOp   33 33 0   0%
hyper_parallel / core / shard / ops / parallel_norm.py LayerNormDistributedOp   31 5 0   84%
hyper_parallel / core / shard / ops / parallel_norm.py (no class)   6 0 0   100%
hyper_parallel / core / shard / ops / parallel_one_hot_ext.py OneHotExtDistributedOp   149 149 0   0%
hyper_parallel / core / shard / ops / parallel_one_hot_ext.py (no class)   20 0 0   100%
hyper_parallel / core / shard / ops / parallel_ops.py DistributedOp   12 6 0   50%
hyper_parallel / core / shard / ops / parallel_ops.py (no class)   6 0 0   100%
hyper_parallel / core / shard / ops / parallel_ops_register.py (no class)   5 0 0   100%
hyper_parallel / core / shard / ops / parallel_pad.py PadDistributedOp   29 7 0   76%
hyper_parallel / core / shard / ops / parallel_pad.py (no class)   3 0 0   100%
hyper_parallel / core / shard / ops / parallel_reduce.py ReduceExtDistributedOpBase   96 21 0   78%
hyper_parallel / core / shard / ops / parallel_reduce.py SumExtDistributedOp   1 0 0   100%
hyper_parallel / core / shard / ops / parallel_reduce.py MeanExtDistributedOp   1 0 0   100%
hyper_parallel / core / shard / ops / parallel_reduce.py ReduceMaxDistributedOp   1 0 0   100%
hyper_parallel / core / shard / ops / parallel_reduce.py ProdExtDistributedOp   1 0 0   100%
hyper_parallel / core / shard / ops / parallel_reduce.py AllExtDistributedOp   1 0 0   100%
hyper_parallel / core / shard / ops / parallel_reduce.py MaxDistributedOp   22 3 0   86%
hyper_parallel / core / shard / ops / parallel_reduce.py (no class)   32 0 0   100%
hyper_parallel / core / shard / ops / parallel_repeat_interleave.py RepeatInterleaveDistributedOp   29 5 0   83%
hyper_parallel / core / shard / ops / parallel_repeat_interleave.py (no class)   4 0 0   100%
hyper_parallel / core / shard / ops / parallel_reshape.py ReshapeDistributedOp   115 32 0   72%
hyper_parallel / core / shard / ops / parallel_reshape.py (no class)   26 9 0   65%
hyper_parallel / core / shard / ops / parallel_scatter_update.py ScatterUpdateDistributedOp   36 11 0   69%
hyper_parallel / core / shard / ops / parallel_scatter_update.py (no class)   5 0 0   100%
hyper_parallel / core / shard / ops / parallel_slice.py SliceDistributedOp   29 29 0   0%
hyper_parallel / core / shard / ops / parallel_slice.py (no class)   5 0 0   100%
hyper_parallel / core / shard / ops / parallel_slice_ext.py SliceExtDistributedOp   6 6 0   0%
hyper_parallel / core / shard / ops / parallel_slice_ext.py (no class)   3 0 0   100%
hyper_parallel / core / shard / ops / parallel_sort.py SortDistributedOp   22 4 0   82%
hyper_parallel / core / shard / ops / parallel_sort.py (no class)   3 0 0   100%
hyper_parallel / core / shard / ops / parallel_split.py SplitWithSizeDistributedOp   9 9 0   0%
hyper_parallel / core / shard / ops / parallel_split.py SplitWithSizeViewDistributedOp   9 9 0   0%
hyper_parallel / core / shard / ops / parallel_split.py SplitDistributedOp   27 7 0   74%
hyper_parallel / core / shard / ops / parallel_split.py SplitTensorDistributedOp   12 12 0   0%
hyper_parallel / core / shard / ops / parallel_split.py SplitTensorViewDistributedOp   12 2 0   83%
hyper_parallel / core / shard / ops / parallel_split.py (no class)   12 0 0   100%
hyper_parallel / core / shard / ops / parallel_squeeze.py SqueezeDistributedOp   76 20 0   74%
hyper_parallel / core / shard / ops / parallel_squeeze.py (no class)   15 0 0   100%
hyper_parallel / core / shard / ops / parallel_topk.py TopKDistributedOp   15 2 0   87%
hyper_parallel / core / shard / ops / parallel_topk.py (no class)   3 0 0   100%
hyper_parallel / core / shard / ops / parallel_transpose.py TransposeDistributedOp   34 8 0   76%
hyper_parallel / core / shard / ops / parallel_transpose.py (no class)   4 0 0   100%
hyper_parallel / core / shard / ops / parallel_tuple_elementwise.py TupleElementWiseDistributedOp   3 1 0   67%
hyper_parallel / core / shard / ops / parallel_tuple_elementwise.py (no class)   3 0 0   100%
hyper_parallel / core / shard / ops / parallel_unbind.py UnbindDistributedOp   19 1 0   95%
hyper_parallel / core / shard / ops / parallel_unbind.py (no class)   4 0 0   100%
hyper_parallel / core / shard / sharding_plan.py ShardingPlan   0 0 0   100%
hyper_parallel / core / shard / sharding_plan.py (no class)   8 0 0   100%
hyper_parallel / core / tensor_redistribution.py TensorRedistribution   183 32 0   83%
hyper_parallel / core / tensor_redistribution.py (no class)   26 0 0   100%
hyper_parallel / core / utils.py (no class)   14 0 0   100%
hyper_parallel / platform / __init__.py (no class)   1 0 0   100%
hyper_parallel / platform / mindspore / __init__.py (no class)   0 0 0   100%
hyper_parallel / platform / mindspore / custom_pass / __init__.py (no class)   7 0 0   100%
hyper_parallel / platform / mindspore / custom_pass / build_plugin.py (no class)   56 56 0   0%
hyper_parallel / platform / mindspore / dtensor.py DTensorBase   66 19 0   71%
hyper_parallel / platform / mindspore / dtensor.py (no class)   23 0 0   100%
hyper_parallel / platform / mindspore / hsdp / __init__.py (no class)   0 0 0   100%
hyper_parallel / platform / mindspore / hsdp / async_grad_hook.py MindSporeHSDPAsyncGradHook   6 6 0   0%
hyper_parallel / platform / mindspore / hsdp / async_grad_hook.py (no class)   3 0 0   100%
hyper_parallel / platform / mindspore / hsdp / grad_hook.py MindSporeHSDPGradHook   21 8 0   62%
hyper_parallel / platform / mindspore / hsdp / grad_hook.py (no class)   10 0 0   100%
hyper_parallel / platform / mindspore / hsdp / param.py MindSporeHSDPParam   20 0 0   100%
hyper_parallel / platform / mindspore / hsdp / param.py (no class)   9 0 0   100%
hyper_parallel / platform / mindspore / hsdp / scheduler.py MindSporeHSDPScheduler   76 51 0   33%
hyper_parallel / platform / mindspore / hsdp / scheduler.py MindSporeHSDPScheduler._get_parameter_forward_hook.ForwardHookNet   4 4 0   0%
hyper_parallel / platform / mindspore / hsdp / scheduler.py (no class)   29 0 0   100%
hyper_parallel / platform / mindspore / hsdp / state.py MindSporeHSDPState   17 1 0   94%
hyper_parallel / platform / mindspore / hsdp / state.py (no class)   19 0 0   100%
hyper_parallel / platform / mindspore / parameter_init.py (no class)   32 3 0   91%
hyper_parallel / platform / mindspore / pipeline_parallel / __init__.py (no class)   0 0 0   100%
hyper_parallel / platform / mindspore / pipeline_parallel / _utils.py _MicroBatch   41 41 0   0%
hyper_parallel / platform / mindspore / pipeline_parallel / _utils.py (no class)   39 39 0   0%
hyper_parallel / platform / mindspore / pipeline_parallel / stage.py PipelineStageBase   66 66 0   0%
hyper_parallel / platform / mindspore / pipeline_parallel / stage.py (no class)   13 0 0   100%
hyper_parallel / platform / mindspore / platform.py MindSporePlatform   179 66 0   63%
hyper_parallel / platform / mindspore / platform.py (no class)   168 0 0   100%
hyper_parallel / platform / mindspore / platform_graph.py MindSporeGraphPlatform   10 10 0   0%
hyper_parallel / platform / mindspore / platform_graph.py (no class)   17 0 0   100%
hyper_parallel / platform / platform.py PlatformType   0 0 0   100%
hyper_parallel / platform / platform.py Platform   98 65 0   34%
hyper_parallel / platform / platform.py (no class)   161 4 0   98%
hyper_parallel / platform / torch / __init__.py (no class)   0 0 0   100%
hyper_parallel / platform / torch / activation_checkpoint / __init__.py (no class)   1 0 0   100%
hyper_parallel / platform / torch / activation_checkpoint / activation_swap.py ActivationPolicy   0 0 0   100%
hyper_parallel / platform / torch / activation_checkpoint / activation_swap.py AsyncSaveOnCpu   30 8 0   73%
hyper_parallel / platform / torch / activation_checkpoint / activation_swap.py ActivationWrapper   15 9 0   40%
hyper_parallel / platform / torch / activation_checkpoint / activation_swap.py SwapWrapper   4 0 0   100%
hyper_parallel / platform / torch / activation_checkpoint / activation_swap.py (no class)   37 0 0   100%
hyper_parallel / platform / torch / activation_checkpoint / sac.py _VersionWrapper   6 4 0   33%
hyper_parallel / platform / torch / activation_checkpoint / sac.py SelectiveCheckpointContext   1 0 0   100%
hyper_parallel / platform / torch / activation_checkpoint / sac.py _CachingTorchDispatchMode   25 3 0   88%
hyper_parallel / platform / torch / activation_checkpoint / sac.py _CachedTorchDispatchMode   26 23 0   12%
hyper_parallel / platform / torch / activation_checkpoint / sac.py (no class)   47 11 0   77%
hyper_parallel / platform / torch / dtensor.py DTensorBase   76 51 0   33%
hyper_parallel / platform / torch / dtensor.py (no class)   45 0 0   100%
hyper_parallel / platform / torch / fully_shard / __init__.py (no class)   0 0 0   100%
hyper_parallel / platform / torch / fully_shard / async_grad_hook.py TorchHSDPAsyncGradHook   7 7 0   0%
hyper_parallel / platform / torch / fully_shard / async_grad_hook.py (no class)   3 3 0   0%
hyper_parallel / platform / torch / fully_shard / grad_hook.py TorchHSDPGradHook   7 7 0   0%
hyper_parallel / platform / torch / fully_shard / grad_hook.py (no class)   3 3 0   0%
hyper_parallel / platform / torch / fully_shard / hook_function.py PostBackwardFunction   30 11 0   63%
hyper_parallel / platform / torch / fully_shard / hook_function.py (no class)   8 0 0   100%
hyper_parallel / platform / torch / fully_shard / param.py TorchHSDPParamV2   278 161 0   42%
hyper_parallel / platform / torch / fully_shard / param.py (no class)   58 1 0   98%
hyper_parallel / platform / torch / fully_shard / scheduler.py TorchHSDPSchedulerV2   43 12 0   72%
hyper_parallel / platform / torch / fully_shard / scheduler.py (no class)   20 0 0   100%
hyper_parallel / platform / torch / fully_shard / state.py TorchHSDPStateV2   96 62 0   35%
hyper_parallel / platform / torch / fully_shard / state.py (no class)   22 0 0   100%
hyper_parallel / platform / torch / fully_shard / utils.py MixedPrecisionPolicy   0 0 0   100%
hyper_parallel / platform / torch / fully_shard / utils.py OffloadPolicy   0 0 0   100%
hyper_parallel / platform / torch / fully_shard / utils.py CPUOffloadPolicy   0 0 0   100%
hyper_parallel / platform / torch / fully_shard / utils.py DataParallelMeshInfo   2 1 0   50%
hyper_parallel / platform / torch / fully_shard / utils.py FSDPMeshInfo   6 1 0   83%
hyper_parallel / platform / torch / fully_shard / utils.py DDPMeshInfo   6 6 0   0%
hyper_parallel / platform / torch / fully_shard / utils.py HSDPMeshInfo   1 1 0   0%
hyper_parallel / platform / torch / fully_shard / utils.py (no class)   31 0 0   100%
hyper_parallel / platform / torch / function_override.py DTensorBackwardHookFunction   25 9 0   64%
hyper_parallel / platform / torch / function_override.py ExtendBackwardHook   5 3 0   40%
hyper_parallel / platform / torch / function_override.py (no class)   14 0 0   100%
hyper_parallel / platform / torch / group_utils.py (no class)   86 18 0   79%
hyper_parallel / platform / torch / hsdp / __init__.py (no class)   0 0 0   100%
hyper_parallel / platform / torch / hsdp / async_grad_hook.py TorchHSDPAsyncGradHook   7 4 0   43%
hyper_parallel / platform / torch / hsdp / async_grad_hook.py (no class)   3 0 0   100%
hyper_parallel / platform / torch / hsdp / grad_hook.py TorchHSDPGradHook   7 4 0   43%
hyper_parallel / platform / torch / hsdp / grad_hook.py (no class)   3 0 0   100%
hyper_parallel / platform / torch / hsdp / hook_function.py PostBackwardFunction   29 7 0   76%
hyper_parallel / platform / torch / hsdp / hook_function.py (no class)   8 0 0   100%
hyper_parallel / platform / torch / hsdp / param.py TorchHSDPParam   14 0 0   100%
hyper_parallel / platform / torch / hsdp / param.py (no class)   8 0 0   100%
hyper_parallel / platform / torch / hsdp / scheduler.py TorchHSDPScheduler   39 11 0   72%
hyper_parallel / platform / torch / hsdp / scheduler.py (no class)   21 0 0   100%
hyper_parallel / platform / torch / hsdp / state.py TorchHSDPState   9 1 0   89%
hyper_parallel / platform / torch / hsdp / state.py (no class)   5 0 0   100%
hyper_parallel / platform / torch / pipeline_parallel / __init__.py (no class)   0 0 0   100%
hyper_parallel / platform / torch / pipeline_parallel / _utils.py _MicroBatch   42 10 0   76%
hyper_parallel / platform / torch / pipeline_parallel / _utils.py (no class)   7 0 0   100%
hyper_parallel / platform / torch / pipeline_parallel / stage.py PipelineStageBase   61 10 0   84%
hyper_parallel / platform / torch / pipeline_parallel / stage.py (no class)   14 0 0   100%
hyper_parallel / platform / torch / platform.py TorchPlatform   205 67 0   67%
hyper_parallel / platform / torch / platform.py (no class)   167 1 0   99%
Total     11439 3642 1   68%

No items found using the specified filter.