Coverage report: 65%

Files Functions Classes

coverage.py v7.13.1, created at 2026-03-16 05:20 +0800

File class   statements missing excluded   coverage
hyper_parallel / __init__.py (no class)   13 0 0   100%
hyper_parallel / collectives / __init__.py (no class)   0 0 0   100%
hyper_parallel / collectives / cc.py (no class)   18 1 0   94%
hyper_parallel / core / __init__.py (no class)   0 0 0   100%
hyper_parallel / core / activation_checkpoint / __init__.py (no class)   2 0 0   100%
hyper_parallel / core / activation_checkpoint / activation_checkpoint.py CheckpointPolicy   0 0 0   100%
hyper_parallel / core / activation_checkpoint / activation_checkpoint.py (no class)   17 0 0   100%
hyper_parallel / core / activation_checkpoint / swap.py SwapTensor   54 38 0   30%
hyper_parallel / core / activation_checkpoint / swap.py Storage   30 14 0   53%
hyper_parallel / core / activation_checkpoint / swap.py SwapGroup   37 16 0   57%
hyper_parallel / core / activation_checkpoint / swap.py SwapManager   82 22 0   73%
hyper_parallel / core / activation_checkpoint / swap.py (no class)   49 0 0   100%
hyper_parallel / core / checkpoint / __init__.py (no class)   11 0 0   100%
hyper_parallel / core / checkpoint / api.py (no class)   83 11 0   87%
hyper_parallel / core / checkpoint / filesystem_storage.py FileSystemWriter   66 8 0   88%
hyper_parallel / core / checkpoint / filesystem_storage.py FileSystemReader   44 11 0   75%
hyper_parallel / core / checkpoint / filesystem_storage.py (no class)   62 4 0   94%
hyper_parallel / core / checkpoint / layout.py (no class)   46 2 0   96%
hyper_parallel / core / checkpoint / loader.py (no class)   11 1 0   91%
hyper_parallel / core / checkpoint / metadata.py MetadataIndex   0 0 0   100%
hyper_parallel / core / checkpoint / metadata.py ChunkStorageMetadata   0 0 0   100%
hyper_parallel / core / checkpoint / metadata.py TensorProperties   0 0 0   100%
hyper_parallel / core / checkpoint / metadata.py BytesStorageMetadata   0 0 0   100%
hyper_parallel / core / checkpoint / metadata.py TensorStorageMetadata   0 0 0   100%
hyper_parallel / core / checkpoint / metadata.py Metadata   0 0 0   100%
hyper_parallel / core / checkpoint / metadata.py (no class)   29 0 0   100%
hyper_parallel / core / checkpoint / planner.py WriteItemType   0 0 0   100%
hyper_parallel / core / checkpoint / planner.py LoadItemType   0 0 0   100%
hyper_parallel / core / checkpoint / planner.py WriteItem   21 2 0   90%
hyper_parallel / core / checkpoint / planner.py ReadItem   0 0 0   100%
hyper_parallel / core / checkpoint / planner.py SavePlan   0 0 0   100%
hyper_parallel / core / checkpoint / planner.py LoadPlan   0 0 0   100%
hyper_parallel / core / checkpoint / planner.py SavePlanner   0 0 0   100%
hyper_parallel / core / checkpoint / planner.py LoadPlanner   0 0 0   100%
hyper_parallel / core / checkpoint / planner.py (no class)   62 0 0   100%
hyper_parallel / core / checkpoint / reshard.py ReshardHandler   80 2 0   98%
hyper_parallel / core / checkpoint / reshard.py (no class)   79 6 0   92%
hyper_parallel / core / checkpoint / saver.py (no class)   11 1 0   91%
hyper_parallel / core / checkpoint / standard_planner.py StandardSavePlanner   109 9 0   92%
hyper_parallel / core / checkpoint / standard_planner.py StandardLoadPlanner   58 13 0   78%
hyper_parallel / core / checkpoint / standard_planner.py (no class)   45 1 0   98%
hyper_parallel / core / checkpoint / storage.py StorageInfo   0 0 0   100%
hyper_parallel / core / checkpoint / storage.py WriteResult   0 0 0   100%
hyper_parallel / core / checkpoint / storage.py StorageWriter   0 0 0   100%
hyper_parallel / core / checkpoint / storage.py StorageReader   0 0 0   100%
hyper_parallel / core / checkpoint / storage.py (no class)   42 0 0   100%
hyper_parallel / core / checkpoint / util.py (no class)   82 12 0   85%
hyper_parallel / core / device_mesh.py DeviceMesh   344 103 0   70%
hyper_parallel / core / device_mesh.py (no class)   98 3 0   97%
hyper_parallel / core / dtensor.py SkipDTensorDispatch   4 0 0   100%
hyper_parallel / core / dtensor.py DTensor   39 8 0   79%
hyper_parallel / core / dtensor.py (no class)   71 0 0   100%
hyper_parallel / core / fully_shard / __init__.py (no class)   2 0 0   100%
hyper_parallel / core / fully_shard / api.py _UnshardHandle   4 0 0   100%
hyper_parallel / core / fully_shard / api.py HSDPModule   112 42 0   62%
hyper_parallel / core / fully_shard / api.py (no class)   80 24 0   70%
hyper_parallel / core / fully_shard / hsdp_param.py HSDPParamV2   23 23 0   0%
hyper_parallel / core / fully_shard / hsdp_param.py (no class)   29 0 0   100%
hyper_parallel / core / fully_shard / hsdp_scheduler.py HSDPSchedulerContext   1 0 0   100%
hyper_parallel / core / fully_shard / hsdp_scheduler.py HSDPSchedulerV2   64 12 0   81%
hyper_parallel / core / fully_shard / hsdp_scheduler.py (no class)   21 0 0   100%
hyper_parallel / core / fully_shard / hsdp_state.py HSDPState   31 4 0   87%
hyper_parallel / core / fully_shard / hsdp_state.py (no class)   11 0 0   100%
hyper_parallel / core / fully_shard / hsdp_utils.py HSDPConfigV2   7 0 0   100%
hyper_parallel / core / fully_shard / hsdp_utils.py ShardedState   0 0 0   100%
hyper_parallel / core / fully_shard / hsdp_utils.py FSDPSchedulerState   0 0 0   100%
hyper_parallel / core / fully_shard / hsdp_utils.py ParamModuleInfo   0 0 0   100%
hyper_parallel / core / fully_shard / hsdp_utils.py (no class)   54 7 0   87%
hyper_parallel / core / fully_shard / utils.py MixedPrecisionPolicy   0 0 0   100%
hyper_parallel / core / fully_shard / utils.py OffloadPolicy   0 0 0   100%
hyper_parallel / core / fully_shard / utils.py CPUOffloadPolicy   0 0 0   100%
hyper_parallel / core / fully_shard / utils.py DataParallelMeshInfo   2 1 0   50%
hyper_parallel / core / fully_shard / utils.py FSDPMeshInfo   6 1 0   83%
hyper_parallel / core / fully_shard / utils.py DDPMeshInfo   6 1 0   83%
hyper_parallel / core / fully_shard / utils.py HSDPMeshInfo   1 0 0   100%
hyper_parallel / core / fully_shard / utils.py (no class)   34 0 0   100%
hyper_parallel / core / hsdp / __init__.py (no class)   2 0 0   100%
hyper_parallel / core / hsdp / api.py HSDPCell   35 30 0   14%
hyper_parallel / core / hsdp / api.py (no class)   66 18 0   73%
hyper_parallel / core / hsdp / hsdp_async_grad_hook.py HSDPAsyncGradHook   78 78 0   0%
hyper_parallel / core / hsdp / hsdp_async_grad_hook.py (no class)   10 0 0   100%
hyper_parallel / core / hsdp / hsdp_grad_buffer.py HSDPGradBuffer   145 145 0   0%
hyper_parallel / core / hsdp / hsdp_grad_buffer.py (no class)   19 0 0   100%
hyper_parallel / core / hsdp / hsdp_grad_hook.py HSDPGradHook   102 63 0   38%
hyper_parallel / core / hsdp / hsdp_grad_hook.py (no class)   11 0 0   100%
hyper_parallel / core / hsdp / hsdp_param.py HSDPParam   217 52 0   76%
hyper_parallel / core / hsdp / hsdp_param.py (no class)   20 0 0   100%
hyper_parallel / core / hsdp / hsdp_param_buffer.py HSDPParamBuffer   57 57 0   0%
hyper_parallel / core / hsdp / hsdp_param_buffer.py (no class)   9 0 0   100%
hyper_parallel / core / hsdp / hsdp_scheduler.py FSDPSchedulerState   0 0 0   100%
hyper_parallel / core / hsdp / hsdp_scheduler.py HSDPScheduler   62 41 0   34%
hyper_parallel / core / hsdp / hsdp_scheduler.py (no class)   27 0 0   100%
hyper_parallel / core / hsdp / hsdp_state.py HSDPState   98 83 0   15%
hyper_parallel / core / hsdp / hsdp_state.py (no class)   13 0 0   100%
hyper_parallel / core / hsdp / hsdp_utils.py OptimizerLevel   0 0 0   100%
hyper_parallel / core / hsdp / hsdp_utils.py GroupInfo   3 0 0   100%
hyper_parallel / core / hsdp / hsdp_utils.py HSDPConfig   11 0 0   100%
hyper_parallel / core / hsdp / hsdp_utils.py (no class)   9 0 0   100%
hyper_parallel / core / init_weights.py (no class)   11 2 0   82%
hyper_parallel / core / layout.py Layout   226 46 0   80%
hyper_parallel / core / layout.py (no class)   116 0 0   100%
hyper_parallel / core / parameter_init.py (no class)   4 0 0   100%
hyper_parallel / core / pipeline_parallel / __init__.py (no class)   4 0 0   100%
hyper_parallel / core / pipeline_parallel / scheduler.py MetaStepType   0 0 0   100%
hyper_parallel / core / pipeline_parallel / scheduler.py MetaStep   16 10 0   38%
hyper_parallel / core / pipeline_parallel / scheduler.py PipelineScheduleRuntime   92 15 0   84%
hyper_parallel / core / pipeline_parallel / scheduler.py ScheduleGPipe   17 17 0   0%
hyper_parallel / core / pipeline_parallel / scheduler.py Schedule1F1B   44 44 0   0%
hyper_parallel / core / pipeline_parallel / scheduler.py ScheduleInterleaved1F1B   63 8 0   87%
hyper_parallel / core / pipeline_parallel / scheduler.py (no class)   244 138 0   43%
hyper_parallel / core / pipeline_parallel / stage.py SharedParameterInfo   11 11 0   0%
hyper_parallel / core / pipeline_parallel / stage.py PipelineStage   183 64 0   65%
hyper_parallel / core / pipeline_parallel / stage.py (no class)   35 0 0   100%
hyper_parallel / core / pipeline_parallel / utils.py BatchDimSpec   11 11 0   0%
hyper_parallel / core / pipeline_parallel / utils.py _RecvInfo   5 0 0   100%
hyper_parallel / core / pipeline_parallel / utils.py (no class)   17 0 0   100%
hyper_parallel / core / placement_types.py Placement   4 3 0   25%
hyper_parallel / core / placement_types.py Shard   11 4 0   64%
hyper_parallel / core / placement_types.py Replicate   4 2 0   50%
hyper_parallel / core / placement_types.py Partial   13 7 0   46%
hyper_parallel / core / placement_types.py (no class)   29 0 0   100%
hyper_parallel / core / random.py _PhiloxState   8 3 0   62%
hyper_parallel / core / random.py _RNGStateTracker   9 2 0   78%
hyper_parallel / core / random.py OffsetBasedRNGTracker   39 3 0   92%
hyper_parallel / core / random.py (no class)   128 17 0   87%
hyper_parallel / core / redistribute_infer.py Status   0 0 0   100%
hyper_parallel / core / redistribute_infer.py TensorMap   14 8 0   43%
hyper_parallel / core / redistribute_infer.py DevMat   47 25 0   47%
hyper_parallel / core / redistribute_infer.py RedistributionOperatorInfer   236 94 0   60%
hyper_parallel / core / redistribute_infer.py (no class)   34 0 0   100%
hyper_parallel / core / shard / __init__.py (no class)   0 0 0   100%
hyper_parallel / core / shard / _op_dispatch.py LayoutCacheKey   9 1 0   89%
hyper_parallel / core / shard / _op_dispatch.py LayoutCacheManager   9 0 0   100%
hyper_parallel / core / shard / _op_dispatch.py OpDispatcher   408 202 0   50%
hyper_parallel / core / shard / _op_dispatch.py (no class)   52 0 0   100%
hyper_parallel / core / shard / api.py parallelize_value_and_grad.CellWrapper   4 1 0   75%
hyper_parallel / core / shard / api.py (no class)   301 101 0   66%
hyper_parallel / core / shard / local_func.py (no class)   48 8 0   83%
hyper_parallel / core / shard / ops / __init__.py (no class)   0 0 0   100%
hyper_parallel / core / shard / ops / parallel_activation_with_axis.py ActivationWithAxisDistributedOp   26 26 0   0%
hyper_parallel / core / shard / ops / parallel_activation_with_axis.py (no class)   4 0 0   100%
hyper_parallel / core / shard / ops / parallel_argmax_with_value_ops.py ArgMaxWithValueDistributedOp   31 31 0   0%
hyper_parallel / core / shard / ops / parallel_argmax_with_value_ops.py (no class)   3 0 0   100%
hyper_parallel / core / shard / ops / parallel_concat.py ConcatDistributedOp   28 28 0   0%
hyper_parallel / core / shard / ops / parallel_concat.py (no class)   4 0 0   100%
hyper_parallel / core / shard / ops / parallel_cumsum.py CumsumDistributedOp   26 15 0   42%
hyper_parallel / core / shard / ops / parallel_cumsum.py (no class)   4 0 0   100%
hyper_parallel / core / shard / ops / parallel_elementwise.py ElementWiseDistributedOp   201 63 0   69%
hyper_parallel / core / shard / ops / parallel_elementwise.py ElementWiseWithPartialDistributedOp   2 0 0   100%
hyper_parallel / core / shard / ops / parallel_elementwise.py AddDistributedOp   19 3 0   84%
hyper_parallel / core / shard / ops / parallel_elementwise.py (no class)   23 0 0   100%
hyper_parallel / core / shard / ops / parallel_embedding.py EmbeddingDistributedOp   22 22 0   0%
hyper_parallel / core / shard / ops / parallel_embedding.py (no class)   4 0 0   100%
hyper_parallel / core / shard / ops / parallel_expand.py ExpandDistributedOp   41 5 1   88%
hyper_parallel / core / shard / ops / parallel_expand.py ExpandAsDistributedOp   41 7 0   83%
hyper_parallel / core / shard / ops / parallel_expand.py (no class)   6 0 0   100%
hyper_parallel / core / shard / ops / parallel_expand_dims.py ExpandDimsDistributedOp   22 6 0   73%
hyper_parallel / core / shard / ops / parallel_expand_dims.py (no class)   4 0 0   100%
hyper_parallel / core / shard / ops / parallel_flash_attention_score.py ParallelFlashAttention   92 21 0   77%
hyper_parallel / core / shard / ops / parallel_flash_attention_score.py (no class)   19 0 0   100%
hyper_parallel / core / shard / ops / parallel_gather.py IndexSelectDistributedOp   21 21 0   0%
hyper_parallel / core / shard / ops / parallel_gather.py GatherDistributedOp   21 21 0   0%
hyper_parallel / core / shard / ops / parallel_gather.py GatherNdDistributedOp   75 23 0   69%
hyper_parallel / core / shard / ops / parallel_gather.py (no class)   14 0 0   100%
hyper_parallel / core / shard / ops / parallel_isin.py IsinDistributedOp   20 20 0   0%
hyper_parallel / core / shard / ops / parallel_isin.py (no class)   4 0 0   100%
hyper_parallel / core / shard / ops / parallel_masked_scatter.py MaskedScatterDistributedOp   9 9 0   0%
hyper_parallel / core / shard / ops / parallel_masked_scatter.py (no class)   3 0 0   100%
hyper_parallel / core / shard / ops / parallel_matmul.py MatMulExtDistributedOp   26 6 0   77%
hyper_parallel / core / shard / ops / parallel_matmul.py MatMulDistributedOp   33 12 0   64%
hyper_parallel / core / shard / ops / parallel_matmul.py BaseBatchMatMulDistributedOp   30 7 0   77%
hyper_parallel / core / shard / ops / parallel_matmul.py BatchMatMulExtDistributedOp   15 3 0   80%
hyper_parallel / core / shard / ops / parallel_matmul.py BatchMatMulDistributedOp   24 8 0   67%
hyper_parallel / core / shard / ops / parallel_matmul.py LinearDistributedOp   47 47 0   0%
hyper_parallel / core / shard / ops / parallel_matmul.py (no class)   18 0 0   100%
hyper_parallel / core / shard / ops / parallel_ms_flash_attention_score.py FlashAttentionScoreDistributedOp   433 126 0   71%
hyper_parallel / core / shard / ops / parallel_ms_flash_attention_score.py (no class)   69 2 0   97%
hyper_parallel / core / shard / ops / parallel_multinomial.py MultinomialDistributedOp   15 8 0   47%
hyper_parallel / core / shard / ops / parallel_multinomial.py (no class)   4 0 0   100%
hyper_parallel / core / shard / ops / parallel_new_ones.py NewOnesDistributedOp   12 3 0   75%
hyper_parallel / core / shard / ops / parallel_new_ones.py (no class)   4 0 0   100%
hyper_parallel / core / shard / ops / parallel_norm.py NormDistributedOp   33 33 0   0%
hyper_parallel / core / shard / ops / parallel_norm.py LayerNormDistributedOp   31 16 0   48%
hyper_parallel / core / shard / ops / parallel_norm.py (no class)   6 0 0   100%
hyper_parallel / core / shard / ops / parallel_npu_flash_attention_score.py FlashAttentionScoreDistributedOp   439 156 0   64%
hyper_parallel / core / shard / ops / parallel_npu_flash_attention_score.py (no class)   53 0 0   100%
hyper_parallel / core / shard / ops / parallel_one_hot_ext.py OneHotExtDistributedOp   149 32 0   79%
hyper_parallel / core / shard / ops / parallel_one_hot_ext.py (no class)   20 0 0   100%
hyper_parallel / core / shard / ops / parallel_ops.py DistributedOp   12 6 0   50%
hyper_parallel / core / shard / ops / parallel_ops.py (no class)   6 0 0   100%
hyper_parallel / core / shard / ops / parallel_ops_register.py (no class)   5 0 0   100%
hyper_parallel / core / shard / ops / parallel_pad.py PadDistributedOp   29 29 0   0%
hyper_parallel / core / shard / ops / parallel_pad.py (no class)   3 0 0   100%
hyper_parallel / core / shard / ops / parallel_reduce.py ReduceExtDistributedOpBase   96 22 0   77%
hyper_parallel / core / shard / ops / parallel_reduce.py SumExtDistributedOp   1 0 0   100%
hyper_parallel / core / shard / ops / parallel_reduce.py MeanExtDistributedOp   1 0 0   100%
hyper_parallel / core / shard / ops / parallel_reduce.py ReduceMaxDistributedOp   1 0 0   100%
hyper_parallel / core / shard / ops / parallel_reduce.py ProdExtDistributedOp   1 0 0   100%
hyper_parallel / core / shard / ops / parallel_reduce.py AllExtDistributedOp   1 0 0   100%
hyper_parallel / core / shard / ops / parallel_reduce.py MaxDistributedOp   22 21 0   5%
hyper_parallel / core / shard / ops / parallel_reduce.py (no class)   32 0 0   100%
hyper_parallel / core / shard / ops / parallel_repeat_interleave.py RepeatInterleaveDistributedOp   29 29 0   0%
hyper_parallel / core / shard / ops / parallel_repeat_interleave.py (no class)   4 0 0   100%
hyper_parallel / core / shard / ops / parallel_reshape.py ReshapeDistributedOp   115 115 0   0%
hyper_parallel / core / shard / ops / parallel_reshape.py (no class)   26 14 0   46%
hyper_parallel / core / shard / ops / parallel_scaled_dot_product_attention.py ScaledDotProductAttentionDistributedOp   155 95 0   39%
hyper_parallel / core / shard / ops / parallel_scaled_dot_product_attention.py (no class)   22 0 0   100%
hyper_parallel / core / shard / ops / parallel_scatter.py ScatterDistributedOp   29 29 0   0%
hyper_parallel / core / shard / ops / parallel_scatter.py (no class)   3 0 0   100%
hyper_parallel / core / shard / ops / parallel_scatter_update.py ScatterUpdateDistributedOp   36 11 0   69%
hyper_parallel / core / shard / ops / parallel_scatter_update.py (no class)   5 0 0   100%
hyper_parallel / core / shard / ops / parallel_slice.py SliceDistributedOp   29 29 0   0%
hyper_parallel / core / shard / ops / parallel_slice.py (no class)   5 0 0   100%
hyper_parallel / core / shard / ops / parallel_slice_ext.py SliceExtDistributedOp   6 6 0   0%
hyper_parallel / core / shard / ops / parallel_slice_ext.py (no class)   3 0 0   100%
hyper_parallel / core / shard / ops / parallel_sort.py SortDistributedOp   22 22 0   0%
hyper_parallel / core / shard / ops / parallel_sort.py (no class)   3 0 0   100%
hyper_parallel / core / shard / ops / parallel_split.py SplitWithSizeDistributedOp   9 9 0   0%
hyper_parallel / core / shard / ops / parallel_split.py SplitWithSizeViewDistributedOp   9 9 0   0%
hyper_parallel / core / shard / ops / parallel_split.py SplitDistributedOp   27 27 0   0%
hyper_parallel / core / shard / ops / parallel_split.py SplitTensorDistributedOp   12 12 0   0%
hyper_parallel / core / shard / ops / parallel_split.py SplitTensorViewDistributedOp   12 12 0   0%
hyper_parallel / core / shard / ops / parallel_split.py (no class)   12 0 0   100%
hyper_parallel / core / shard / ops / parallel_squeeze.py SqueezeDistributedOp   76 42 0   45%
hyper_parallel / core / shard / ops / parallel_squeeze.py (no class)   15 0 0   100%
hyper_parallel / core / shard / ops / parallel_topk.py TopKDistributedOp   15 15 0   0%
hyper_parallel / core / shard / ops / parallel_topk.py (no class)   3 0 0   100%
hyper_parallel / core / shard / ops / parallel_transpose.py TransposeDistributedOp   34 7 0   79%
hyper_parallel / core / shard / ops / parallel_transpose.py (no class)   4 0 0   100%
hyper_parallel / core / shard / ops / parallel_tuple_elementwise.py TupleElementWiseDistributedOp   3 3 0   0%
hyper_parallel / core / shard / ops / parallel_tuple_elementwise.py (no class)   3 0 0   100%
hyper_parallel / core / shard / ops / parallel_unbind.py UnbindDistributedOp   19 19 0   0%
hyper_parallel / core / shard / ops / parallel_unbind.py (no class)   4 0 0   100%
hyper_parallel / core / shard / sharding_plan.py ShardingPlan   0 0 0   100%
hyper_parallel / core / shard / sharding_plan.py (no class)   8 0 0   100%
hyper_parallel / core / tensor_redistribution.py TensorRedistribution   183 35 0   81%
hyper_parallel / core / tensor_redistribution.py (no class)   26 0 0   100%
hyper_parallel / core / utils / __init__.py (no class)   3 0 0   100%
hyper_parallel / core / utils / clip_grad.py (no class)   6 0 0   100%
hyper_parallel / core / utils / shape_utils.py (no class)   14 0 0   100%
hyper_parallel / platform / __init__.py (no class)   1 0 0   100%
hyper_parallel / platform / mindspore / __init__.py (no class)   0 0 0   100%
hyper_parallel / platform / mindspore / custom_pass / __init__.py (no class)   7 7 0   0%
hyper_parallel / platform / mindspore / custom_pass / build_plugin.py (no class)   56 56 0   0%
hyper_parallel / platform / mindspore / dtensor.py DTensorBase   66 20 0   70%
hyper_parallel / platform / mindspore / dtensor.py (no class)   23 0 0   100%
hyper_parallel / platform / mindspore / fully_shard / __init__.py (no class)   0 0 0   100%
hyper_parallel / platform / mindspore / fully_shard / hook_function.py PostBackwardFunction   30 9 0   70%
hyper_parallel / platform / mindspore / fully_shard / hook_function.py (no class)   8 0 0   100%
hyper_parallel / platform / mindspore / fully_shard / param.py MindSporeHSDPParamV2   231 75 0   68%
hyper_parallel / platform / mindspore / fully_shard / param.py (no class)   78 8 0   90%
hyper_parallel / platform / mindspore / fully_shard / scheduler.py MindSporeHSDPSchedulerV2   43 4 0   91%
hyper_parallel / platform / mindspore / fully_shard / scheduler.py (no class)   20 0 0   100%
hyper_parallel / platform / mindspore / fully_shard / state.py MindSporeHSDPStateV2   97 27 0   72%
hyper_parallel / platform / mindspore / fully_shard / state.py (no class)   24 1 0   96%
hyper_parallel / platform / mindspore / hsdp / __init__.py (no class)   0 0 0   100%
hyper_parallel / platform / mindspore / hsdp / async_grad_hook.py MindSporeHSDPAsyncGradHook   6 6 0   0%
hyper_parallel / platform / mindspore / hsdp / async_grad_hook.py (no class)   3 0 0   100%
hyper_parallel / platform / mindspore / hsdp / grad_hook.py MindSporeHSDPGradHook   21 10 0   52%
hyper_parallel / platform / mindspore / hsdp / grad_hook.py (no class)   10 0 0   100%
hyper_parallel / platform / mindspore / hsdp / param.py MindSporeHSDPParam   20 5 0   75%
hyper_parallel / platform / mindspore / hsdp / param.py (no class)   9 0 0   100%
hyper_parallel / platform / mindspore / hsdp / scheduler.py MindSporeHSDPScheduler   76 64 0   16%
hyper_parallel / platform / mindspore / hsdp / scheduler.py MindSporeHSDPScheduler._get_parameter_forward_hook.ForwardHookNet   4 4 0   0%
hyper_parallel / platform / mindspore / hsdp / scheduler.py (no class)   29 0 0   100%
hyper_parallel / platform / mindspore / hsdp / state.py MindSporeHSDPState   17 7 0   59%
hyper_parallel / platform / mindspore / hsdp / state.py (no class)   19 0 0   100%
hyper_parallel / platform / mindspore / init_weights.py (no class)   31 26 0   16%
hyper_parallel / platform / mindspore / parameter_init.py (no class)   32 3 0   91%
hyper_parallel / platform / mindspore / pipeline_parallel / __init__.py (no class)   0 0 0   100%
hyper_parallel / platform / mindspore / pipeline_parallel / _utils.py _MicroBatch   41 41 0   0%
hyper_parallel / platform / mindspore / pipeline_parallel / _utils.py (no class)   39 39 0   0%
hyper_parallel / platform / mindspore / pipeline_parallel / stage.py PipelineStageBase   66 66 0   0%
hyper_parallel / platform / mindspore / pipeline_parallel / stage.py (no class)   13 0 0   100%
hyper_parallel / platform / mindspore / platform.py MindSporePlatform   213 107 0   50%
hyper_parallel / platform / mindspore / platform.py (no class)   183 0 0   100%
hyper_parallel / platform / mindspore / platform_graph.py MindSporeGraphPlatform   10 10 0   0%
hyper_parallel / platform / mindspore / platform_graph.py (no class)   17 0 0   100%
hyper_parallel / platform / platform.py PlatformType   0 0 0   100%
hyper_parallel / platform / platform.py Platform   110 96 0   13%
hyper_parallel / platform / platform.py (no class)   174 4 0   98%
hyper_parallel / platform / torch / __init__.py (no class)   0 0 0   100%
hyper_parallel / platform / torch / activation_checkpoint / __init__.py (no class)   1 0 0   100%
hyper_parallel / platform / torch / activation_checkpoint / activation_swap.py ActivationPolicy   0 0 0   100%
hyper_parallel / platform / torch / activation_checkpoint / activation_swap.py AsyncSaveOnCpu   30 9 0   70%
hyper_parallel / platform / torch / activation_checkpoint / activation_swap.py ActivationWrapper   15 15 0   0%
hyper_parallel / platform / torch / activation_checkpoint / activation_swap.py SwapWrapper   4 4 0   0%
hyper_parallel / platform / torch / activation_checkpoint / activation_swap.py (no class)   37 2 0   95%
hyper_parallel / platform / torch / activation_checkpoint / sac.py _VersionWrapper   6 4 0   33%
hyper_parallel / platform / torch / activation_checkpoint / sac.py SelectiveCheckpointContext   1 0 0   100%
hyper_parallel / platform / torch / activation_checkpoint / sac.py _CachingTorchDispatchMode   25 3 0   88%
hyper_parallel / platform / torch / activation_checkpoint / sac.py _CachedTorchDispatchMode   26 23 0   12%
hyper_parallel / platform / torch / activation_checkpoint / sac.py (no class)   47 11 0   77%
hyper_parallel / platform / torch / clip_grad.py (no class)   243 58 0   76%
hyper_parallel / platform / torch / dtensor.py DTensorBase   76 50 0   34%
hyper_parallel / platform / torch / dtensor.py (no class)   46 0 0   100%
hyper_parallel / platform / torch / fully_shard / __init__.py (no class)   0 0 0   100%
hyper_parallel / platform / torch / fully_shard / hook_function.py PostBackwardFunction   30 11 0   63%
hyper_parallel / platform / torch / fully_shard / hook_function.py (no class)   8 0 0   100%
hyper_parallel / platform / torch / fully_shard / param.py TorchHSDPParamV2   264 93 0   65%
hyper_parallel / platform / torch / fully_shard / param.py (no class)   57 1 0   98%
hyper_parallel / platform / torch / fully_shard / scheduler.py TorchHSDPSchedulerV2   57 23 0   60%
hyper_parallel / platform / torch / fully_shard / scheduler.py (no class)   22 0 0   100%
hyper_parallel / platform / torch / fully_shard / state.py TorchHSDPStateV2   109 50 0   54%
hyper_parallel / platform / torch / fully_shard / state.py (no class)   24 3 0   88%
hyper_parallel / platform / torch / fully_shard / state_dict_utils.py (no class)   45 45 0   0%
hyper_parallel / platform / torch / function_override.py DTensorBackwardHookFunction   25 9 0   64%
hyper_parallel / platform / torch / function_override.py ExtendBackwardHook   5 3 0   40%
hyper_parallel / platform / torch / function_override.py (no class)   14 0 0   100%
hyper_parallel / platform / torch / group_utils.py (no class)   88 18 0   80%
hyper_parallel / platform / torch / hsdp / __init__.py (no class)   0 0 0   100%
hyper_parallel / platform / torch / hsdp / async_grad_hook.py TorchHSDPAsyncGradHook   7 7 0   0%
hyper_parallel / platform / torch / hsdp / async_grad_hook.py (no class)   3 3 0   0%
hyper_parallel / platform / torch / hsdp / grad_hook.py TorchHSDPGradHook   7 7 0   0%
hyper_parallel / platform / torch / hsdp / grad_hook.py (no class)   3 3 0   0%
hyper_parallel / platform / torch / hsdp / hook_function.py PostBackwardFunction   29 29 0   0%
hyper_parallel / platform / torch / hsdp / hook_function.py (no class)   8 8 0   0%
hyper_parallel / platform / torch / hsdp / param.py TorchHSDPParam   14 14 0   0%
hyper_parallel / platform / torch / hsdp / param.py (no class)   8 8 0   0%
hyper_parallel / platform / torch / hsdp / scheduler.py TorchHSDPScheduler   39 39 0   0%
hyper_parallel / platform / torch / hsdp / scheduler.py (no class)   21 21 0   0%
hyper_parallel / platform / torch / hsdp / state.py TorchHSDPState   9 9 0   0%
hyper_parallel / platform / torch / hsdp / state.py (no class)   5 5 0   0%
hyper_parallel / platform / torch / init_weights.py (no class)   26 0 0   100%
hyper_parallel / platform / torch / pipeline_parallel / __init__.py (no class)   0 0 0   100%
hyper_parallel / platform / torch / pipeline_parallel / _utils.py _MicroBatch   42 14 0   67%
hyper_parallel / platform / torch / pipeline_parallel / _utils.py (no class)   7 0 0   100%
hyper_parallel / platform / torch / pipeline_parallel / stage.py PipelineStageBase   61 10 0   84%
hyper_parallel / platform / torch / pipeline_parallel / stage.py (no class)   14 0 0   100%
hyper_parallel / platform / torch / platform.py TorchPlatform   222 100 0   55%
hyper_parallel / platform / torch / platform.py (no class)   178 1 0   99%
Total     13071 4623 1   65%

No items found using the specified filter.