| hyper_parallel / __init__.py |
(no class) |
|
12 |
0 |
0 |
|
100% |
| hyper_parallel / collectives / __init__.py |
(no class) |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / collectives / cc.py |
(no class) |
|
14 |
0 |
0 |
|
100% |
| hyper_parallel / core / __init__.py |
(no class) |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / activation_checkpoint / __init__.py |
(no class) |
|
2 |
0 |
0 |
|
100% |
| hyper_parallel / core / activation_checkpoint / activation_checkpoint.py |
CheckpointPolicy |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / activation_checkpoint / activation_checkpoint.py |
(no class) |
|
17 |
0 |
0 |
|
100% |
| hyper_parallel / core / activation_checkpoint / swap.py |
SwapTensor |
|
54 |
37 |
0 |
|
31% |
| hyper_parallel / core / activation_checkpoint / swap.py |
Storage |
|
30 |
14 |
0 |
|
53% |
| hyper_parallel / core / activation_checkpoint / swap.py |
SwapGroup |
|
37 |
16 |
0 |
|
57% |
| hyper_parallel / core / activation_checkpoint / swap.py |
SwapManager |
|
82 |
22 |
0 |
|
73% |
| hyper_parallel / core / activation_checkpoint / swap.py |
(no class) |
|
49 |
0 |
0 |
|
100% |
| hyper_parallel / core / checkpoint / __init__.py |
(no class) |
|
11 |
0 |
0 |
|
100% |
| hyper_parallel / core / checkpoint / api.py |
(no class) |
|
83 |
11 |
0 |
|
87% |
| hyper_parallel / core / checkpoint / filesystem_storage.py |
FileSystemWriter |
|
66 |
8 |
0 |
|
88% |
| hyper_parallel / core / checkpoint / filesystem_storage.py |
FileSystemReader |
|
44 |
11 |
0 |
|
75% |
| hyper_parallel / core / checkpoint / filesystem_storage.py |
(no class) |
|
62 |
4 |
0 |
|
94% |
| hyper_parallel / core / checkpoint / layout.py |
(no class) |
|
46 |
6 |
0 |
|
87% |
| hyper_parallel / core / checkpoint / loader.py |
(no class) |
|
11 |
1 |
0 |
|
91% |
| hyper_parallel / core / checkpoint / metadata.py |
MetadataIndex |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / checkpoint / metadata.py |
ChunkStorageMetadata |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / checkpoint / metadata.py |
TensorProperties |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / checkpoint / metadata.py |
BytesStorageMetadata |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / checkpoint / metadata.py |
TensorStorageMetadata |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / checkpoint / metadata.py |
Metadata |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / checkpoint / metadata.py |
(no class) |
|
29 |
0 |
0 |
|
100% |
| hyper_parallel / core / checkpoint / planner.py |
WriteItemType |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / checkpoint / planner.py |
LoadItemType |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / checkpoint / planner.py |
WriteItem |
|
21 |
2 |
0 |
|
90% |
| hyper_parallel / core / checkpoint / planner.py |
ReadItem |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / checkpoint / planner.py |
SavePlan |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / checkpoint / planner.py |
LoadPlan |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / checkpoint / planner.py |
SavePlanner |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / checkpoint / planner.py |
LoadPlanner |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / checkpoint / planner.py |
(no class) |
|
62 |
0 |
0 |
|
100% |
| hyper_parallel / core / checkpoint / reshard.py |
ReshardHandler |
|
80 |
2 |
0 |
|
98% |
| hyper_parallel / core / checkpoint / reshard.py |
(no class) |
|
79 |
6 |
0 |
|
92% |
| hyper_parallel / core / checkpoint / saver.py |
(no class) |
|
11 |
1 |
0 |
|
91% |
| hyper_parallel / core / checkpoint / standard_planner.py |
StandardSavePlanner |
|
109 |
9 |
0 |
|
92% |
| hyper_parallel / core / checkpoint / standard_planner.py |
StandardLoadPlanner |
|
58 |
13 |
0 |
|
78% |
| hyper_parallel / core / checkpoint / standard_planner.py |
(no class) |
|
45 |
1 |
0 |
|
98% |
| hyper_parallel / core / checkpoint / storage.py |
StorageInfo |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / checkpoint / storage.py |
WriteResult |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / checkpoint / storage.py |
StorageWriter |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / checkpoint / storage.py |
StorageReader |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / checkpoint / storage.py |
(no class) |
|
42 |
0 |
0 |
|
100% |
| hyper_parallel / core / checkpoint / util.py |
(no class) |
|
82 |
12 |
0 |
|
85% |
| hyper_parallel / core / device_mesh.py |
DeviceMesh |
|
331 |
129 |
0 |
|
61% |
| hyper_parallel / core / device_mesh.py |
(no class) |
|
101 |
3 |
0 |
|
97% |
| hyper_parallel / core / dtensor.py |
SkipDTensorDispatch |
|
4 |
0 |
0 |
|
100% |
| hyper_parallel / core / dtensor.py |
DTensor |
|
41 |
7 |
0 |
|
83% |
| hyper_parallel / core / dtensor.py |
(no class) |
|
60 |
0 |
0 |
|
100% |
| hyper_parallel / core / fully_shard / __init__.py |
(no class) |
|
2 |
0 |
0 |
|
100% |
| hyper_parallel / core / fully_shard / api.py |
_UnshardHandle |
|
4 |
4 |
0 |
|
0% |
| hyper_parallel / core / fully_shard / api.py |
HSDPModule |
|
122 |
117 |
0 |
|
4% |
| hyper_parallel / core / fully_shard / api.py |
(no class) |
|
119 |
37 |
0 |
|
69% |
| hyper_parallel / core / fully_shard / hsdp_async_grad_hook.py |
HSDPAsyncGradHook |
|
78 |
78 |
0 |
|
0% |
| hyper_parallel / core / fully_shard / hsdp_async_grad_hook.py |
(no class) |
|
10 |
0 |
0 |
|
100% |
| hyper_parallel / core / fully_shard / hsdp_grad_buffer.py |
HSDPGradBuffer |
|
145 |
145 |
0 |
|
0% |
| hyper_parallel / core / fully_shard / hsdp_grad_buffer.py |
(no class) |
|
19 |
0 |
0 |
|
100% |
| hyper_parallel / core / fully_shard / hsdp_grad_hook.py |
HSDPGradHook |
|
102 |
102 |
0 |
|
0% |
| hyper_parallel / core / fully_shard / hsdp_grad_hook.py |
(no class) |
|
11 |
0 |
0 |
|
100% |
| hyper_parallel / core / fully_shard / hsdp_param.py |
HSDPParamV2 |
|
27 |
27 |
0 |
|
0% |
| hyper_parallel / core / fully_shard / hsdp_param.py |
(no class) |
|
33 |
0 |
0 |
|
100% |
| hyper_parallel / core / fully_shard / hsdp_param_buffer.py |
HSDPParamBuffer |
|
74 |
74 |
0 |
|
0% |
| hyper_parallel / core / fully_shard / hsdp_param_buffer.py |
(no class) |
|
10 |
0 |
0 |
|
100% |
| hyper_parallel / core / fully_shard / hsdp_scheduler.py |
HSDPSchedulerContext |
|
5 |
0 |
0 |
|
100% |
| hyper_parallel / core / fully_shard / hsdp_scheduler.py |
HSDPSchedulerV2 |
|
84 |
50 |
0 |
|
40% |
| hyper_parallel / core / fully_shard / hsdp_scheduler.py |
(no class) |
|
29 |
0 |
0 |
|
100% |
| hyper_parallel / core / fully_shard / hsdp_state.py |
HSDPState |
|
113 |
82 |
0 |
|
27% |
| hyper_parallel / core / fully_shard / hsdp_state.py |
(no class) |
|
18 |
0 |
0 |
|
100% |
| hyper_parallel / core / fully_shard / hsdp_utils.py |
OptimizerLevel |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / fully_shard / hsdp_utils.py |
GroupInfo |
|
3 |
3 |
0 |
|
0% |
| hyper_parallel / core / fully_shard / hsdp_utils.py |
HSDPConfigV2 |
|
10 |
0 |
0 |
|
100% |
| hyper_parallel / core / fully_shard / hsdp_utils.py |
ShardedState |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / fully_shard / hsdp_utils.py |
FSDPSchedulerState |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / fully_shard / hsdp_utils.py |
ParamModuleInfo |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / fully_shard / hsdp_utils.py |
ExtensionsData |
|
2 |
2 |
0 |
|
0% |
| hyper_parallel / core / fully_shard / hsdp_utils.py |
(no class) |
|
56 |
7 |
0 |
|
88% |
| hyper_parallel / core / hsdp / __init__.py |
(no class) |
|
2 |
0 |
0 |
|
100% |
| hyper_parallel / core / hsdp / api.py |
HSDPCell |
|
35 |
9 |
0 |
|
74% |
| hyper_parallel / core / hsdp / api.py |
(no class) |
|
66 |
14 |
0 |
|
79% |
| hyper_parallel / core / hsdp / hsdp_async_grad_hook.py |
HSDPAsyncGradHook |
|
78 |
21 |
0 |
|
73% |
| hyper_parallel / core / hsdp / hsdp_async_grad_hook.py |
(no class) |
|
10 |
0 |
0 |
|
100% |
| hyper_parallel / core / hsdp / hsdp_grad_buffer.py |
HSDPGradBuffer |
|
145 |
76 |
0 |
|
48% |
| hyper_parallel / core / hsdp / hsdp_grad_buffer.py |
(no class) |
|
19 |
0 |
0 |
|
100% |
| hyper_parallel / core / hsdp / hsdp_grad_hook.py |
HSDPGradHook |
|
102 |
39 |
0 |
|
62% |
| hyper_parallel / core / hsdp / hsdp_grad_hook.py |
(no class) |
|
11 |
0 |
0 |
|
100% |
| hyper_parallel / core / hsdp / hsdp_param.py |
HSDPParam |
|
219 |
16 |
0 |
|
93% |
| hyper_parallel / core / hsdp / hsdp_param.py |
(no class) |
|
20 |
0 |
0 |
|
100% |
| hyper_parallel / core / hsdp / hsdp_param_buffer.py |
HSDPParamBuffer |
|
57 |
10 |
0 |
|
82% |
| hyper_parallel / core / hsdp / hsdp_param_buffer.py |
(no class) |
|
9 |
0 |
0 |
|
100% |
| hyper_parallel / core / hsdp / hsdp_scheduler.py |
FSDPSchedulerState |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / hsdp / hsdp_scheduler.py |
HSDPScheduler |
|
62 |
9 |
0 |
|
85% |
| hyper_parallel / core / hsdp / hsdp_scheduler.py |
(no class) |
|
27 |
0 |
0 |
|
100% |
| hyper_parallel / core / hsdp / hsdp_state.py |
HSDPState |
|
98 |
23 |
0 |
|
77% |
| hyper_parallel / core / hsdp / hsdp_state.py |
(no class) |
|
13 |
0 |
0 |
|
100% |
| hyper_parallel / core / hsdp / hsdp_utils.py |
OptimizerLevel |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / hsdp / hsdp_utils.py |
GroupInfo |
|
3 |
0 |
0 |
|
100% |
| hyper_parallel / core / hsdp / hsdp_utils.py |
HSDPConfig |
|
11 |
0 |
0 |
|
100% |
| hyper_parallel / core / hsdp / hsdp_utils.py |
(no class) |
|
9 |
0 |
0 |
|
100% |
| hyper_parallel / core / layout.py |
Layout |
|
218 |
29 |
0 |
|
87% |
| hyper_parallel / core / layout.py |
(no class) |
|
116 |
0 |
0 |
|
100% |
| hyper_parallel / core / parameter_init.py |
(no class) |
|
4 |
0 |
0 |
|
100% |
| hyper_parallel / core / pipeline_parallel / __init__.py |
(no class) |
|
4 |
0 |
0 |
|
100% |
| hyper_parallel / core / pipeline_parallel / scheduler.py |
MetaStepType |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / pipeline_parallel / scheduler.py |
MetaStep |
|
16 |
10 |
0 |
|
38% |
| hyper_parallel / core / pipeline_parallel / scheduler.py |
PipelineScheduleRuntime |
|
92 |
15 |
0 |
|
84% |
| hyper_parallel / core / pipeline_parallel / scheduler.py |
ScheduleGPipe |
|
17 |
17 |
0 |
|
0% |
| hyper_parallel / core / pipeline_parallel / scheduler.py |
Schedule1F1B |
|
44 |
44 |
0 |
|
0% |
| hyper_parallel / core / pipeline_parallel / scheduler.py |
ScheduleInterleaved1F1B |
|
63 |
8 |
0 |
|
87% |
| hyper_parallel / core / pipeline_parallel / scheduler.py |
(no class) |
|
244 |
138 |
0 |
|
43% |
| hyper_parallel / core / pipeline_parallel / stage.py |
SharedParameterInfo |
|
11 |
11 |
0 |
|
0% |
| hyper_parallel / core / pipeline_parallel / stage.py |
PipelineStage |
|
183 |
53 |
0 |
|
71% |
| hyper_parallel / core / pipeline_parallel / stage.py |
(no class) |
|
35 |
0 |
0 |
|
100% |
| hyper_parallel / core / pipeline_parallel / utils.py |
BatchDimSpec |
|
11 |
11 |
0 |
|
0% |
| hyper_parallel / core / pipeline_parallel / utils.py |
_RecvInfo |
|
5 |
0 |
0 |
|
100% |
| hyper_parallel / core / pipeline_parallel / utils.py |
(no class) |
|
17 |
0 |
0 |
|
100% |
| hyper_parallel / core / placement_types.py |
Placement |
|
4 |
4 |
0 |
|
0% |
| hyper_parallel / core / placement_types.py |
Shard |
|
11 |
7 |
0 |
|
36% |
| hyper_parallel / core / placement_types.py |
Replicate |
|
4 |
4 |
0 |
|
0% |
| hyper_parallel / core / placement_types.py |
Partial |
|
13 |
10 |
0 |
|
23% |
| hyper_parallel / core / placement_types.py |
(no class) |
|
29 |
0 |
0 |
|
100% |
| hyper_parallel / core / random.py |
_PhiloxState |
|
8 |
3 |
0 |
|
62% |
| hyper_parallel / core / random.py |
_RNGStateTracker |
|
9 |
2 |
0 |
|
78% |
| hyper_parallel / core / random.py |
OffsetBasedRNGTracker |
|
40 |
2 |
0 |
|
95% |
| hyper_parallel / core / random.py |
(no class) |
|
128 |
35 |
0 |
|
73% |
| hyper_parallel / core / redistribute_infer.py |
Status |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / redistribute_infer.py |
TensorMap |
|
14 |
8 |
0 |
|
43% |
| hyper_parallel / core / redistribute_infer.py |
DevMat |
|
47 |
25 |
0 |
|
47% |
| hyper_parallel / core / redistribute_infer.py |
RedistributionOperatorInfer |
|
236 |
95 |
0 |
|
60% |
| hyper_parallel / core / redistribute_infer.py |
(no class) |
|
34 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / __init__.py |
(no class) |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / _op_dispatch.py |
LayoutCacheKey |
|
9 |
1 |
0 |
|
89% |
| hyper_parallel / core / shard / _op_dispatch.py |
LayoutCacheManager |
|
9 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / _op_dispatch.py |
OpDispatcher |
|
374 |
97 |
0 |
|
74% |
| hyper_parallel / core / shard / _op_dispatch.py |
(no class) |
|
49 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / api.py |
parallelize_value_and_grad.CellWrapper |
|
4 |
1 |
0 |
|
75% |
| hyper_parallel / core / shard / api.py |
(no class) |
|
301 |
68 |
0 |
|
77% |
| hyper_parallel / core / shard / local_func.py |
(no class) |
|
48 |
8 |
0 |
|
83% |
| hyper_parallel / core / shard / ops / __init__.py |
(no class) |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_activation_with_axis.py |
ActivationWithAxisDistributedOp |
|
26 |
3 |
0 |
|
88% |
| hyper_parallel / core / shard / ops / parallel_activation_with_axis.py |
(no class) |
|
4 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_argmax_with_value_ops.py |
ArgMaxWithValueDistributedOp |
|
31 |
31 |
0 |
|
0% |
| hyper_parallel / core / shard / ops / parallel_argmax_with_value_ops.py |
(no class) |
|
3 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_concat.py |
ConcatDistributedOp |
|
28 |
28 |
0 |
|
0% |
| hyper_parallel / core / shard / ops / parallel_concat.py |
(no class) |
|
4 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_cumsum.py |
CumsumDistributedOp |
|
26 |
4 |
0 |
|
85% |
| hyper_parallel / core / shard / ops / parallel_cumsum.py |
(no class) |
|
4 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_elementwise.py |
ElementWiseDistributedOp |
|
201 |
62 |
0 |
|
69% |
| hyper_parallel / core / shard / ops / parallel_elementwise.py |
ElementWiseWithPartialDistributedOp |
|
2 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_elementwise.py |
AddDistributedOp |
|
19 |
1 |
0 |
|
95% |
| hyper_parallel / core / shard / ops / parallel_elementwise.py |
(no class) |
|
23 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_embedding.py |
EmbeddingDistributedOp |
|
22 |
1 |
0 |
|
95% |
| hyper_parallel / core / shard / ops / parallel_embedding.py |
(no class) |
|
4 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_expand.py |
ExpandDistributedOp |
|
41 |
5 |
1 |
|
88% |
| hyper_parallel / core / shard / ops / parallel_expand.py |
ExpandAsDistributedOp |
|
41 |
7 |
0 |
|
83% |
| hyper_parallel / core / shard / ops / parallel_expand.py |
(no class) |
|
6 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_expand_dims.py |
ExpandDimsDistributedOp |
|
22 |
6 |
0 |
|
73% |
| hyper_parallel / core / shard / ops / parallel_expand_dims.py |
(no class) |
|
4 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_flash_attention_score.py |
ParallelFlashAttention |
|
92 |
21 |
0 |
|
77% |
| hyper_parallel / core / shard / ops / parallel_flash_attention_score.py |
(no class) |
|
19 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_gather.py |
IndexSelectDistributedOp |
|
21 |
21 |
0 |
|
0% |
| hyper_parallel / core / shard / ops / parallel_gather.py |
GatherDistributedOp |
|
21 |
21 |
0 |
|
0% |
| hyper_parallel / core / shard / ops / parallel_gather.py |
GatherNdDistributedOp |
|
75 |
23 |
0 |
|
69% |
| hyper_parallel / core / shard / ops / parallel_gather.py |
(no class) |
|
14 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_isin.py |
IsinDistributedOp |
|
20 |
2 |
0 |
|
90% |
| hyper_parallel / core / shard / ops / parallel_isin.py |
(no class) |
|
4 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_masked_scatter.py |
MaskedScatterDistributedOp |
|
9 |
1 |
0 |
|
89% |
| hyper_parallel / core / shard / ops / parallel_masked_scatter.py |
(no class) |
|
3 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_matmul.py |
MatMulExtDistributedOp |
|
26 |
6 |
0 |
|
77% |
| hyper_parallel / core / shard / ops / parallel_matmul.py |
MatMulDistributedOp |
|
33 |
10 |
0 |
|
70% |
| hyper_parallel / core / shard / ops / parallel_matmul.py |
BaseBatchMatMulDistributedOp |
|
30 |
7 |
0 |
|
77% |
| hyper_parallel / core / shard / ops / parallel_matmul.py |
BatchMatMulExtDistributedOp |
|
15 |
3 |
0 |
|
80% |
| hyper_parallel / core / shard / ops / parallel_matmul.py |
BatchMatMulDistributedOp |
|
24 |
8 |
0 |
|
67% |
| hyper_parallel / core / shard / ops / parallel_matmul.py |
LinearDistributedOp |
|
47 |
10 |
0 |
|
79% |
| hyper_parallel / core / shard / ops / parallel_matmul.py |
(no class) |
|
18 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_multinomial.py |
MultinomialDistributedOp |
|
15 |
1 |
0 |
|
93% |
| hyper_parallel / core / shard / ops / parallel_multinomial.py |
(no class) |
|
4 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_new_ones.py |
NewOnesDistributedOp |
|
12 |
2 |
0 |
|
83% |
| hyper_parallel / core / shard / ops / parallel_new_ones.py |
(no class) |
|
4 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_norm.py |
NormDistributedOp |
|
33 |
33 |
0 |
|
0% |
| hyper_parallel / core / shard / ops / parallel_norm.py |
LayerNormDistributedOp |
|
31 |
5 |
0 |
|
84% |
| hyper_parallel / core / shard / ops / parallel_norm.py |
(no class) |
|
6 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_one_hot_ext.py |
OneHotExtDistributedOp |
|
149 |
149 |
0 |
|
0% |
| hyper_parallel / core / shard / ops / parallel_one_hot_ext.py |
(no class) |
|
20 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_ops.py |
DistributedOp |
|
12 |
6 |
0 |
|
50% |
| hyper_parallel / core / shard / ops / parallel_ops.py |
(no class) |
|
6 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_ops_register.py |
(no class) |
|
5 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_pad.py |
PadDistributedOp |
|
29 |
7 |
0 |
|
76% |
| hyper_parallel / core / shard / ops / parallel_pad.py |
(no class) |
|
3 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_reduce.py |
ReduceExtDistributedOpBase |
|
96 |
21 |
0 |
|
78% |
| hyper_parallel / core / shard / ops / parallel_reduce.py |
SumExtDistributedOp |
|
1 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_reduce.py |
MeanExtDistributedOp |
|
1 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_reduce.py |
ReduceMaxDistributedOp |
|
1 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_reduce.py |
ProdExtDistributedOp |
|
1 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_reduce.py |
AllExtDistributedOp |
|
1 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_reduce.py |
MaxDistributedOp |
|
22 |
3 |
0 |
|
86% |
| hyper_parallel / core / shard / ops / parallel_reduce.py |
(no class) |
|
32 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_repeat_interleave.py |
RepeatInterleaveDistributedOp |
|
29 |
5 |
0 |
|
83% |
| hyper_parallel / core / shard / ops / parallel_repeat_interleave.py |
(no class) |
|
4 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_reshape.py |
ReshapeDistributedOp |
|
115 |
32 |
0 |
|
72% |
| hyper_parallel / core / shard / ops / parallel_reshape.py |
(no class) |
|
26 |
9 |
0 |
|
65% |
| hyper_parallel / core / shard / ops / parallel_scatter_update.py |
ScatterUpdateDistributedOp |
|
36 |
11 |
0 |
|
69% |
| hyper_parallel / core / shard / ops / parallel_scatter_update.py |
(no class) |
|
5 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_slice.py |
SliceDistributedOp |
|
29 |
29 |
0 |
|
0% |
| hyper_parallel / core / shard / ops / parallel_slice.py |
(no class) |
|
5 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_slice_ext.py |
SliceExtDistributedOp |
|
6 |
6 |
0 |
|
0% |
| hyper_parallel / core / shard / ops / parallel_slice_ext.py |
(no class) |
|
3 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_sort.py |
SortDistributedOp |
|
22 |
4 |
0 |
|
82% |
| hyper_parallel / core / shard / ops / parallel_sort.py |
(no class) |
|
3 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_split.py |
SplitWithSizeDistributedOp |
|
9 |
9 |
0 |
|
0% |
| hyper_parallel / core / shard / ops / parallel_split.py |
SplitWithSizeViewDistributedOp |
|
9 |
9 |
0 |
|
0% |
| hyper_parallel / core / shard / ops / parallel_split.py |
SplitDistributedOp |
|
27 |
7 |
0 |
|
74% |
| hyper_parallel / core / shard / ops / parallel_split.py |
SplitTensorDistributedOp |
|
12 |
12 |
0 |
|
0% |
| hyper_parallel / core / shard / ops / parallel_split.py |
SplitTensorViewDistributedOp |
|
12 |
2 |
0 |
|
83% |
| hyper_parallel / core / shard / ops / parallel_split.py |
(no class) |
|
12 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_squeeze.py |
SqueezeDistributedOp |
|
76 |
20 |
0 |
|
74% |
| hyper_parallel / core / shard / ops / parallel_squeeze.py |
(no class) |
|
15 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_topk.py |
TopKDistributedOp |
|
15 |
2 |
0 |
|
87% |
| hyper_parallel / core / shard / ops / parallel_topk.py |
(no class) |
|
3 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_transpose.py |
TransposeDistributedOp |
|
34 |
8 |
0 |
|
76% |
| hyper_parallel / core / shard / ops / parallel_transpose.py |
(no class) |
|
4 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_tuple_elementwise.py |
TupleElementWiseDistributedOp |
|
3 |
1 |
0 |
|
67% |
| hyper_parallel / core / shard / ops / parallel_tuple_elementwise.py |
(no class) |
|
3 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / ops / parallel_unbind.py |
UnbindDistributedOp |
|
19 |
1 |
0 |
|
95% |
| hyper_parallel / core / shard / ops / parallel_unbind.py |
(no class) |
|
4 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / sharding_plan.py |
ShardingPlan |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / core / shard / sharding_plan.py |
(no class) |
|
8 |
0 |
0 |
|
100% |
| hyper_parallel / core / tensor_redistribution.py |
TensorRedistribution |
|
183 |
32 |
0 |
|
83% |
| hyper_parallel / core / tensor_redistribution.py |
(no class) |
|
26 |
0 |
0 |
|
100% |
| hyper_parallel / core / utils.py |
(no class) |
|
14 |
0 |
0 |
|
100% |
| hyper_parallel / platform / __init__.py |
(no class) |
|
1 |
0 |
0 |
|
100% |
| hyper_parallel / platform / mindspore / __init__.py |
(no class) |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / platform / mindspore / custom_pass / __init__.py |
(no class) |
|
7 |
0 |
0 |
|
100% |
| hyper_parallel / platform / mindspore / custom_pass / build_plugin.py |
(no class) |
|
56 |
56 |
0 |
|
0% |
| hyper_parallel / platform / mindspore / dtensor.py |
DTensorBase |
|
66 |
19 |
0 |
|
71% |
| hyper_parallel / platform / mindspore / dtensor.py |
(no class) |
|
23 |
0 |
0 |
|
100% |
| hyper_parallel / platform / mindspore / hsdp / __init__.py |
(no class) |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / platform / mindspore / hsdp / async_grad_hook.py |
MindSporeHSDPAsyncGradHook |
|
6 |
6 |
0 |
|
0% |
| hyper_parallel / platform / mindspore / hsdp / async_grad_hook.py |
(no class) |
|
3 |
0 |
0 |
|
100% |
| hyper_parallel / platform / mindspore / hsdp / grad_hook.py |
MindSporeHSDPGradHook |
|
21 |
8 |
0 |
|
62% |
| hyper_parallel / platform / mindspore / hsdp / grad_hook.py |
(no class) |
|
10 |
0 |
0 |
|
100% |
| hyper_parallel / platform / mindspore / hsdp / param.py |
MindSporeHSDPParam |
|
20 |
0 |
0 |
|
100% |
| hyper_parallel / platform / mindspore / hsdp / param.py |
(no class) |
|
9 |
0 |
0 |
|
100% |
| hyper_parallel / platform / mindspore / hsdp / scheduler.py |
MindSporeHSDPScheduler |
|
76 |
51 |
0 |
|
33% |
| hyper_parallel / platform / mindspore / hsdp / scheduler.py |
MindSporeHSDPScheduler._get_parameter_forward_hook.ForwardHookNet |
|
4 |
4 |
0 |
|
0% |
| hyper_parallel / platform / mindspore / hsdp / scheduler.py |
(no class) |
|
29 |
0 |
0 |
|
100% |
| hyper_parallel / platform / mindspore / hsdp / state.py |
MindSporeHSDPState |
|
17 |
1 |
0 |
|
94% |
| hyper_parallel / platform / mindspore / hsdp / state.py |
(no class) |
|
19 |
0 |
0 |
|
100% |
| hyper_parallel / platform / mindspore / parameter_init.py |
(no class) |
|
32 |
3 |
0 |
|
91% |
| hyper_parallel / platform / mindspore / pipeline_parallel / __init__.py |
(no class) |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / platform / mindspore / pipeline_parallel / _utils.py |
_MicroBatch |
|
41 |
41 |
0 |
|
0% |
| hyper_parallel / platform / mindspore / pipeline_parallel / _utils.py |
(no class) |
|
39 |
39 |
0 |
|
0% |
| hyper_parallel / platform / mindspore / pipeline_parallel / stage.py |
PipelineStageBase |
|
66 |
66 |
0 |
|
0% |
| hyper_parallel / platform / mindspore / pipeline_parallel / stage.py |
(no class) |
|
13 |
0 |
0 |
|
100% |
| hyper_parallel / platform / mindspore / platform.py |
MindSporePlatform |
|
179 |
66 |
0 |
|
63% |
| hyper_parallel / platform / mindspore / platform.py |
(no class) |
|
168 |
0 |
0 |
|
100% |
| hyper_parallel / platform / mindspore / platform_graph.py |
MindSporeGraphPlatform |
|
10 |
10 |
0 |
|
0% |
| hyper_parallel / platform / mindspore / platform_graph.py |
(no class) |
|
17 |
0 |
0 |
|
100% |
| hyper_parallel / platform / platform.py |
PlatformType |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / platform / platform.py |
Platform |
|
98 |
65 |
0 |
|
34% |
| hyper_parallel / platform / platform.py |
(no class) |
|
161 |
4 |
0 |
|
98% |
| hyper_parallel / platform / torch / __init__.py |
(no class) |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / activation_checkpoint / __init__.py |
(no class) |
|
1 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / activation_checkpoint / activation_swap.py |
ActivationPolicy |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / activation_checkpoint / activation_swap.py |
AsyncSaveOnCpu |
|
30 |
8 |
0 |
|
73% |
| hyper_parallel / platform / torch / activation_checkpoint / activation_swap.py |
ActivationWrapper |
|
15 |
9 |
0 |
|
40% |
| hyper_parallel / platform / torch / activation_checkpoint / activation_swap.py |
SwapWrapper |
|
4 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / activation_checkpoint / activation_swap.py |
(no class) |
|
37 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / activation_checkpoint / sac.py |
_VersionWrapper |
|
6 |
4 |
0 |
|
33% |
| hyper_parallel / platform / torch / activation_checkpoint / sac.py |
SelectiveCheckpointContext |
|
1 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / activation_checkpoint / sac.py |
_CachingTorchDispatchMode |
|
25 |
3 |
0 |
|
88% |
| hyper_parallel / platform / torch / activation_checkpoint / sac.py |
_CachedTorchDispatchMode |
|
26 |
23 |
0 |
|
12% |
| hyper_parallel / platform / torch / activation_checkpoint / sac.py |
(no class) |
|
47 |
11 |
0 |
|
77% |
| hyper_parallel / platform / torch / dtensor.py |
DTensorBase |
|
76 |
51 |
0 |
|
33% |
| hyper_parallel / platform / torch / dtensor.py |
(no class) |
|
45 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / fully_shard / __init__.py |
(no class) |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / fully_shard / async_grad_hook.py |
TorchHSDPAsyncGradHook |
|
7 |
7 |
0 |
|
0% |
| hyper_parallel / platform / torch / fully_shard / async_grad_hook.py |
(no class) |
|
3 |
3 |
0 |
|
0% |
| hyper_parallel / platform / torch / fully_shard / grad_hook.py |
TorchHSDPGradHook |
|
7 |
7 |
0 |
|
0% |
| hyper_parallel / platform / torch / fully_shard / grad_hook.py |
(no class) |
|
3 |
3 |
0 |
|
0% |
| hyper_parallel / platform / torch / fully_shard / hook_function.py |
PostBackwardFunction |
|
30 |
11 |
0 |
|
63% |
| hyper_parallel / platform / torch / fully_shard / hook_function.py |
(no class) |
|
8 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / fully_shard / param.py |
TorchHSDPParamV2 |
|
278 |
161 |
0 |
|
42% |
| hyper_parallel / platform / torch / fully_shard / param.py |
(no class) |
|
58 |
1 |
0 |
|
98% |
| hyper_parallel / platform / torch / fully_shard / scheduler.py |
TorchHSDPSchedulerV2 |
|
43 |
12 |
0 |
|
72% |
| hyper_parallel / platform / torch / fully_shard / scheduler.py |
(no class) |
|
20 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / fully_shard / state.py |
TorchHSDPStateV2 |
|
96 |
62 |
0 |
|
35% |
| hyper_parallel / platform / torch / fully_shard / state.py |
(no class) |
|
22 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / fully_shard / utils.py |
MixedPrecisionPolicy |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / fully_shard / utils.py |
OffloadPolicy |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / fully_shard / utils.py |
CPUOffloadPolicy |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / fully_shard / utils.py |
DataParallelMeshInfo |
|
2 |
1 |
0 |
|
50% |
| hyper_parallel / platform / torch / fully_shard / utils.py |
FSDPMeshInfo |
|
6 |
1 |
0 |
|
83% |
| hyper_parallel / platform / torch / fully_shard / utils.py |
DDPMeshInfo |
|
6 |
6 |
0 |
|
0% |
| hyper_parallel / platform / torch / fully_shard / utils.py |
HSDPMeshInfo |
|
1 |
1 |
0 |
|
0% |
| hyper_parallel / platform / torch / fully_shard / utils.py |
(no class) |
|
31 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / function_override.py |
DTensorBackwardHookFunction |
|
25 |
9 |
0 |
|
64% |
| hyper_parallel / platform / torch / function_override.py |
ExtendBackwardHook |
|
5 |
3 |
0 |
|
40% |
| hyper_parallel / platform / torch / function_override.py |
(no class) |
|
14 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / group_utils.py |
(no class) |
|
86 |
18 |
0 |
|
79% |
| hyper_parallel / platform / torch / hsdp / __init__.py |
(no class) |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / hsdp / async_grad_hook.py |
TorchHSDPAsyncGradHook |
|
7 |
4 |
0 |
|
43% |
| hyper_parallel / platform / torch / hsdp / async_grad_hook.py |
(no class) |
|
3 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / hsdp / grad_hook.py |
TorchHSDPGradHook |
|
7 |
4 |
0 |
|
43% |
| hyper_parallel / platform / torch / hsdp / grad_hook.py |
(no class) |
|
3 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / hsdp / hook_function.py |
PostBackwardFunction |
|
29 |
7 |
0 |
|
76% |
| hyper_parallel / platform / torch / hsdp / hook_function.py |
(no class) |
|
8 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / hsdp / param.py |
TorchHSDPParam |
|
14 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / hsdp / param.py |
(no class) |
|
8 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / hsdp / scheduler.py |
TorchHSDPScheduler |
|
39 |
11 |
0 |
|
72% |
| hyper_parallel / platform / torch / hsdp / scheduler.py |
(no class) |
|
21 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / hsdp / state.py |
TorchHSDPState |
|
9 |
1 |
0 |
|
89% |
| hyper_parallel / platform / torch / hsdp / state.py |
(no class) |
|
5 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / pipeline_parallel / __init__.py |
(no class) |
|
0 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / pipeline_parallel / _utils.py |
_MicroBatch |
|
42 |
10 |
0 |
|
76% |
| hyper_parallel / platform / torch / pipeline_parallel / _utils.py |
(no class) |
|
7 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / pipeline_parallel / stage.py |
PipelineStageBase |
|
61 |
10 |
0 |
|
84% |
| hyper_parallel / platform / torch / pipeline_parallel / stage.py |
(no class) |
|
14 |
0 |
0 |
|
100% |
| hyper_parallel / platform / torch / platform.py |
TorchPlatform |
|
205 |
67 |
0 |
|
67% |
| hyper_parallel / platform / torch / platform.py |
(no class) |
|
167 |
1 |
0 |
|
99% |