File tree Expand file tree Collapse file tree 1 file changed +6
-13
lines changed Expand file tree Collapse file tree 1 file changed +6
-13
lines changed Original file line number Diff line number Diff line change @@ -119,21 +119,14 @@ def aten_autobucketing_reordering_pass(
119119def configure_inductor_for_autobucketing (mode : str = "aten" ):
120120 # allow configuring inductor comms optimizations from torchtitan commandline
121121 if mode == "aten" :
122- from autoparallel .auto_bucketing import (
123- aten_autobucketing_config ,
124- aten_autobucketing_reordering_pass ,
125- )
126-
127- # this is from the stacked pr in https://github.com/pytorch/pytorch/pull/163960
128- torch ._inductor .config .reorder_for_peak_memory = False
129- torch ._inductor .config .reorder_for_compute_comm_overlap = False
130- aten_autobucketing_reordering_pass = partial (
131- aten_autobucketing_reordering_pass ,
132- configs = aten_autobucketing_config , # type: ignore
122+ torch ._inductor .config .aten_distributed_optimizations .enable_overlap_scheduling = (
123+ True
133124 )
134- torch ._inductor .config .post_grad_custom_post_pass = (
135- aten_autobucketing_reordering_pass # type: ignore
125+ torch ._inductor .config .aten_distributed_optimizations . collective_bucketing = (
126+ True
136127 )
128+ torch ._inductor .config .aten_distributed_optimizations .insert_overlap_deps = True
129+ torch ._inductor .config .aten_distributed_optimizations .max_compute_pre_fetch = 10
137130 elif mode == "inductor" :
138131 from autoparallel .auto_bucketing import (
139132 simple_fsdp_autobucketing_reordering_pass ,
You can’t perform that action at this time.
0 commit comments