@@ -158,9 +158,14 @@ def monkey_patch_initialize_model_parallel(
158158
159159 # Build the tensor model-parallel groups.
160160 num_tensor_model_parallel_groups : int = world_size // tensor_model_parallel_size
161- assert (
162- sglang .srt .distributed .parallel_state ._TP is None
163- ), "tensor model parallel group is already initialized"
161+ ############################################################################
162+ ## This is a patch code for sgalng
163+ ## Ignore parallel state already set alert
164+ # assert (
165+ # sglang.srt.distributed.parallel_state._TP is None
166+ # ), "tensor model parallel group is already initialized"
167+ ## End of patch
168+ ############################################################################
164169 group_ranks = []
165170 for i in range (num_tensor_model_parallel_groups ):
166171 ranks = list (range (i * tensor_model_parallel_size , (i + 1 ) * tensor_model_parallel_size ))
@@ -199,9 +204,14 @@ def monkey_patch_initialize_model_parallel(
199204 moe_ep_size = expert_model_parallel_size
200205
201206 moe_tp_size = tensor_model_parallel_size // moe_ep_size
202- assert (
203- sglang .srt .distributed .parallel_state ._MOE_EP is None
204- ), "expert model parallel group is already initialized"
207+ ############################################################################
208+ ## This is a patch code for sgalng
209+ ## Ignore parallel state already set alert
210+ # assert (
211+ # sglang.srt.distributed.parallel_state._MOE_EP is None
212+ # ), "expert model parallel group is already initialized"
213+ ## End of patch
214+ ############################################################################
205215 group_ranks = []
206216 for i in range (num_tensor_model_parallel_groups ):
207217 for j in range (moe_tp_size ):
@@ -220,9 +230,14 @@ def monkey_patch_initialize_model_parallel(
220230 )
221231 )
222232
223- assert (
224- sglang .srt .distributed .parallel_state ._MOE_TP is None
225- ), "expert model parallel group is already initialized"
233+ ############################################################################
234+ ## This is a patch code for sgalng
235+ ## Ignore parallel state already set alert
236+ # assert (
237+ # sglang.srt.distributed.parallel_state._MOE_TP is None
238+ # ), "expert model parallel group is already initialized"
239+ ## End of patch
240+ ############################################################################
226241 group_ranks = []
227242 for i in range (num_tensor_model_parallel_groups ):
228243 for j in range (moe_ep_size ):
@@ -243,9 +258,14 @@ def monkey_patch_initialize_model_parallel(
243258
244259 # Build the pipeline model-parallel groups.
245260 num_pipeline_model_parallel_groups : int = world_size // pipeline_model_parallel_size
246- assert (
247- sglang .srt .distributed .parallel_state ._PP is None
248- ), "pipeline model parallel group is already initialized"
261+ ############################################################################
262+ ## This is a patch code for sgalng
263+ ## Ignore parallel state already set alert
264+ # assert (
265+ # sglang.srt.distributed.parallel_state._PP is None
266+ # ), "pipeline model parallel group is already initialized"
267+ ## End of patch
268+ ############################################################################
249269 group_ranks = []
250270 for i in range (num_pipeline_model_parallel_groups ):
251271 ranks = list (range (i , world_size , num_pipeline_model_parallel_groups ))
0 commit comments