|
|
@@ -145,7 +145,7 @@ def _get_parameter_broadcast(): |
|
|
parallel_mode = auto_parallel_context().get_parallel_mode() |
|
|
parallel_mode = auto_parallel_context().get_parallel_mode() |
|
|
parameter_broadcast = auto_parallel_context().get_parameter_broadcast() |
|
|
parameter_broadcast = auto_parallel_context().get_parameter_broadcast() |
|
|
|
|
|
|
|
|
if parallel_mode in ("data_parallel", "hybrid_parallel") and parameter_broadcast is False and get_seed is None: |
|
|
|
|
|
|
|
|
if parallel_mode in ("data_parallel", "hybrid_parallel") and parameter_broadcast is False and get_seed() is None: |
|
|
logger.warning("You are suggested to use mindspore.context.set_auto_parallel_context(parameter_broadcast=True)" |
|
|
logger.warning("You are suggested to use mindspore.context.set_auto_parallel_context(parameter_broadcast=True)" |
|
|
" or mindspore.common.set_seed() to share parameters among multi-devices.") |
|
|
" or mindspore.common.set_seed() to share parameters among multi-devices.") |
|
|
|
|
|
|
|
|
|