diff --git a/oneflow/core/job_rewriter/multi_tensor_model_update.cpp b/oneflow/core/job_rewriter/multi_tensor_model_update.cpp index a2d4e83ad83..334b0860744 100644 --- a/oneflow/core/job_rewriter/multi_tensor_model_update.cpp +++ b/oneflow/core/job_rewriter/multi_tensor_model_update.cpp @@ -231,10 +231,6 @@ Maybe MultiTensorModelUpdatePass::Apply(const OpGraph& op_graph, } const user_op::UserOpConfWrapper model_update_user_conf( find_model_update_update_node->op().op_conf()); - // Multi tensor update pass only support for CUDA currently. - if (find_model_update_update_node->parallel_desc().device_type() != DeviceType::kCUDA) { - continue; - } // Multi tensor update pass only support Data Parallel. bool if_data_parallel = true; diff --git a/python/oneflow/nn/optimizer/adamw.py b/python/oneflow/nn/optimizer/adamw.py index 10ed9e12640..17e650598f7 100644 --- a/python/oneflow/nn/optimizer/adamw.py +++ b/python/oneflow/nn/optimizer/adamw.py @@ -163,10 +163,6 @@ def __init__( warnings.warn("Fused Adamw is not supported when amsgrad=True.") param_group["fused"] = False - if param_group["fused"] and not param.is_cuda: - warnings.warn("Fused Adamw only support cuda parameters.") - param_group["fused"] = False - self._op_with_amsgrad = ( flow.stateful_op("adam_update") .Input("model")