Browse Source

Serving, output build exception info

pull/210/head
xuyongfei 5 years ago
parent
commit
3bca8cb1e9
3 changed files with 5 additions and 3 deletions
  1. +2
    -1
      mindspore_serving/ccsrc/worker/inference/mindspore_model_wrap.cc
  2. +1
    -2
      mindspore_serving/master/__init__.py
  3. +2
    -0
      mindspore_serving/master/context.py

+ 2
- 1
mindspore_serving/ccsrc/worker/inference/mindspore_model_wrap.cc View File

@@ -99,7 +99,8 @@ Status MindSporeModelWrap::LoadModelFromFile(serving::DeviceType device_type, ui
} catch (std::runtime_error &ex) {
return INFER_STATUS_LOG_ERROR(FAILED)
<< "Load model from file failed, model file: " << file_name << ", device_type: '" << device_type
<< "', device_id: " << device_id << ", model type: " << model_type << ", options: " << other_options;
<< "', device_id: " << device_id << ", model type: " << model_type << ", options: " << other_options
<< ", build error detail: " << ex.what();
}

ApiModelInfo api_model_info;


+ 1
- 2
mindspore_serving/master/__init__.py View File

@@ -22,6 +22,5 @@ __all__.extend([
"start_grpc_server",
'start_restful_server',
'start_master_server',
'stop',
'context'
'stop'
])

+ 2
- 0
mindspore_serving/master/context.py View File

@@ -16,6 +16,8 @@
from mindspore_serving._mindspore_serving import MasterContext_
from mindspore_serving.common import check_type

__all__ = ["set_max_request_buffer_count"]

_context = MasterContext_.get_instance()




Loading…
Cancel
Save