Browse Source

fix cmake compile options

tags/v1.0.0
hexia 5 years ago
parent
commit
23641e618a
16 changed files with 2 additions and 25 deletions
  1. +2
    -0
      cmake/external_libs/libevent.cmake
  2. +0
    -1
      serving/acl/acl_session.h
  3. +0
    -2
      serving/acl/dvpp_process.cc
  4. +0
    -2
      serving/acl/model_process.cc
  5. +0
    -2
      serving/acl/model_process.h
  6. +0
    -2
      serving/core/http_process.cc
  7. +0
    -3
      serving/core/server.cc
  8. +0
    -1
      serving/core/server.h
  9. +0
    -2
      serving/core/serving_tensor.cc
  10. +0
    -2
      serving/core/serving_tensor.h
  11. +0
    -2
      serving/core/session.cc
  12. +0
    -2
      serving/core/session.h
  13. +0
    -1
      serving/core/util/file_system_operation.cc
  14. +0
    -1
      serving/core/util/option_parser.cc
  15. +0
    -1
      serving/core/version_control/model.cc
  16. +0
    -1
      serving/core/version_control/version_controller.cc

+ 2
- 0
cmake/external_libs/libevent.cmake View File

@@ -1,3 +1,5 @@
set(libevent_CFLAGS "-fstack-protector-all -D_FORTIFY_SOURCE=2 -O2")
set(libevent_LDFLAGS "-Wl,-z,now")
mindspore_add_pkg(libevent
VER 2.1.12
LIBS event event_pthreads


+ 0
- 1
serving/acl/acl_session.h View File

@@ -29,7 +29,6 @@

namespace mindspore {
namespace inference {

class AclSession : public InferSession {
public:
AclSession();


+ 0
- 2
serving/acl/dvpp_process.cc View File

@@ -23,7 +23,6 @@

namespace mindspore {
namespace inference {

DvppProcess::DvppProcess() {}

DvppProcess::~DvppProcess() {}
@@ -1134,6 +1133,5 @@ Status DvppProcess::InitWithJsonConfig(const std::string &json_config) {
}
return SUCCESS;
}

} // namespace inference
} // namespace mindspore

+ 0
- 2
serving/acl/model_process.cc View File

@@ -22,7 +22,6 @@

namespace mindspore {
namespace inference {

Status ModelProcess::PreInitModelResource() {
model_desc_ = aclmdlCreateDesc();
aclError acl_ret = aclmdlGetDesc(model_desc_, model_id_);
@@ -428,6 +427,5 @@ size_t ModelProcess::GetBatchSize() const {
}
return static_cast<size_t>(input_infos_[0].dims[0]);
}

} // namespace inference
} // namespace mindspore

+ 0
- 2
serving/acl/model_process.h View File

@@ -25,7 +25,6 @@

namespace mindspore {
namespace inference {

struct AclTensorInfo {
void *device_data;
size_t buffer_size;
@@ -78,7 +77,6 @@ class ModelProcess {
void DestroyInputsBuffer();
void DestroyOutputsBuffer();
};

} // namespace inference
} // namespace mindspore



+ 0
- 2
serving/core/http_process.cc View File

@@ -32,7 +32,6 @@ using nlohmann::json;

namespace mindspore {
namespace serving {

const int BUF_MAX = 0x7FFFFFFF;
static constexpr char HTTP_DATA[] = "data";
static constexpr char HTTP_TENSOR[] = "tensor";
@@ -550,6 +549,5 @@ void http_handler_msg(struct evhttp_request *const req, void *const arg) {
evbuffer_free(retbuff);
MSI_TIME_STAMP_END(TotalRestfulPredict)
}

} // namespace serving
} // namespace mindspore

+ 0
- 3
serving/core/server.cc View File

@@ -40,7 +40,6 @@ using ms_serving::PredictRequest;

namespace mindspore {
namespace serving {

namespace {
static const uint32_t uint32max = 0x7FFFFFFF;
std::promise<void> exit_requested;
@@ -66,7 +65,6 @@ grpc::Status CreatGRPCStatus(const Status &status) {
return grpc::Status::CANCELLED;
}
}

} // namespace

// Service Implement
@@ -114,7 +112,6 @@ static std::pair<struct evhttp *, struct event_base *> NewHttpServer() {
auto listener =
evconnlistener_new_bind(eb, nullptr, nullptr, LEV_OPT_REUSEABLE | LEV_OPT_CLOSE_ON_EXEC | LEV_OPT_CLOSE_ON_FREE, -1,
reinterpret_cast<struct sockaddr *>(&sin), sizeof(sin));

if (listener == nullptr) {
MSI_LOG_ERROR << "Serving Error: RESTful server start failed, create http listener faild, port " << http_port;
std::cout << "Serving Error: RESTful server start failed, create http listener faild, port " << http_port


+ 0
- 1
serving/core/server.h View File

@@ -19,7 +19,6 @@
#include "util/status.h"
namespace mindspore {
namespace serving {

class Server {
public:
Server() = default;


+ 0
- 2
serving/core/serving_tensor.cc View File

@@ -27,7 +27,6 @@ using std::vector;

namespace mindspore {
namespace serving {

using inference::DataType;
using inference::InferTensorBase;

@@ -191,6 +190,5 @@ const inference::InferImagesBase *ServingImagesRequest::operator[](size_t index)
}
return &(cache_[index]);
}

} // namespace serving
} // namespace mindspore

+ 0
- 2
serving/core/serving_tensor.h View File

@@ -25,7 +25,6 @@

namespace mindspore {
namespace serving {

class MS_API ServingTensor : public inference::InferTensorBase {
public:
// the other's lifetime must longer than this object
@@ -101,7 +100,6 @@ class ServingImagesRequest : public inference::ImagesRequestBase {
const ms_serving::PredictRequest &request_;
std::vector<ServingImages> cache_;
};

} // namespace serving
} // namespace mindspore
#endif // MINDSPORE_SERVING_TENSOR_H_

+ 0
- 2
serving/core/session.cc View File

@@ -35,7 +35,6 @@ using ms_serving::PredictRequest;

namespace mindspore {
namespace serving {

Status Session::CreatDeviceSession(const std::string &device, uint32_t device_id) {
session_ = inference::InferSession::CreateSession(device, device_id);
if (session_ == nullptr) {
@@ -151,6 +150,5 @@ Status Session::GetModelInputsInfo(std::vector<inference::InferTensor> &tensor_l
}
return ret;
}

} // namespace serving
} // namespace mindspore

+ 0
- 2
serving/core/session.h View File

@@ -28,7 +28,6 @@

namespace mindspore {
namespace serving {

using inference::FAILED;
using inference::INVALID_INPUTS;
using inference::Status;
@@ -57,7 +56,6 @@ class Session {

Status PredictInner(const PredictRequest &request, PredictReply &reply);
};

} // namespace serving
} // namespace mindspore
#endif // MINDSPORE_SERVER_H

+ 0
- 1
serving/core/util/file_system_operation.cc View File

@@ -29,7 +29,6 @@

namespace mindspore {
namespace serving {

bool DirOrFileExist(const std::string &file_path) {
int ret = access(file_path.c_str(), 0);
return (ret == -1) ? false : true;


+ 0
- 1
serving/core/util/option_parser.cc View File

@@ -23,7 +23,6 @@

namespace mindspore {
namespace serving {

bool StartWith(const std::string &str, const std::string &expected) {
return expected.empty() ||
(str.size() >= expected.size() && memcmp(str.data(), expected.data(), expected.size()) == 0);


+ 0
- 1
serving/core/version_control/model.cc View File

@@ -19,7 +19,6 @@

namespace mindspore {
namespace serving {

MindSporeModel::MindSporeModel(const std::string &model_name, const std::string &model_path,
const std::string &model_version, const time_t &last_update_time)
: model_name_(model_name),


+ 0
- 1
serving/core/version_control/version_controller.cc View File

@@ -25,7 +25,6 @@

namespace mindspore {
namespace serving {

volatile bool stop_poll = false;

std::string GetVersionFromPath(const std::string &path) {


Loading…
Cancel
Save