| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428 | /* * * Copyright 2015 gRPC authors. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * *     http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. * */#include <chrono>#include <memory>#include <mutex>#include <sstream>#include <string>#include <thread>#include <vector>#include <grpc/grpc.h>#include <grpc/support/alloc.h>#include <grpc/support/log.h>#include <grpc/support/time.h>#include <grpcpp/channel.h>#include <grpcpp/client_context.h>#include <grpcpp/server.h>#include <grpcpp/server_builder.h>#include "src/core/lib/profiling/timers.h"#include "src/proto/grpc/testing/benchmark_service.grpc.pb.h"#include "test/cpp/qps/client.h"#include "test/cpp/qps/interarrival.h"#include "test/cpp/qps/usage_timer.h"namespace grpc {namespace testing {static std::unique_ptr<BenchmarkService::Stub> BenchmarkStubCreator(    const std::shared_ptr<Channel>& ch) {  return BenchmarkService::NewStub(ch);}class SynchronousClient    : public ClientImpl<BenchmarkService::Stub, SimpleRequest> { public:  explicit SynchronousClient(const ClientConfig& config)      : ClientImpl<BenchmarkService::Stub, SimpleRequest>(            config, BenchmarkStubCreator) {    num_threads_ =        config.outstanding_rpcs_per_channel() * config.client_channels();    responses_.resize(num_threads_);    SetupLoadTest(config, num_threads_);  }  ~SynchronousClient() override {}  virtual bool InitThreadFuncImpl(size_t thread_idx) = 0;  virtual bool ThreadFuncImpl(HistogramEntry* entry, size_t thread_idx) = 0;  void ThreadFunc(size_t thread_idx, Thread* t) override {    if (!InitThreadFuncImpl(thread_idx)) {      return;    }    for (;;) {      // run the loop body      HistogramEntry entry;      const bool thread_still_ok = ThreadFuncImpl(&entry, thread_idx);      t->UpdateHistogram(&entry);      if (!thread_still_ok || ThreadCompleted()) {        return;      }    }  } protected:  // WaitToIssue returns false if we realize that we need to break out  bool WaitToIssue(int thread_idx) {    if (!closed_loop_) {      const gpr_timespec next_issue_time = NextIssueTime(thread_idx);      // Avoid sleeping for too long continuously because we might      // need to terminate before then. This is an issue since      // exponential distribution can occasionally produce bad outliers      while (true) {        const gpr_timespec one_sec_delay =            gpr_time_add(gpr_now(GPR_CLOCK_MONOTONIC),                         gpr_time_from_seconds(1, GPR_TIMESPAN));        if (gpr_time_cmp(next_issue_time, one_sec_delay) <= 0) {          gpr_sleep_until(next_issue_time);          return true;        } else {          gpr_sleep_until(one_sec_delay);          if (gpr_atm_acq_load(&thread_pool_done_) != static_cast<gpr_atm>(0)) {            return false;          }        }      }    }    return true;  }  size_t num_threads_;  std::vector<SimpleResponse> responses_;};class SynchronousUnaryClient final : public SynchronousClient { public:  explicit SynchronousUnaryClient(const ClientConfig& config)      : SynchronousClient(config) {    StartThreads(num_threads_);  }  ~SynchronousUnaryClient() override {}  bool InitThreadFuncImpl(size_t /*thread_idx*/) override { return true; }  bool ThreadFuncImpl(HistogramEntry* entry, size_t thread_idx) override {    if (!WaitToIssue(thread_idx)) {      return true;    }    auto* stub = channels_[thread_idx % channels_.size()].get_stub();    double start = UsageTimer::Now();    GPR_TIMER_SCOPE("SynchronousUnaryClient::ThreadFunc", 0);    grpc::ClientContext context;    grpc::Status s =        stub->UnaryCall(&context, request_, &responses_[thread_idx]);    if (s.ok()) {      entry->set_value((UsageTimer::Now() - start) * 1e9);    }    entry->set_status(s.error_code());    return true;  } private:  void DestroyMultithreading() final { EndThreads(); }};template <class StreamType>class SynchronousStreamingClient : public SynchronousClient { public:  explicit SynchronousStreamingClient(const ClientConfig& config)      : SynchronousClient(config),        context_(num_threads_),        stream_(num_threads_),        stream_mu_(num_threads_),        shutdown_(num_threads_),        messages_per_stream_(config.messages_per_stream()),        messages_issued_(num_threads_) {    StartThreads(num_threads_);  }  ~SynchronousStreamingClient() override {    CleanupAllStreams([this](size_t thread_idx) {      // Don't log any kind of error since we may have canceled this      stream_[thread_idx]->Finish().IgnoreError();    });  } protected:  std::vector<grpc::ClientContext> context_;  std::vector<std::unique_ptr<StreamType>> stream_;  // stream_mu_ is only needed when changing an element of stream_ or context_  std::vector<std::mutex> stream_mu_;  // use struct Bool rather than bool because vector<bool> is not concurrent  struct Bool {    bool val;    Bool() : val(false) {}  };  std::vector<Bool> shutdown_;  const int messages_per_stream_;  std::vector<int> messages_issued_;  void FinishStream(HistogramEntry* entry, size_t thread_idx) {    Status s = stream_[thread_idx]->Finish();    // don't set the value since the stream is failed and shouldn't be timed    entry->set_status(s.error_code());    if (!s.ok()) {      std::lock_guard<std::mutex> l(stream_mu_[thread_idx]);      if (!shutdown_[thread_idx].val) {        gpr_log(GPR_ERROR, "Stream %" PRIuPTR " received an error %s",                thread_idx, s.error_message().c_str());      }    }    // Lock the stream_mu_ now because the client context could change    std::lock_guard<std::mutex> l(stream_mu_[thread_idx]);    context_[thread_idx].~ClientContext();    new (&context_[thread_idx]) ClientContext();  }  void CleanupAllStreams(const std::function<void(size_t)>& cleaner) {    std::vector<std::thread> cleanup_threads;    for (size_t i = 0; i < num_threads_; i++) {      cleanup_threads.emplace_back([this, i, cleaner] {        std::lock_guard<std::mutex> l(stream_mu_[i]);        shutdown_[i].val = true;        if (stream_[i]) {          cleaner(i);        }      });    }    for (auto& th : cleanup_threads) {      th.join();    }  } private:  void DestroyMultithreading() final {    CleanupAllStreams(        [this](size_t thread_idx) { context_[thread_idx].TryCancel(); });    EndThreads();  }};class SynchronousStreamingPingPongClient final    : public SynchronousStreamingClient<          grpc::ClientReaderWriter<SimpleRequest, SimpleResponse>> { public:  explicit SynchronousStreamingPingPongClient(const ClientConfig& config)      : SynchronousStreamingClient(config) {}  ~SynchronousStreamingPingPongClient() override {    CleanupAllStreams(        [this](size_t thread_idx) { stream_[thread_idx]->WritesDone(); });  } private:  bool InitThreadFuncImpl(size_t thread_idx) override {    auto* stub = channels_[thread_idx % channels_.size()].get_stub();    std::lock_guard<std::mutex> l(stream_mu_[thread_idx]);    if (!shutdown_[thread_idx].val) {      stream_[thread_idx] = stub->StreamingCall(&context_[thread_idx]);    } else {      return false;    }    messages_issued_[thread_idx] = 0;    return true;  }  bool ThreadFuncImpl(HistogramEntry* entry, size_t thread_idx) override {    if (!WaitToIssue(thread_idx)) {      return true;    }    GPR_TIMER_SCOPE("SynchronousStreamingPingPongClient::ThreadFunc", 0);    double start = UsageTimer::Now();    if (stream_[thread_idx]->Write(request_) &&        stream_[thread_idx]->Read(&responses_[thread_idx])) {      entry->set_value((UsageTimer::Now() - start) * 1e9);      // don't set the status since there isn't one yet      if ((messages_per_stream_ != 0) &&          (++messages_issued_[thread_idx] < messages_per_stream_)) {        return true;      } else if (messages_per_stream_ == 0) {        return true;      } else {        // Fall through to the below resetting code after finish      }    }    stream_[thread_idx]->WritesDone();    FinishStream(entry, thread_idx);    auto* stub = channels_[thread_idx % channels_.size()].get_stub();    std::lock_guard<std::mutex> l(stream_mu_[thread_idx]);    if (!shutdown_[thread_idx].val) {      stream_[thread_idx] = stub->StreamingCall(&context_[thread_idx]);    } else {      stream_[thread_idx].reset();      return false;    }    messages_issued_[thread_idx] = 0;    return true;  }};class SynchronousStreamingFromClientClient final    : public SynchronousStreamingClient<grpc::ClientWriter<SimpleRequest>> { public:  explicit SynchronousStreamingFromClientClient(const ClientConfig& config)      : SynchronousStreamingClient(config), last_issue_(num_threads_) {}  ~SynchronousStreamingFromClientClient() override {    CleanupAllStreams(        [this](size_t thread_idx) { stream_[thread_idx]->WritesDone(); });  } private:  std::vector<double> last_issue_;  bool InitThreadFuncImpl(size_t thread_idx) override {    auto* stub = channels_[thread_idx % channels_.size()].get_stub();    std::lock_guard<std::mutex> l(stream_mu_[thread_idx]);    if (!shutdown_[thread_idx].val) {      stream_[thread_idx] = stub->StreamingFromClient(&context_[thread_idx],                                                      &responses_[thread_idx]);    } else {      return false;    }    last_issue_[thread_idx] = UsageTimer::Now();    return true;  }  bool ThreadFuncImpl(HistogramEntry* entry, size_t thread_idx) override {    // Figure out how to make histogram sensible if this is rate-paced    if (!WaitToIssue(thread_idx)) {      return true;    }    GPR_TIMER_SCOPE("SynchronousStreamingFromClientClient::ThreadFunc", 0);    if (stream_[thread_idx]->Write(request_)) {      double now = UsageTimer::Now();      entry->set_value((now - last_issue_[thread_idx]) * 1e9);      last_issue_[thread_idx] = now;      return true;    }    stream_[thread_idx]->WritesDone();    FinishStream(entry, thread_idx);    auto* stub = channels_[thread_idx % channels_.size()].get_stub();    std::lock_guard<std::mutex> l(stream_mu_[thread_idx]);    if (!shutdown_[thread_idx].val) {      stream_[thread_idx] = stub->StreamingFromClient(&context_[thread_idx],                                                      &responses_[thread_idx]);    } else {      stream_[thread_idx].reset();      return false;    }    return true;  }};class SynchronousStreamingFromServerClient final    : public SynchronousStreamingClient<grpc::ClientReader<SimpleResponse>> { public:  explicit SynchronousStreamingFromServerClient(const ClientConfig& config)      : SynchronousStreamingClient(config), last_recv_(num_threads_) {}  ~SynchronousStreamingFromServerClient() override {} private:  std::vector<double> last_recv_;  bool InitThreadFuncImpl(size_t thread_idx) override {    auto* stub = channels_[thread_idx % channels_.size()].get_stub();    std::lock_guard<std::mutex> l(stream_mu_[thread_idx]);    if (!shutdown_[thread_idx].val) {      stream_[thread_idx] =          stub->StreamingFromServer(&context_[thread_idx], request_);    } else {      return false;    }    last_recv_[thread_idx] = UsageTimer::Now();    return true;  }  bool ThreadFuncImpl(HistogramEntry* entry, size_t thread_idx) override {    GPR_TIMER_SCOPE("SynchronousStreamingFromServerClient::ThreadFunc", 0);    if (stream_[thread_idx]->Read(&responses_[thread_idx])) {      double now = UsageTimer::Now();      entry->set_value((now - last_recv_[thread_idx]) * 1e9);      last_recv_[thread_idx] = now;      return true;    }    FinishStream(entry, thread_idx);    auto* stub = channels_[thread_idx % channels_.size()].get_stub();    std::lock_guard<std::mutex> l(stream_mu_[thread_idx]);    if (!shutdown_[thread_idx].val) {      stream_[thread_idx] =          stub->StreamingFromServer(&context_[thread_idx], request_);    } else {      stream_[thread_idx].reset();      return false;    }    return true;  }};class SynchronousStreamingBothWaysClient final    : public SynchronousStreamingClient<          grpc::ClientReaderWriter<SimpleRequest, SimpleResponse>> { public:  explicit SynchronousStreamingBothWaysClient(const ClientConfig& config)      : SynchronousStreamingClient(config) {}  ~SynchronousStreamingBothWaysClient() override {    CleanupAllStreams(        [this](size_t thread_idx) { stream_[thread_idx]->WritesDone(); });  } private:  bool InitThreadFuncImpl(size_t thread_idx) override {    auto* stub = channels_[thread_idx % channels_.size()].get_stub();    std::lock_guard<std::mutex> l(stream_mu_[thread_idx]);    if (!shutdown_[thread_idx].val) {      stream_[thread_idx] = stub->StreamingBothWays(&context_[thread_idx]);    } else {      return false;    }    return true;  }  bool ThreadFuncImpl(HistogramEntry* /*entry*/,                      size_t /*thread_idx*/) override {    // TODO (vjpai): Do this    return true;  }};std::unique_ptr<Client> CreateSynchronousClient(const ClientConfig& config) {  GPR_ASSERT(!config.use_coalesce_api());  // not supported yet.  switch (config.rpc_type()) {    case UNARY:      return std::unique_ptr<Client>(new SynchronousUnaryClient(config));    case STREAMING:      return std::unique_ptr<Client>(          new SynchronousStreamingPingPongClient(config));    case STREAMING_FROM_CLIENT:      return std::unique_ptr<Client>(          new SynchronousStreamingFromClientClient(config));    case STREAMING_FROM_SERVER:      return std::unique_ptr<Client>(          new SynchronousStreamingFromServerClient(config));    case STREAMING_BOTH_WAYS:      return std::unique_ptr<Client>(          new SynchronousStreamingBothWaysClient(config));    default:      assert(false);      return nullptr;  }}}  // namespace testing}  // namespace grpc
 |