| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509 | 
							- /*
 
-  *
 
-  * Copyright 2015, Google Inc.
 
-  * All rights reserved.
 
-  *
 
-  * Redistribution and use in source and binary forms, with or without
 
-  * modification, are permitted provided that the following conditions are
 
-  * met:
 
-  *
 
-  *     * Redistributions of source code must retain the above copyright
 
-  * notice, this list of conditions and the following disclaimer.
 
-  *     * Redistributions in binary form must reproduce the above
 
-  * copyright notice, this list of conditions and the following disclaimer
 
-  * in the documentation and/or other materials provided with the
 
-  * distribution.
 
-  *     * Neither the name of Google Inc. nor the names of its
 
-  * contributors may be used to endorse or promote products derived from
 
-  * this software without specific prior written permission.
 
-  *
 
-  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
 
-  * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
 
-  * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
 
-  * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
 
-  * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
 
-  * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
 
-  * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
 
-  * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
 
-  * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
 
-  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
 
-  * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
 
-  *
 
-  */
 
- #include <cinttypes>
 
- #include <deque>
 
- #include <list>
 
- #include <thread>
 
- #include <unordered_map>
 
- #include <vector>
 
- #include <grpc++/channel.h>
 
- #include <grpc++/client_context.h>
 
- #include <grpc++/create_channel.h>
 
- #include <grpc/support/alloc.h>
 
- #include <grpc/support/host_port.h>
 
- #include <grpc/support/log.h>
 
- #include "src/core/lib/support/env.h"
 
- #include "src/proto/grpc/testing/services.grpc.pb.h"
 
- #include "test/core/util/port.h"
 
- #include "test/core/util/test_config.h"
 
- #include "test/cpp/qps/driver.h"
 
- #include "test/cpp/qps/histogram.h"
 
- #include "test/cpp/qps/qps_worker.h"
 
- #include "test/cpp/qps/stats.h"
 
- using std::list;
 
- using std::thread;
 
- using std::unique_ptr;
 
- using std::deque;
 
- using std::vector;
 
- namespace grpc {
 
- namespace testing {
 
- static std::string get_host(const std::string& worker) {
 
-   char* host;
 
-   char* port;
 
-   gpr_split_host_port(worker.c_str(), &host, &port);
 
-   const string s(host);
 
-   gpr_free(host);
 
-   gpr_free(port);
 
-   return s;
 
- }
 
- static std::unordered_map<string, std::deque<int>> get_hosts_and_cores(
 
-     const deque<string>& workers) {
 
-   std::unordered_map<string, std::deque<int>> hosts;
 
-   for (auto it = workers.begin(); it != workers.end(); it++) {
 
-     const string host = get_host(*it);
 
-     if (hosts.find(host) == hosts.end()) {
 
-       auto stub = WorkerService::NewStub(
 
-           CreateChannel(*it, InsecureChannelCredentials()));
 
-       grpc::ClientContext ctx;
 
-       ctx.set_fail_fast(false);
 
-       CoreRequest dummy;
 
-       CoreResponse cores;
 
-       grpc::Status s = stub->CoreCount(&ctx, dummy, &cores);
 
-       GPR_ASSERT(s.ok());
 
-       std::deque<int> dq;
 
-       for (int i = 0; i < cores.cores(); i++) {
 
-         dq.push_back(i);
 
-       }
 
-       hosts[host] = dq;
 
-     }
 
-   }
 
-   return hosts;
 
- }
 
- static deque<string> get_workers(const string& name) {
 
-   char* env = gpr_getenv(name.c_str());
 
-   if (!env) return deque<string>();
 
-   deque<string> out;
 
-   char* p = env;
 
-   for (;;) {
 
-     char* comma = strchr(p, ',');
 
-     if (comma) {
 
-       out.emplace_back(p, comma);
 
-       p = comma + 1;
 
-     } else {
 
-       out.emplace_back(p);
 
-       gpr_free(env);
 
-       return out;
 
-     }
 
-   }
 
- }
 
- // helpers for postprocess_scenario_result
 
- static double WallTime(ClientStats s) { return s.time_elapsed(); }
 
- static double SystemTime(ClientStats s) { return s.time_system(); }
 
- static double UserTime(ClientStats s) { return s.time_user(); }
 
- static double ServerWallTime(ServerStats s) { return s.time_elapsed(); }
 
- static double ServerSystemTime(ServerStats s) { return s.time_system(); }
 
- static double ServerUserTime(ServerStats s) { return s.time_user(); }
 
- static int Cores(int n) { return n; }
 
- // Postprocess ScenarioResult and populate result summary.
 
- static void postprocess_scenario_result(ScenarioResult* result) {
 
-   Histogram histogram;
 
-   histogram.MergeProto(result->latencies());
 
-   auto qps = histogram.Count() / average(result->client_stats(), WallTime);
 
-   auto qps_per_server_core = qps / sum(result->server_cores(), Cores);
 
-   result->mutable_summary()->set_qps(qps);
 
-   result->mutable_summary()->set_qps_per_server_core(qps_per_server_core);
 
-   result->mutable_summary()->set_latency_50(histogram.Percentile(50));
 
-   result->mutable_summary()->set_latency_90(histogram.Percentile(90));
 
-   result->mutable_summary()->set_latency_95(histogram.Percentile(95));
 
-   result->mutable_summary()->set_latency_99(histogram.Percentile(99));
 
-   result->mutable_summary()->set_latency_999(histogram.Percentile(99.9));
 
-   auto server_system_time = 100.0 *
 
-                             sum(result->server_stats(), ServerSystemTime) /
 
-                             sum(result->server_stats(), ServerWallTime);
 
-   auto server_user_time = 100.0 * sum(result->server_stats(), ServerUserTime) /
 
-                           sum(result->server_stats(), ServerWallTime);
 
-   auto client_system_time = 100.0 * sum(result->client_stats(), SystemTime) /
 
-                             sum(result->client_stats(), WallTime);
 
-   auto client_user_time = 100.0 * sum(result->client_stats(), UserTime) /
 
-                           sum(result->client_stats(), WallTime);
 
-   result->mutable_summary()->set_server_system_time(server_system_time);
 
-   result->mutable_summary()->set_server_user_time(server_user_time);
 
-   result->mutable_summary()->set_client_system_time(client_system_time);
 
-   result->mutable_summary()->set_client_user_time(client_user_time);
 
- }
 
- // Namespace for classes and functions used only in RunScenario
 
- // Using this rather than local definitions to workaround gcc-4.4 limitations
 
- // regarding using templates without linkage
 
- namespace runsc {
 
- // ClientContext allocator
 
- static ClientContext* AllocContext(list<ClientContext>* contexts) {
 
-   contexts->emplace_back();
 
-   auto context = &contexts->back();
 
-   context->set_fail_fast(false);
 
-   return context;
 
- }
 
- struct ServerData {
 
-   unique_ptr<WorkerService::Stub> stub;
 
-   unique_ptr<ClientReaderWriter<ServerArgs, ServerStatus>> stream;
 
- };
 
- struct ClientData {
 
-   unique_ptr<WorkerService::Stub> stub;
 
-   unique_ptr<ClientReaderWriter<ClientArgs, ClientStatus>> stream;
 
- };
 
- }  // namespace runsc
 
- std::unique_ptr<ScenarioResult> RunScenario(
 
-     const ClientConfig& initial_client_config, size_t num_clients,
 
-     const ServerConfig& initial_server_config, size_t num_servers,
 
-     int warmup_seconds, int benchmark_seconds, int spawn_local_worker_count) {
 
-   // ClientContext allocations (all are destroyed at scope exit)
 
-   list<ClientContext> contexts;
 
-   // To be added to the result, containing the final configuration used for
 
-   // client and config (including host, etc.)
 
-   ClientConfig result_client_config;
 
-   const ServerConfig result_server_config = initial_server_config;
 
-   // Get client, server lists
 
-   auto workers = get_workers("QPS_WORKERS");
 
-   ClientConfig client_config = initial_client_config;
 
-   // Spawn some local workers if desired
 
-   vector<unique_ptr<QpsWorker>> local_workers;
 
-   for (int i = 0; i < abs(spawn_local_worker_count); i++) {
 
-     // act as if we're a new test -- gets a good rng seed
 
-     static bool called_init = false;
 
-     if (!called_init) {
 
-       char args_buf[100];
 
-       strcpy(args_buf, "some-benchmark");
 
-       char* args[] = {args_buf};
 
-       grpc_test_init(1, args);
 
-       called_init = true;
 
-     }
 
-     int driver_port = grpc_pick_unused_port_or_die();
 
-     local_workers.emplace_back(new QpsWorker(driver_port));
 
-     char addr[256];
 
-     sprintf(addr, "localhost:%d", driver_port);
 
-     if (spawn_local_worker_count < 0) {
 
-       workers.push_front(addr);
 
-     } else {
 
-       workers.push_back(addr);
 
-     }
 
-   }
 
-   // Setup the hosts and core counts
 
-   auto hosts_cores = get_hosts_and_cores(workers);
 
-   // if num_clients is set to <=0, do dynamic sizing: all workers
 
-   // except for servers are clients
 
-   if (num_clients <= 0) {
 
-     num_clients = workers.size() - num_servers;
 
-   }
 
-   // TODO(ctiller): support running multiple configurations, and binpack
 
-   // client/server pairs
 
-   // to available workers
 
-   GPR_ASSERT(workers.size() >= num_clients + num_servers);
 
-   // Trim to just what we need
 
-   workers.resize(num_clients + num_servers);
 
-   // Start servers
 
-   using runsc::ServerData;
 
-   // servers is array rather than std::vector to avoid gcc-4.4 issues
 
-   // where class contained in std::vector must have a copy constructor
 
-   auto* servers = new ServerData[num_servers];
 
-   for (size_t i = 0; i < num_servers; i++) {
 
-     gpr_log(GPR_INFO, "Starting server on %s (worker #%" PRIuPTR ")",
 
-             workers[i].c_str(), i);
 
-     servers[i].stub = WorkerService::NewStub(
 
-         CreateChannel(workers[i], InsecureChannelCredentials()));
 
-     ServerConfig server_config = initial_server_config;
 
-     char* host;
 
-     char* driver_port;
 
-     char* cli_target;
 
-     gpr_split_host_port(workers[i].c_str(), &host, &driver_port);
 
-     string host_str(host);
 
-     int server_core_limit = initial_server_config.core_limit();
 
-     int client_core_limit = initial_client_config.core_limit();
 
-     if (server_core_limit == 0 && client_core_limit > 0) {
 
-       // In this case, limit the server cores if it matches the
 
-       // same host as one or more clients
 
-       const auto& dq = hosts_cores.at(host_str);
 
-       bool match = false;
 
-       int limit = dq.size();
 
-       for (size_t cli = 0; cli < num_clients; cli++) {
 
-         if (host_str == get_host(workers[cli + num_servers])) {
 
-           limit -= client_core_limit;
 
-           match = true;
 
-         }
 
-       }
 
-       if (match) {
 
-         GPR_ASSERT(limit > 0);
 
-         server_core_limit = limit;
 
-       }
 
-     }
 
-     if (server_core_limit > 0) {
 
-       auto& dq = hosts_cores.at(host_str);
 
-       GPR_ASSERT(dq.size() >= static_cast<size_t>(server_core_limit));
 
-       for (int core = 0; core < server_core_limit; core++) {
 
-         server_config.add_core_list(dq.front());
 
-         dq.pop_front();
 
-       }
 
-     }
 
-     ServerArgs args;
 
-     *args.mutable_setup() = server_config;
 
-     servers[i].stream =
 
-         servers[i].stub->RunServer(runsc::AllocContext(&contexts));
 
-     if (!servers[i].stream->Write(args)) {
 
-       gpr_log(GPR_ERROR, "Could not write args to server %zu", i);
 
-     }
 
-     ServerStatus init_status;
 
-     if (!servers[i].stream->Read(&init_status)) {
 
-       gpr_log(GPR_ERROR, "Server %zu did not yield initial status", i);
 
-     }
 
-     gpr_join_host_port(&cli_target, host, init_status.port());
 
-     client_config.add_server_targets(cli_target);
 
-     gpr_free(host);
 
-     gpr_free(driver_port);
 
-     gpr_free(cli_target);
 
-   }
 
-   // Targets are all set by now
 
-   result_client_config = client_config;
 
-   // Start clients
 
-   using runsc::ClientData;
 
-   // clients is array rather than std::vector to avoid gcc-4.4 issues
 
-   // where class contained in std::vector must have a copy constructor
 
-   auto* clients = new ClientData[num_clients];
 
-   for (size_t i = 0; i < num_clients; i++) {
 
-     const auto& worker = workers[i + num_servers];
 
-     gpr_log(GPR_INFO, "Starting client on %s (worker #%" PRIuPTR ")",
 
-             worker.c_str(), i + num_servers);
 
-     clients[i].stub = WorkerService::NewStub(
 
-         CreateChannel(worker, InsecureChannelCredentials()));
 
-     ClientConfig per_client_config = client_config;
 
-     int server_core_limit = initial_server_config.core_limit();
 
-     int client_core_limit = initial_client_config.core_limit();
 
-     if ((server_core_limit > 0) || (client_core_limit > 0)) {
 
-       auto& dq = hosts_cores.at(get_host(worker));
 
-       if (client_core_limit == 0) {
 
-         // limit client cores if it matches a server host
 
-         bool match = false;
 
-         int limit = dq.size();
 
-         for (size_t srv = 0; srv < num_servers; srv++) {
 
-           if (get_host(worker) == get_host(workers[srv])) {
 
-             match = true;
 
-           }
 
-         }
 
-         if (match) {
 
-           GPR_ASSERT(limit > 0);
 
-           client_core_limit = limit;
 
-         }
 
-       }
 
-       if (client_core_limit > 0) {
 
-         GPR_ASSERT(dq.size() >= static_cast<size_t>(client_core_limit));
 
-         for (int core = 0; core < client_core_limit; core++) {
 
-           per_client_config.add_core_list(dq.front());
 
-           dq.pop_front();
 
-         }
 
-       }
 
-     }
 
-     ClientArgs args;
 
-     *args.mutable_setup() = per_client_config;
 
-     clients[i].stream =
 
-         clients[i].stub->RunClient(runsc::AllocContext(&contexts));
 
-     if (!clients[i].stream->Write(args)) {
 
-       gpr_log(GPR_ERROR, "Could not write args to client %zu", i);
 
-     }
 
-     ClientStatus init_status;
 
-     if (!clients[i].stream->Read(&init_status)) {
 
-       gpr_log(GPR_ERROR, "Client %zu did not yield initial status", i);
 
-     }
 
-   }
 
-   // Let everything warmup
 
-   gpr_log(GPR_INFO, "Warming up");
 
-   gpr_timespec start = gpr_now(GPR_CLOCK_REALTIME);
 
-   gpr_sleep_until(
 
-       gpr_time_add(start, gpr_time_from_seconds(warmup_seconds, GPR_TIMESPAN)));
 
-   // Start a run
 
-   gpr_log(GPR_INFO, "Starting");
 
-   ServerArgs server_mark;
 
-   server_mark.mutable_mark()->set_reset(true);
 
-   ClientArgs client_mark;
 
-   client_mark.mutable_mark()->set_reset(true);
 
-   for (size_t i = 0; i < num_servers; i++) {
 
-     auto server = &servers[i];
 
-     if (!server->stream->Write(server_mark)) {
 
-       gpr_log(GPR_ERROR, "Couldn't write mark to server %zu", i);
 
-     }
 
-   }
 
-   for (size_t i = 0; i < num_clients; i++) {
 
-     auto client = &clients[i];
 
-     if (!client->stream->Write(client_mark)) {
 
-       gpr_log(GPR_ERROR, "Couldn't write mark to client %zu", i);
 
-     }
 
-   }
 
-   ServerStatus server_status;
 
-   ClientStatus client_status;
 
-   for (size_t i = 0; i < num_servers; i++) {
 
-     auto server = &servers[i];
 
-     if (!server->stream->Read(&server_status)) {
 
-       gpr_log(GPR_ERROR, "Couldn't get status from server %zu", i);
 
-     }
 
-   }
 
-   for (size_t i = 0; i < num_clients; i++) {
 
-     auto client = &clients[i];
 
-     if (!client->stream->Read(&client_status)) {
 
-       gpr_log(GPR_ERROR, "Couldn't get status from client %zu", i);
 
-     }
 
-   }
 
-   // Wait some time
 
-   gpr_log(GPR_INFO, "Running");
 
-   // Use gpr_sleep_until rather than this_thread::sleep_until to support
 
-   // compilers that don't work with this_thread
 
-   gpr_sleep_until(gpr_time_add(
 
-       start,
 
-       gpr_time_from_seconds(warmup_seconds + benchmark_seconds, GPR_TIMESPAN)));
 
-   // Finish a run
 
-   std::unique_ptr<ScenarioResult> result(new ScenarioResult);
 
-   Histogram merged_latencies;
 
-   gpr_log(GPR_INFO, "Finishing clients");
 
-   for (size_t i = 0; i < num_clients; i++) {
 
-     auto client = &clients[i];
 
-     if (!client->stream->Write(client_mark)) {
 
-       gpr_log(GPR_ERROR, "Couldn't write mark to client %zu", i);
 
-     }
 
-     if (!client->stream->WritesDone()) {
 
-       gpr_log(GPR_ERROR, "Failed WritesDone for client %zu", i);
 
-     }
 
-   }
 
-   for (size_t i = 0; i < num_clients; i++) {
 
-     auto client = &clients[i];
 
-     // Read the client final status
 
-     if (client->stream->Read(&client_status)) {
 
-       gpr_log(GPR_INFO, "Received final status from client %zu", i);
 
-       const auto& stats = client_status.stats();
 
-       merged_latencies.MergeProto(stats.latencies());
 
-       result->add_client_stats()->CopyFrom(stats);
 
-       // That final status should be the last message on the client stream
 
-       GPR_ASSERT(!client->stream->Read(&client_status));
 
-     } else {
 
-       gpr_log(GPR_ERROR, "Couldn't get final status from client %zu", i);
 
-     }
 
-   }
 
-   for (size_t i = 0; i < num_clients; i++) {
 
-     auto client = &clients[i];
 
-     Status s = client->stream->Finish();
 
-     result->add_client_success(s.ok());
 
-     if (!s.ok()) {
 
-       gpr_log(GPR_ERROR, "Client %zu had an error %s", i,
 
-               s.error_message().c_str());
 
-     }
 
-   }
 
-   delete[] clients;
 
-   merged_latencies.FillProto(result->mutable_latencies());
 
-   gpr_log(GPR_INFO, "Finishing servers");
 
-   for (size_t i = 0; i < num_servers; i++) {
 
-     auto server = &servers[i];
 
-     if (!server->stream->Write(server_mark)) {
 
-       gpr_log(GPR_ERROR, "Couldn't write mark to server %zu", i);
 
-     }
 
-     if (!server->stream->WritesDone()) {
 
-       gpr_log(GPR_ERROR, "Failed WritesDone for server %zu", i);
 
-     }
 
-   }
 
-   for (size_t i = 0; i < num_servers; i++) {
 
-     auto server = &servers[i];
 
-     // Read the server final status
 
-     if (server->stream->Read(&server_status)) {
 
-       gpr_log(GPR_INFO, "Received final status from server %zu", i);
 
-       result->add_server_stats()->CopyFrom(server_status.stats());
 
-       result->add_server_cores(server_status.cores());
 
-       // That final status should be the last message on the server stream
 
-       GPR_ASSERT(!server->stream->Read(&server_status));
 
-     } else {
 
-       gpr_log(GPR_ERROR, "Couldn't get final status from server %zu", i);
 
-     }
 
-   }
 
-   for (size_t i = 0; i < num_servers; i++) {
 
-     auto server = &servers[i];
 
-     Status s = server->stream->Finish();
 
-     result->add_server_success(s.ok());
 
-     if (!s.ok()) {
 
-       gpr_log(GPR_ERROR, "Server %zu had an error %s", i,
 
-               s.error_message().c_str());
 
-     }
 
-   }
 
-   delete[] servers;
 
-   postprocess_scenario_result(result.get());
 
-   return result;
 
- }
 
- bool RunQuit() {
 
-   // Get client, server lists
 
-   bool result = true;
 
-   auto workers = get_workers("QPS_WORKERS");
 
-   for (size_t i = 0; i < workers.size(); i++) {
 
-     auto stub = WorkerService::NewStub(
 
-         CreateChannel(workers[i], InsecureChannelCredentials()));
 
-     Void dummy;
 
-     grpc::ClientContext ctx;
 
-     ctx.set_fail_fast(false);
 
-     Status s = stub->QuitWorker(&ctx, dummy, &dummy);
 
-     if (!s.ok()) {
 
-       gpr_log(GPR_ERROR, "Worker %zu could not be properly quit because %s", i,
 
-               s.error_message().c_str());
 
-       result = false;
 
-     }
 
-   }
 
-   return result;
 
- }
 
- }  // namespace testing
 
- }  // namespace grpc
 
 
  |