123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686 |
- /*
- *
- * Copyright 2015 gRPC authors.
- *
- * Licensed under the Apache License, Version 2.0 (the "License");
- * you may not use this file except in compliance with the License.
- * You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- *
- */
- #include "test/cpp/qps/driver.h"
- #include <cinttypes>
- #include <deque>
- #include <list>
- #include <thread>
- #include <unordered_map>
- #include <vector>
- #include <grpc/support/alloc.h>
- #include <grpc/support/log.h>
- #include <grpc/support/string_util.h>
- #include <grpcpp/channel.h>
- #include <grpcpp/client_context.h>
- #include <grpcpp/create_channel.h>
- #include "src/core/lib/gpr/env.h"
- #include "src/core/lib/gprpp/host_port.h"
- #include "src/core/lib/profiling/timers.h"
- #include "src/proto/grpc/testing/worker_service.grpc.pb.h"
- #include "test/core/util/port.h"
- #include "test/core/util/test_config.h"
- #include "test/cpp/qps/client.h"
- #include "test/cpp/qps/histogram.h"
- #include "test/cpp/qps/qps_worker.h"
- #include "test/cpp/qps/stats.h"
- #include "test/cpp/util/test_credentials_provider.h"
- using std::deque;
- using std::list;
- using std::unique_ptr;
- using std::vector;
- namespace grpc {
- namespace testing {
- static std::string get_host(const std::string& worker) {
- absl::string_view host;
- absl::string_view port;
- grpc_core::SplitHostPort(worker.c_str(), &host, &port);
- return std::string(host.data(), host.size());
- }
- static deque<string> get_workers(const string& env_name) {
- deque<string> out;
- char* env = gpr_getenv(env_name.c_str());
- if (!env) {
- env = gpr_strdup("");
- }
- char* p = env;
- if (strlen(env) != 0) {
- for (;;) {
- char* comma = strchr(p, ',');
- if (comma) {
- out.emplace_back(p, comma);
- p = comma + 1;
- } else {
- out.emplace_back(p);
- break;
- }
- }
- }
- if (out.empty()) {
- gpr_log(GPR_ERROR,
- "Environment variable \"%s\" does not contain a list of QPS "
- "workers to use. Set it to a comma-separated list of "
- "hostname:port pairs, starting with hosts that should act as "
- "servers. E.g. export "
- "%s=\"serverhost1:1234,clienthost1:1234,clienthost2:1234\"",
- env_name.c_str(), env_name.c_str());
- }
- gpr_free(env);
- return out;
- }
- std::string GetCredType(
- const std::string& worker_addr,
- const std::map<std::string, std::string>& per_worker_credential_types,
- const std::string& credential_type) {
- auto it = per_worker_credential_types.find(worker_addr);
- if (it != per_worker_credential_types.end()) {
- return it->second;
- }
- return credential_type;
- }
- // helpers for postprocess_scenario_result
- static double WallTime(const ClientStats& s) { return s.time_elapsed(); }
- static double SystemTime(const ClientStats& s) { return s.time_system(); }
- static double UserTime(const ClientStats& s) { return s.time_user(); }
- static double CliPollCount(const ClientStats& s) { return s.cq_poll_count(); }
- static double SvrPollCount(const ServerStats& s) { return s.cq_poll_count(); }
- static double ServerWallTime(const ServerStats& s) { return s.time_elapsed(); }
- static double ServerSystemTime(const ServerStats& s) { return s.time_system(); }
- static double ServerUserTime(const ServerStats& s) { return s.time_user(); }
- static double ServerTotalCpuTime(const ServerStats& s) {
- return s.total_cpu_time();
- }
- static double ServerIdleCpuTime(const ServerStats& s) {
- return s.idle_cpu_time();
- }
- static int Cores(int n) { return n; }
- static bool IsSuccess(const Status& s) {
- if (s.ok()) return true;
- // Since we shutdown servers and clients at the same time, they both can
- // observe cancellation. Thus, we consider CANCELLED as good status.
- if (static_cast<StatusCode>(s.error_code()) == StatusCode::CANCELLED) {
- return true;
- }
- // Since we shutdown servers and clients at the same time, server can close
- // the socket before the client attempts to do that, and vice versa. Thus
- // receiving a "Socket closed" error is fine.
- if (s.error_message() == "Socket closed") return true;
- return false;
- }
- // Postprocess ScenarioResult and populate result summary.
- static void postprocess_scenario_result(ScenarioResult* result) {
- // Get latencies from ScenarioResult latencies histogram and populate to
- // result summary.
- Histogram histogram;
- histogram.MergeProto(result->latencies());
- result->mutable_summary()->set_latency_50(histogram.Percentile(50));
- result->mutable_summary()->set_latency_90(histogram.Percentile(90));
- result->mutable_summary()->set_latency_95(histogram.Percentile(95));
- result->mutable_summary()->set_latency_99(histogram.Percentile(99));
- result->mutable_summary()->set_latency_999(histogram.Percentile(99.9));
- // Calculate qps and cpu load for each client and then aggregate results for
- // all clients
- double qps = 0;
- double client_system_cpu_load = 0, client_user_cpu_load = 0;
- for (int i = 0; i < result->client_stats_size(); i++) {
- auto client_stat = result->client_stats(i);
- qps += client_stat.latencies().count() / client_stat.time_elapsed();
- client_system_cpu_load +=
- client_stat.time_system() / client_stat.time_elapsed();
- client_user_cpu_load +=
- client_stat.time_user() / client_stat.time_elapsed();
- }
- // Calculate cpu load for each server and then aggregate results for all
- // servers
- double server_system_cpu_load = 0, server_user_cpu_load = 0;
- for (int i = 0; i < result->server_stats_size(); i++) {
- auto server_stat = result->server_stats(i);
- server_system_cpu_load +=
- server_stat.time_system() / server_stat.time_elapsed();
- server_user_cpu_load +=
- server_stat.time_user() / server_stat.time_elapsed();
- }
- result->mutable_summary()->set_qps(qps);
- // Populate the percentage of cpu load to result summary.
- result->mutable_summary()->set_server_system_time(100 *
- server_system_cpu_load);
- result->mutable_summary()->set_server_user_time(100 * server_user_cpu_load);
- result->mutable_summary()->set_client_system_time(100 *
- client_system_cpu_load);
- result->mutable_summary()->set_client_user_time(100 * client_user_cpu_load);
- // For Non-linux platform, get_cpu_usage() is not implemented. Thus,
- // ServerTotalCpuTime and ServerIdleCpuTime are both 0.
- if (average(result->server_stats(), ServerTotalCpuTime) == 0) {
- result->mutable_summary()->set_server_cpu_usage(0);
- } else {
- auto server_cpu_usage =
- 100 - 100 * average(result->server_stats(), ServerIdleCpuTime) /
- average(result->server_stats(), ServerTotalCpuTime);
- result->mutable_summary()->set_server_cpu_usage(server_cpu_usage);
- }
- // Calculate and populate successful request per second and failed requests
- // per seconds to result summary.
- auto time_estimate = average(result->client_stats(), WallTime);
- if (result->request_results_size() > 0) {
- int64_t successes = 0;
- int64_t failures = 0;
- for (int i = 0; i < result->request_results_size(); i++) {
- const RequestResultCount& rrc = result->request_results(i);
- if (rrc.status_code() == 0) {
- successes += rrc.count();
- } else {
- failures += rrc.count();
- }
- }
- result->mutable_summary()->set_successful_requests_per_second(
- successes / time_estimate);
- result->mutable_summary()->set_failed_requests_per_second(failures /
- time_estimate);
- }
- // Fill in data for other metrics required in result summary
- auto qps_per_server_core = qps / sum(result->server_cores(), Cores);
- result->mutable_summary()->set_qps_per_server_core(qps_per_server_core);
- result->mutable_summary()->set_client_polls_per_request(
- sum(result->client_stats(), CliPollCount) / histogram.Count());
- result->mutable_summary()->set_server_polls_per_request(
- sum(result->server_stats(), SvrPollCount) / histogram.Count());
- auto server_queries_per_cpu_sec =
- histogram.Count() / (sum(result->server_stats(), ServerSystemTime) +
- sum(result->server_stats(), ServerUserTime));
- auto client_queries_per_cpu_sec =
- histogram.Count() / (sum(result->client_stats(), SystemTime) +
- sum(result->client_stats(), UserTime));
- result->mutable_summary()->set_server_queries_per_cpu_sec(
- server_queries_per_cpu_sec);
- result->mutable_summary()->set_client_queries_per_cpu_sec(
- client_queries_per_cpu_sec);
- }
- struct ClientData {
- unique_ptr<WorkerService::Stub> stub;
- unique_ptr<ClientReaderWriter<ClientArgs, ClientStatus>> stream;
- };
- struct ServerData {
- unique_ptr<WorkerService::Stub> stub;
- unique_ptr<ClientReaderWriter<ServerArgs, ServerStatus>> stream;
- };
- static void FinishClients(const std::vector<ClientData>& clients,
- const ClientArgs& client_mark) {
- gpr_log(GPR_INFO, "Finishing clients");
- for (size_t i = 0, i_end = clients.size(); i < i_end; i++) {
- auto client = &clients[i];
- if (!client->stream->Write(client_mark)) {
- gpr_log(GPR_ERROR, "Couldn't write mark to client %zu", i);
- GPR_ASSERT(false);
- }
- if (!client->stream->WritesDone()) {
- gpr_log(GPR_ERROR, "Failed WritesDone for client %zu", i);
- GPR_ASSERT(false);
- }
- }
- }
- static void ReceiveFinalStatusFromClients(
- const std::vector<ClientData>& clients, Histogram& merged_latencies,
- std::unordered_map<int, int64_t>& merged_statuses, ScenarioResult& result) {
- gpr_log(GPR_INFO, "Receiving final status from clients");
- ClientStatus client_status;
- for (size_t i = 0, i_end = clients.size(); i < i_end; i++) {
- auto client = &clients[i];
- // Read the client final status
- if (client->stream->Read(&client_status)) {
- gpr_log(GPR_INFO, "Received final status from client %zu", i);
- const auto& stats = client_status.stats();
- merged_latencies.MergeProto(stats.latencies());
- for (int i = 0; i < stats.request_results_size(); i++) {
- merged_statuses[stats.request_results(i).status_code()] +=
- stats.request_results(i).count();
- }
- result.add_client_stats()->CopyFrom(stats);
- // Check that final status was should be the last message on the client
- // stream.
- // TODO(jtattermusch): note that that waiting for Read to return can take
- // long on some scenarios (e.g. unconstrained streaming_from_server). See
- // https://github.com/grpc/grpc/blob/3bd0cd208ea549760a2daf595f79b91b247fe240/test/cpp/qps/server_async.cc#L176
- // where the shutdown delay pretty much determines the wait here.
- GPR_ASSERT(!client->stream->Read(&client_status));
- } else {
- gpr_log(GPR_ERROR, "Couldn't get final status from client %zu", i);
- GPR_ASSERT(false);
- }
- }
- }
- static void ShutdownClients(const std::vector<ClientData>& clients,
- ScenarioResult& result) {
- gpr_log(GPR_INFO, "Shutdown clients");
- for (size_t i = 0, i_end = clients.size(); i < i_end; i++) {
- auto client = &clients[i];
- Status s = client->stream->Finish();
- // Since we shutdown servers and clients at the same time, clients can
- // observe cancellation. Thus, we consider both OK and CANCELLED as good
- // status.
- const bool success = IsSuccess(s);
- result.add_client_success(success);
- if (!success) {
- gpr_log(GPR_ERROR, "Client %zu had an error %s", i,
- s.error_message().c_str());
- GPR_ASSERT(false);
- }
- }
- }
- static void FinishServers(const std::vector<ServerData>& servers,
- const ServerArgs& server_mark) {
- gpr_log(GPR_INFO, "Finishing servers");
- for (size_t i = 0, i_end = servers.size(); i < i_end; i++) {
- auto server = &servers[i];
- if (!server->stream->Write(server_mark)) {
- gpr_log(GPR_ERROR, "Couldn't write mark to server %zu", i);
- GPR_ASSERT(false);
- }
- if (!server->stream->WritesDone()) {
- gpr_log(GPR_ERROR, "Failed WritesDone for server %zu", i);
- GPR_ASSERT(false);
- }
- }
- }
- static void ReceiveFinalStatusFromServer(const std::vector<ServerData>& servers,
- ScenarioResult& result) {
- gpr_log(GPR_INFO, "Receiving final status from servers");
- ServerStatus server_status;
- for (size_t i = 0, i_end = servers.size(); i < i_end; i++) {
- auto server = &servers[i];
- // Read the server final status
- if (server->stream->Read(&server_status)) {
- gpr_log(GPR_INFO, "Received final status from server %zu", i);
- result.add_server_stats()->CopyFrom(server_status.stats());
- result.add_server_cores(server_status.cores());
- // That final status should be the last message on the server stream
- GPR_ASSERT(!server->stream->Read(&server_status));
- } else {
- gpr_log(GPR_ERROR, "Couldn't get final status from server %zu", i);
- GPR_ASSERT(false);
- }
- }
- }
- static void ShutdownServers(const std::vector<ServerData>& servers,
- ScenarioResult& result) {
- gpr_log(GPR_INFO, "Shutdown servers");
- for (size_t i = 0, i_end = servers.size(); i < i_end; i++) {
- auto server = &servers[i];
- Status s = server->stream->Finish();
- // Since we shutdown servers and clients at the same time, servers can
- // observe cancellation. Thus, we consider both OK and CANCELLED as good
- // status.
- const bool success = IsSuccess(s);
- result.add_server_success(success);
- if (!success) {
- gpr_log(GPR_ERROR, "Server %zu had an error %s", i,
- s.error_message().c_str());
- GPR_ASSERT(false);
- }
- }
- }
- std::vector<grpc::testing::Server*>* g_inproc_servers = nullptr;
- std::unique_ptr<ScenarioResult> RunScenario(
- const ClientConfig& initial_client_config, size_t num_clients,
- const ServerConfig& initial_server_config, size_t num_servers,
- int warmup_seconds, int benchmark_seconds, int spawn_local_worker_count,
- const std::string& qps_server_target_override,
- const std::string& credential_type,
- const std::map<std::string, std::string>& per_worker_credential_types,
- bool run_inproc, int32_t median_latency_collection_interval_millis) {
- if (run_inproc) {
- g_inproc_servers = new std::vector<grpc::testing::Server*>;
- }
- // Log everything from the driver
- gpr_set_log_verbosity(GPR_LOG_SEVERITY_DEBUG);
- // ClientContext allocations (all are destroyed at scope exit)
- list<ClientContext> contexts;
- auto alloc_context = [](list<ClientContext>* contexts) {
- contexts->emplace_back();
- auto context = &contexts->back();
- context->set_wait_for_ready(true);
- return context;
- };
- // To be added to the result, containing the final configuration used for
- // client and config (including host, etc.)
- ClientConfig result_client_config;
- // Get client, server lists; ignore if inproc test
- auto workers = (!run_inproc) ? get_workers("QPS_WORKERS") : deque<string>();
- ClientConfig client_config = initial_client_config;
- // Spawn some local workers if desired
- vector<unique_ptr<QpsWorker>> local_workers;
- for (int i = 0; i < abs(spawn_local_worker_count); i++) {
- // act as if we're a new test -- gets a good rng seed
- static bool called_init = false;
- if (!called_init) {
- char args_buf[100];
- strcpy(args_buf, "some-benchmark");
- char* args[] = {args_buf};
- grpc_test_init(1, args);
- called_init = true;
- }
- char addr[256];
- // we use port # of -1 to indicate inproc
- int driver_port = (!run_inproc) ? grpc_pick_unused_port_or_die() : -1;
- local_workers.emplace_back(new QpsWorker(driver_port, 0, credential_type));
- sprintf(addr, "localhost:%d", driver_port);
- if (spawn_local_worker_count < 0) {
- workers.push_front(addr);
- } else {
- workers.push_back(addr);
- }
- }
- GPR_ASSERT(!workers.empty());
- // if num_clients is set to <=0, do dynamic sizing: all workers
- // except for servers are clients
- if (num_clients <= 0) {
- num_clients = workers.size() - num_servers;
- }
- // TODO(ctiller): support running multiple configurations, and binpack
- // client/server pairs
- // to available workers
- GPR_ASSERT(workers.size() >= num_clients + num_servers);
- // Trim to just what we need
- workers.resize(num_clients + num_servers);
- // Start servers
- std::vector<ServerData> servers(num_servers);
- std::unordered_map<string, std::deque<int>> hosts_cores;
- ChannelArguments channel_args;
- for (size_t i = 0; i < num_servers; i++) {
- gpr_log(GPR_INFO, "Starting server on %s (worker #%" PRIuPTR ")",
- workers[i].c_str(), i);
- if (!run_inproc) {
- servers[i].stub = WorkerService::NewStub(grpc::CreateTestChannel(
- workers[i],
- GetCredType(workers[i], per_worker_credential_types, credential_type),
- nullptr /* call creds */, {} /* interceptor creators */));
- } else {
- servers[i].stub = WorkerService::NewStub(
- local_workers[i]->InProcessChannel(channel_args));
- }
- const ServerConfig& server_config = initial_server_config;
- if (server_config.core_limit() != 0) {
- gpr_log(GPR_ERROR,
- "server config core limit is set but ignored by driver");
- GPR_ASSERT(false);
- }
- ServerArgs args;
- *args.mutable_setup() = server_config;
- servers[i].stream = servers[i].stub->RunServer(alloc_context(&contexts));
- if (!servers[i].stream->Write(args)) {
- gpr_log(GPR_ERROR, "Could not write args to server %zu", i);
- GPR_ASSERT(false);
- }
- ServerStatus init_status;
- if (!servers[i].stream->Read(&init_status)) {
- gpr_log(GPR_ERROR, "Server %zu did not yield initial status", i);
- GPR_ASSERT(false);
- }
- if (run_inproc) {
- std::string cli_target(INPROC_NAME_PREFIX);
- cli_target += std::to_string(i);
- client_config.add_server_targets(cli_target);
- } else {
- std::string host = get_host(workers[i]);
- std::string cli_target =
- grpc_core::JoinHostPort(host.c_str(), init_status.port());
- client_config.add_server_targets(cli_target.c_str());
- }
- }
- if (qps_server_target_override.length() > 0) {
- // overriding the qps server target only makes since if there is <= 1
- // servers
- GPR_ASSERT(num_servers <= 1);
- client_config.clear_server_targets();
- client_config.add_server_targets(qps_server_target_override);
- }
- client_config.set_median_latency_collection_interval_millis(
- median_latency_collection_interval_millis);
- // Targets are all set by now
- result_client_config = client_config;
- // Start clients
- std::vector<ClientData> clients(num_clients);
- size_t channels_allocated = 0;
- for (size_t i = 0; i < num_clients; i++) {
- const auto& worker = workers[i + num_servers];
- gpr_log(GPR_INFO, "Starting client on %s (worker #%" PRIuPTR ")",
- worker.c_str(), i + num_servers);
- if (!run_inproc) {
- clients[i].stub = WorkerService::NewStub(grpc::CreateTestChannel(
- worker,
- GetCredType(worker, per_worker_credential_types, credential_type),
- nullptr /* call creds */, {} /* interceptor creators */));
- } else {
- clients[i].stub = WorkerService::NewStub(
- local_workers[i + num_servers]->InProcessChannel(channel_args));
- }
- ClientConfig per_client_config = client_config;
- if (initial_client_config.core_limit() != 0) {
- gpr_log(GPR_ERROR, "client config core limit set but ignored");
- GPR_ASSERT(false);
- }
- // Reduce channel count so that total channels specified is held regardless
- // of the number of clients available
- size_t num_channels =
- (client_config.client_channels() - channels_allocated) /
- (num_clients - i);
- channels_allocated += num_channels;
- gpr_log(GPR_DEBUG, "Client %" PRIdPTR " gets %" PRIdPTR " channels", i,
- num_channels);
- per_client_config.set_client_channels(num_channels);
- ClientArgs args;
- *args.mutable_setup() = per_client_config;
- clients[i].stream = clients[i].stub->RunClient(alloc_context(&contexts));
- if (!clients[i].stream->Write(args)) {
- gpr_log(GPR_ERROR, "Could not write args to client %zu", i);
- GPR_ASSERT(false);
- }
- }
- for (size_t i = 0; i < num_clients; i++) {
- ClientStatus init_status;
- if (!clients[i].stream->Read(&init_status)) {
- gpr_log(GPR_ERROR, "Client %zu did not yield initial status", i);
- GPR_ASSERT(false);
- }
- }
- // Send an initial mark: clients can use this to know that everything is ready
- // to start
- gpr_log(GPR_INFO, "Initiating");
- ServerArgs server_mark;
- server_mark.mutable_mark()->set_reset(true);
- ClientArgs client_mark;
- client_mark.mutable_mark()->set_reset(true);
- ServerStatus server_status;
- ClientStatus client_status;
- for (size_t i = 0; i < num_clients; i++) {
- auto client = &clients[i];
- if (!client->stream->Write(client_mark)) {
- gpr_log(GPR_ERROR, "Couldn't write mark to client %zu", i);
- GPR_ASSERT(false);
- }
- }
- for (size_t i = 0; i < num_clients; i++) {
- auto client = &clients[i];
- if (!client->stream->Read(&client_status)) {
- gpr_log(GPR_ERROR, "Couldn't get status from client %zu", i);
- GPR_ASSERT(false);
- }
- }
- // Let everything warmup
- gpr_log(GPR_INFO, "Warming up");
- gpr_timespec start = gpr_now(GPR_CLOCK_REALTIME);
- gpr_sleep_until(
- gpr_time_add(start, gpr_time_from_seconds(warmup_seconds, GPR_TIMESPAN)));
- // Start a run
- gpr_log(GPR_INFO, "Starting");
- for (size_t i = 0; i < num_servers; i++) {
- auto server = &servers[i];
- if (!server->stream->Write(server_mark)) {
- gpr_log(GPR_ERROR, "Couldn't write mark to server %zu", i);
- GPR_ASSERT(false);
- }
- }
- for (size_t i = 0; i < num_clients; i++) {
- auto client = &clients[i];
- if (!client->stream->Write(client_mark)) {
- gpr_log(GPR_ERROR, "Couldn't write mark to client %zu", i);
- GPR_ASSERT(false);
- }
- }
- for (size_t i = 0; i < num_servers; i++) {
- auto server = &servers[i];
- if (!server->stream->Read(&server_status)) {
- gpr_log(GPR_ERROR, "Couldn't get status from server %zu", i);
- GPR_ASSERT(false);
- }
- }
- for (size_t i = 0; i < num_clients; i++) {
- auto client = &clients[i];
- if (!client->stream->Read(&client_status)) {
- gpr_log(GPR_ERROR, "Couldn't get status from client %zu", i);
- GPR_ASSERT(false);
- }
- }
- // Wait some time
- gpr_log(GPR_INFO, "Running");
- // Use gpr_sleep_until rather than this_thread::sleep_until to support
- // compilers that don't work with this_thread
- gpr_sleep_until(gpr_time_add(
- start,
- gpr_time_from_seconds(warmup_seconds + benchmark_seconds, GPR_TIMESPAN)));
- gpr_timer_set_enabled(0);
- // Finish a run
- std::unique_ptr<ScenarioResult> result(new ScenarioResult);
- Histogram merged_latencies;
- std::unordered_map<int, int64_t> merged_statuses;
- // For the case where clients lead the test such as UNARY and
- // STREAMING_FROM_CLIENT, clients need to finish completely while a server
- // is running to prevent the clients from being stuck while waiting for
- // the result.
- bool client_finish_first =
- (client_config.rpc_type() != STREAMING_FROM_SERVER);
- FinishClients(clients, client_mark);
- if (!client_finish_first) {
- FinishServers(servers, server_mark);
- }
- ReceiveFinalStatusFromClients(clients, merged_latencies, merged_statuses,
- *result);
- ShutdownClients(clients, *result);
- if (client_finish_first) {
- FinishServers(servers, server_mark);
- }
- ReceiveFinalStatusFromServer(servers, *result);
- ShutdownServers(servers, *result);
- delete g_inproc_servers;
- merged_latencies.FillProto(result->mutable_latencies());
- for (std::unordered_map<int, int64_t>::iterator it = merged_statuses.begin();
- it != merged_statuses.end(); ++it) {
- RequestResultCount* rrc = result->add_request_results();
- rrc->set_status_code(it->first);
- rrc->set_count(it->second);
- }
- postprocess_scenario_result(result.get());
- return result;
- }
- bool RunQuit(
- const std::string& credential_type,
- const std::map<std::string, std::string>& per_worker_credential_types) {
- // Get client, server lists
- bool result = true;
- auto workers = get_workers("QPS_WORKERS");
- if (workers.empty()) {
- return false;
- }
- for (size_t i = 0; i < workers.size(); i++) {
- auto stub = WorkerService::NewStub(grpc::CreateTestChannel(
- workers[i],
- GetCredType(workers[i], per_worker_credential_types, credential_type),
- nullptr /* call creds */, {} /* interceptor creators */));
- Void phony;
- grpc::ClientContext ctx;
- ctx.set_wait_for_ready(true);
- Status s = stub->QuitWorker(&ctx, phony, &phony);
- if (!s.ok()) {
- gpr_log(GPR_ERROR, "Worker %zu could not be properly quit because %s", i,
- s.error_message().c_str());
- result = false;
- }
- }
- return result;
- }
- } // namespace testing
- } // namespace grpc
|