| /* |
| * |
| * Copyright 2015 gRPC authors. |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| * |
| */ |
| |
| #include <chrono> |
| #include <memory> |
| #include <mutex> |
| #include <sstream> |
| #include <string> |
| #include <thread> |
| #include <vector> |
| |
| #include <grpc++/channel.h> |
| #include <grpc++/client_context.h> |
| #include <grpc++/server.h> |
| #include <grpc++/server_builder.h> |
| #include <grpc/grpc.h> |
| #include <grpc/support/alloc.h> |
| #include <grpc/support/host_port.h> |
| #include <grpc/support/log.h> |
| #include <grpc/support/time.h> |
| |
| #include "src/core/lib/profiling/timers.h" |
| #include "src/proto/grpc/testing/services.grpc.pb.h" |
| #include "test/cpp/qps/client.h" |
| #include "test/cpp/qps/interarrival.h" |
| #include "test/cpp/qps/usage_timer.h" |
| |
| namespace grpc { |
| namespace testing { |
| |
| static std::unique_ptr<BenchmarkService::Stub> BenchmarkStubCreator( |
| std::shared_ptr<Channel> ch) { |
| return BenchmarkService::NewStub(ch); |
| } |
| |
| class SynchronousClient |
| : public ClientImpl<BenchmarkService::Stub, SimpleRequest> { |
| public: |
| SynchronousClient(const ClientConfig& config) |
| : ClientImpl<BenchmarkService::Stub, SimpleRequest>( |
| config, BenchmarkStubCreator) { |
| num_threads_ = |
| config.outstanding_rpcs_per_channel() * config.client_channels(); |
| responses_.resize(num_threads_); |
| SetupLoadTest(config, num_threads_); |
| } |
| |
| virtual ~SynchronousClient(){}; |
| |
| protected: |
| // WaitToIssue returns false if we realize that we need to break out |
| bool WaitToIssue(int thread_idx) { |
| if (!closed_loop_) { |
| const gpr_timespec next_issue_time = NextIssueTime(thread_idx); |
| // Avoid sleeping for too long continuously because we might |
| // need to terminate before then. This is an issue since |
| // exponential distribution can occasionally produce bad outliers |
| while (true) { |
| const gpr_timespec one_sec_delay = |
| gpr_time_add(gpr_now(GPR_CLOCK_MONOTONIC), |
| gpr_time_from_seconds(1, GPR_TIMESPAN)); |
| if (gpr_time_cmp(next_issue_time, one_sec_delay) <= 0) { |
| gpr_sleep_until(next_issue_time); |
| return true; |
| } else { |
| gpr_sleep_until(one_sec_delay); |
| if (gpr_atm_acq_load(&thread_pool_done_) != static_cast<gpr_atm>(0)) { |
| return false; |
| } |
| } |
| } |
| } |
| return true; |
| } |
| |
| size_t num_threads_; |
| std::vector<SimpleResponse> responses_; |
| |
| private: |
| void DestroyMultithreading() override final { EndThreads(); } |
| }; |
| |
| class SynchronousUnaryClient final : public SynchronousClient { |
| public: |
| SynchronousUnaryClient(const ClientConfig& config) |
| : SynchronousClient(config) { |
| StartThreads(num_threads_); |
| } |
| ~SynchronousUnaryClient() {} |
| |
| bool ThreadFunc(HistogramEntry* entry, size_t thread_idx) override { |
| if (!WaitToIssue(thread_idx)) { |
| return true; |
| } |
| auto* stub = channels_[thread_idx % channels_.size()].get_stub(); |
| double start = UsageTimer::Now(); |
| GPR_TIMER_SCOPE("SynchronousUnaryClient::ThreadFunc", 0); |
| grpc::ClientContext context; |
| grpc::Status s = |
| stub->UnaryCall(&context, request_, &responses_[thread_idx]); |
| if (s.ok()) { |
| entry->set_value((UsageTimer::Now() - start) * 1e9); |
| } |
| entry->set_status(s.error_code()); |
| return true; |
| } |
| }; |
| |
| template <class StreamType> |
| class SynchronousStreamingClient : public SynchronousClient { |
| public: |
| SynchronousStreamingClient(const ClientConfig& config) |
| : SynchronousClient(config), |
| context_(num_threads_), |
| stream_(num_threads_), |
| messages_per_stream_(config.messages_per_stream()), |
| messages_issued_(num_threads_) { |
| StartThreads(num_threads_); |
| } |
| virtual ~SynchronousStreamingClient() { |
| std::vector<std::thread> cleanup_threads; |
| for (size_t i = 0; i < num_threads_; i++) { |
| cleanup_threads.emplace_back([this, i]() { |
| auto stream = &stream_[i]; |
| if (*stream) { |
| // forcibly cancel the streams, then finish |
| context_[i].TryCancel(); |
| (*stream)->Finish().IgnoreError(); |
| // don't log any error message on !ok since this was canceled |
| } |
| }); |
| } |
| for (auto& th : cleanup_threads) { |
| th.join(); |
| } |
| } |
| |
| protected: |
| std::vector<grpc::ClientContext> context_; |
| std::vector<std::unique_ptr<StreamType>> stream_; |
| const int messages_per_stream_; |
| std::vector<int> messages_issued_; |
| |
| void FinishStream(HistogramEntry* entry, size_t thread_idx) { |
| Status s = stream_[thread_idx]->Finish(); |
| // don't set the value since the stream is failed and shouldn't be timed |
| entry->set_status(s.error_code()); |
| if (!s.ok()) { |
| gpr_log(GPR_ERROR, "Stream %" PRIuPTR " received an error %s", thread_idx, |
| s.error_message().c_str()); |
| } |
| context_[thread_idx].~ClientContext(); |
| new (&context_[thread_idx]) ClientContext(); |
| } |
| }; |
| |
| class SynchronousStreamingPingPongClient final |
| : public SynchronousStreamingClient< |
| grpc::ClientReaderWriter<SimpleRequest, SimpleResponse>> { |
| public: |
| SynchronousStreamingPingPongClient(const ClientConfig& config) |
| : SynchronousStreamingClient(config) { |
| for (size_t thread_idx = 0; thread_idx < num_threads_; thread_idx++) { |
| auto* stub = channels_[thread_idx % channels_.size()].get_stub(); |
| stream_[thread_idx] = stub->StreamingCall(&context_[thread_idx]); |
| messages_issued_[thread_idx] = 0; |
| } |
| } |
| ~SynchronousStreamingPingPongClient() { |
| std::vector<std::thread> cleanup_threads; |
| for (size_t i = 0; i < num_threads_; i++) { |
| cleanup_threads.emplace_back([this, i]() { |
| auto stream = &stream_[i]; |
| if (*stream) { |
| (*stream)->WritesDone(); |
| } |
| }); |
| } |
| for (auto& th : cleanup_threads) { |
| th.join(); |
| } |
| } |
| |
| bool ThreadFunc(HistogramEntry* entry, size_t thread_idx) override { |
| if (!WaitToIssue(thread_idx)) { |
| return true; |
| } |
| GPR_TIMER_SCOPE("SynchronousStreamingPingPongClient::ThreadFunc", 0); |
| double start = UsageTimer::Now(); |
| if (stream_[thread_idx]->Write(request_) && |
| stream_[thread_idx]->Read(&responses_[thread_idx])) { |
| entry->set_value((UsageTimer::Now() - start) * 1e9); |
| // don't set the status since there isn't one yet |
| if ((messages_per_stream_ != 0) && |
| (++messages_issued_[thread_idx] < messages_per_stream_)) { |
| return true; |
| } else if (messages_per_stream_ == 0) { |
| return true; |
| } else { |
| // Fall through to the below resetting code after finish |
| } |
| } |
| stream_[thread_idx]->WritesDone(); |
| FinishStream(entry, thread_idx); |
| auto* stub = channels_[thread_idx % channels_.size()].get_stub(); |
| stream_[thread_idx] = stub->StreamingCall(&context_[thread_idx]); |
| messages_issued_[thread_idx] = 0; |
| return true; |
| } |
| }; |
| |
| class SynchronousStreamingFromClientClient final |
| : public SynchronousStreamingClient<grpc::ClientWriter<SimpleRequest>> { |
| public: |
| SynchronousStreamingFromClientClient(const ClientConfig& config) |
| : SynchronousStreamingClient(config), last_issue_(num_threads_) { |
| for (size_t thread_idx = 0; thread_idx < num_threads_; thread_idx++) { |
| auto* stub = channels_[thread_idx % channels_.size()].get_stub(); |
| stream_[thread_idx] = stub->StreamingFromClient(&context_[thread_idx], |
| &responses_[thread_idx]); |
| last_issue_[thread_idx] = UsageTimer::Now(); |
| } |
| } |
| ~SynchronousStreamingFromClientClient() { |
| std::vector<std::thread> cleanup_threads; |
| for (size_t i = 0; i < num_threads_; i++) { |
| cleanup_threads.emplace_back([this, i]() { |
| auto stream = &stream_[i]; |
| if (*stream) { |
| (*stream)->WritesDone(); |
| } |
| }); |
| } |
| for (auto& th : cleanup_threads) { |
| th.join(); |
| } |
| } |
| |
| bool ThreadFunc(HistogramEntry* entry, size_t thread_idx) override { |
| // Figure out how to make histogram sensible if this is rate-paced |
| if (!WaitToIssue(thread_idx)) { |
| return true; |
| } |
| GPR_TIMER_SCOPE("SynchronousStreamingFromClientClient::ThreadFunc", 0); |
| if (stream_[thread_idx]->Write(request_)) { |
| double now = UsageTimer::Now(); |
| entry->set_value((now - last_issue_[thread_idx]) * 1e9); |
| last_issue_[thread_idx] = now; |
| return true; |
| } |
| stream_[thread_idx]->WritesDone(); |
| FinishStream(entry, thread_idx); |
| auto* stub = channels_[thread_idx % channels_.size()].get_stub(); |
| stream_[thread_idx] = stub->StreamingFromClient(&context_[thread_idx], |
| &responses_[thread_idx]); |
| return true; |
| } |
| |
| private: |
| std::vector<double> last_issue_; |
| }; |
| |
| class SynchronousStreamingFromServerClient final |
| : public SynchronousStreamingClient<grpc::ClientReader<SimpleResponse>> { |
| public: |
| SynchronousStreamingFromServerClient(const ClientConfig& config) |
| : SynchronousStreamingClient(config), last_recv_(num_threads_) { |
| for (size_t thread_idx = 0; thread_idx < num_threads_; thread_idx++) { |
| auto* stub = channels_[thread_idx % channels_.size()].get_stub(); |
| stream_[thread_idx] = |
| stub->StreamingFromServer(&context_[thread_idx], request_); |
| last_recv_[thread_idx] = UsageTimer::Now(); |
| } |
| } |
| bool ThreadFunc(HistogramEntry* entry, size_t thread_idx) override { |
| GPR_TIMER_SCOPE("SynchronousStreamingFromServerClient::ThreadFunc", 0); |
| if (stream_[thread_idx]->Read(&responses_[thread_idx])) { |
| double now = UsageTimer::Now(); |
| entry->set_value((now - last_recv_[thread_idx]) * 1e9); |
| last_recv_[thread_idx] = now; |
| return true; |
| } |
| FinishStream(entry, thread_idx); |
| auto* stub = channels_[thread_idx % channels_.size()].get_stub(); |
| stream_[thread_idx] = |
| stub->StreamingFromServer(&context_[thread_idx], request_); |
| return true; |
| } |
| |
| private: |
| std::vector<double> last_recv_; |
| }; |
| |
| class SynchronousStreamingBothWaysClient final |
| : public SynchronousStreamingClient< |
| grpc::ClientReaderWriter<SimpleRequest, SimpleResponse>> { |
| public: |
| SynchronousStreamingBothWaysClient(const ClientConfig& config) |
| : SynchronousStreamingClient(config) { |
| for (size_t thread_idx = 0; thread_idx < num_threads_; thread_idx++) { |
| auto* stub = channels_[thread_idx % channels_.size()].get_stub(); |
| stream_[thread_idx] = stub->StreamingBothWays(&context_[thread_idx]); |
| } |
| } |
| ~SynchronousStreamingBothWaysClient() { |
| std::vector<std::thread> cleanup_threads; |
| for (size_t i = 0; i < num_threads_; i++) { |
| cleanup_threads.emplace_back([this, i]() { |
| auto stream = &stream_[i]; |
| if (*stream) { |
| (*stream)->WritesDone(); |
| } |
| }); |
| } |
| for (auto& th : cleanup_threads) { |
| th.join(); |
| } |
| } |
| |
| bool ThreadFunc(HistogramEntry* entry, size_t thread_idx) override { |
| // TODO (vjpai): Do this |
| return true; |
| } |
| }; |
| |
| std::unique_ptr<Client> CreateSynchronousClient(const ClientConfig& config) { |
| switch (config.rpc_type()) { |
| case UNARY: |
| return std::unique_ptr<Client>(new SynchronousUnaryClient(config)); |
| case STREAMING: |
| return std::unique_ptr<Client>( |
| new SynchronousStreamingPingPongClient(config)); |
| case STREAMING_FROM_CLIENT: |
| return std::unique_ptr<Client>( |
| new SynchronousStreamingFromClientClient(config)); |
| case STREAMING_FROM_SERVER: |
| return std::unique_ptr<Client>( |
| new SynchronousStreamingFromServerClient(config)); |
| case STREAMING_BOTH_WAYS: |
| return std::unique_ptr<Client>( |
| new SynchronousStreamingBothWaysClient(config)); |
| default: |
| assert(false); |
| return nullptr; |
| } |
| } |
| |
| } // namespace testing |
| } // namespace grpc |